Results 401-500 of 3503 (3416 ASCL, 87 submitted)

[ascl:2207.015]
calviacat: Calibrate star photometry by catalog comparison

calviacat calibrates star photometry by comparison to a catalog, including PanSTARRS 1, ATLAS-RefCat2, and SkyMapper catalogs. Catalog queries are cached so that subsequent calibrations of the same or similar fields can be more quickly executed.

[ascl:1105.013]
CAMB Sources: Number Counts, Lensing & Dark-age 21cm Power Spectra

We relate the observable number of sources per solid angle and redshift to the underlying proper source density and velocity, background evolution and line-of-sight potentials. We give an exact result in the case of linearized perturbations assuming general relativity. This consistently includes contributions of the source density perturbations and redshift distortions, magnification, radial displacement, and various additional linear terms that are small on sub-horizon scales. In addition we calculate the effect on observed luminosities, and hence the result for sources observed as a function of flux, including magnification bias and radial-displacement effects. We give the corresponding linear result for a magnitude-limited survey at low redshift, and discuss the angular power spectrum of the total count distribution. We also calculate the cross-correlation with the CMB polarization and temperature including Doppler source terms, magnification, redshift distortions and other velocity effects for the sources, and discuss why the contribution of redshift distortions is generally small. Finally we relate the result for source number counts to that for the brightness of line radiation, for example 21-cm radiation, from the sources.

[ascl:1102.026]
CAMB: Code for Anisotropies in the Microwave Background

We present a fully covariant and gauge-invariant calculation of the evolution of anisotropies in the cosmic microwave background (CMB) radiation. We use the physically appealing covariant approach to cosmological perturbations, which ensures that all variables are gauge-invariant and have a clear physical interpretation. We derive the complete set of frame-independent, linearised equations describing the (Boltzmann) evolution of anisotropy and inhomogeneity in an almost Friedmann-Robertson-Walker (FRW) cold dark matter (CDM) universe. These equations include the contributions of scalar, vector and tensor modes in a unified manner. Frame-independent equations for scalar and tensor perturbations, which are valid for any value of the background curvature, are obtained straightforwardly from the complete set of equations. We discuss the scalar equations in detail, including the integral solution and relation with the line of sight approach, analytic solutions in the early radiation dominated era, and the numerical solution in the standard CDM model. Our results confirm those obtained by other groups, who have worked carefully with non-covariant methods in specific gauges, but are derived here in a completely transparent fashion.

[ascl:1801.007]
cambmag: Magnetic Fields in CAMB

cambmag is a modification to CAMB (ascl:1102.026) that calculates the compensated magnetic mode in the scalar, vector and tensor case. Previously CAMB included code only for the vectors. It also corrects for tight-coupling issues and adds in the ability to include massive neutrinos when calculating vector modes.

[ascl:1605.006]
CAMELOT: Cloud Archive for MEtadata, Library and Online Toolkit

Ginsburg, Adam; Kruijssen, J. M. Diederik; Longmore, Steven N.; Koch, Eric; Glover, Simon C. O.; Dale, James E.; Commerçon, Benoît; Giannetti, Andrea; McLeod, Anna F.; Testi, Leonardo; Zahorecz, Sarolta; Rathborne, Jill M.; Zhang, Qizhou; Fontani, Francesco; Beltrán, Maite T.; Rivilla, Victor M.

CAMELOT facilitates the comparison of observational data and simulations of molecular clouds and/or star-forming regions. The central component of CAMELOT is a database summarizing the properties of observational data and simulations in the literature through pertinent metadata. The core functionality allows users to upload metadata, search and visualize the contents of the database to find and match observations/simulations over any range of parameter space.

To bridge the fundamental disconnect between inherently 2D observational data and 3D simulations, the code uses key physical properties that, in principle, are straightforward for both observers and simulators to measure — the surface density (Sigma), velocity dispersion (sigma) and radius (R). By determining these in a self-consistent way for all entries in the database, it should be possible to make robust comparisons.

[ascl:1502.015]
Camelus: Counts of Amplified Mass Elevations from Lensing with Ultrafast Simulations

Camelus provides a prediction on weak lensing peak counts from input cosmological parameters. Written in C, it samples halos from a mass function and assigns a profile, carries out ray-tracing simulations, and then counts peaks from ray-tracing maps. The creation of the ray-tracing simulations requires less computing time than N-body runs and the results is in good agreement with full N-body simulations.

[ascl:1505.030]
CANDID: Companion Analysis and Non-Detection in Interferometric Data

Gallenne, A.; Mérand, A.; Kervella, P.; Monnier, J. D.; Schaefer, G. H.; Baron, F.; Breitfelder, J.; Le Bouquin, J. B.; Roettenbacher, R. M.; Gieren, W.; Pietrzynski, G.; McAlister, H.; ten Brummelaar, T.; Sturmann, J.; Sturmann, L.; Turner, N.; Ridgway, S.; Kraus, S.

CANDID finds faint companion around star in interferometric data in the OIFITS format. It allows systematically searching for faint companions in OIFITS data, and if not found, estimates the detection limit. The tool is based on model fitting and Chi2 minimization, with a grid for the starting points of the companion position. It ensures all positions are explored by estimating a-posteriori if the grid is dense enough, and provides an estimate of the optimum grid density.

[ascl:1106.017]
CAOS: Code for Adaptive Optics Systems

The CAOS "system" (where CAOS stands for Code for Adaptive Optics Systems) is properly said a Problem Solving Environment (PSE). It is essentially composed of a graphical programming interface (the CAOS Application Builder) which can load different packages (set of modules). Current publicly distributed packages are the Software Package CAOS (the original adaptive optics package), the Software Package AIRY (an image-reconstruction-oriented package - AIRY stands for Astronomical Image Restoration with interferometrY), the Software Package PAOLAC (a simple CAOS interface for the analytic IDL code PAOLA developed by Laurent Jolissaint - PAOLAC stands for PAOLA within Caos), and a couple of private packages (not publicly distributed but restricted to the corresponding consortia): SPHERE (especially developed for the VLT planet finder SPHERE), and AIRY-LN (a specialized version of AIRY for the LBT instrument LINC-NIRVANA). Another package is also being developed: MAOS (that stands for Multiconjugate Adaptive Optics Simulations), developed for multi-reference multiconjugate AO studies purpose but still in a beta-version form.

[ascl:1404.011]
CAP_LOESS_1D & CAP_LOESS_2D: Recover mean trends from noisy data

CAP_LOESS_1D and CAP_LOESS_2D provide improved implementations of the one-dimensional (Clevelend 1979) and two-dimensional (Cleveland & Devlin 1988) Locally Weighted Regression (LOESS) methods to recover the mean trends of the population from noisy data in one or two dimensions. They include a robust approach to deal with outliers (bad data). The software is available in both IDL and Python versions.

[ascl:2011.002]
CAPTURE: Interferometric pipeline for image creation from GMRT data

CAPTURE (CAsa Pipeline-cum-Toolkit for Upgraded Giant Metrewave Radio Telescope data REduction) produces continuum images from radio interferometric data. Written in Python, it uses CASA (ascl:1107.013) tasks to analyze data obtained by the GMRT. It can produce self-calibrated images in a fully automatic mode or can run in steps to allow the data to be inspected throughout processing.

[ascl:2308.009]
caput: Utilities for building radio astronomy data analysis pipelines

Shaw, J. Richard; Masui, Kiyoshi; Nitsche, Rick; Boskovic, Anja; Zuo, Shifan; Gray, Liam; Fandino, Mateus; Wiebe, Donald V.; Siegel, Seth R.

Caput (Cluster Astronomical Python Utilities) contains utilities for handling large datasets on computer clusters. Written with radio astronomy in mind, the package provides an infrastructure for building, managing and configuring pipelines for data processing. It includes modules for dynamically importing and utilizing mpi4py, in-memory mock-ups of h5py objects, and infrastructure for running data analysis pipelines on computer clusters. Caput features a generic container for holding self-documenting datasets in memory with straightforward syncing to h5py files, and offers specialization for holding time stream data. Caput also includes tools for MPI-parallel analysis and routines for converting between different time representations, dealing with leap seconds, and calculating celestial times.

[ascl:2006.014]
CARACal: Containerized Automated Radio Astronomy Calibration pipeline

Józsa, Gyula I. G.; White, Sarah V.; Thorat, Kshitij; Smirnov, Oleg M.; Serra, Paolo; Ramatsoku, Mpati; Ramaila, Athanaseus J. T.; Perkins, Simon J.; Molnár, Dániel Cs.; Makhathini, Sphesihle; Maccagni, Filippo M.; Kleiner, Dane; Kamphuis, Peter; Hugo, Benjamin V.; de Blok, W. J. G.; Andati, Lexy A. L.

CARACal (Containerized Automated Radio Astronomy Calibration, formerly MeerKATHI) reduces radio-interferometric data. Developed originally as an end-to-end continuum- and line imaging pipeline for MeerKAT, it can also be used with other radio telescopes. CARACal reduces large data sets and produces high-dynamic-range continuum images and spectroscopic data cubes. The pipeline is platform-independent and delivers imaging quality metrics to efficiently assess the data quality.

[ascl:1505.003]
caret: Classification and Regression Training

caret (Classification And REgression Training) provides functions for training and plotting classification and regression models. It contains tools for data splitting, pre-processing, feature selection, model tuning using resampling, and variable importance estimation, as well as other functionality.

[ascl:1404.009]
carma_pack: MCMC sampler for Bayesian inference

carma_pack is an MCMC sampler for performing Bayesian inference on continuous time autoregressive moving average models. These models may be used to model time series with irregular sampling. The MCMC sampler utilizes an adaptive Metropolis algorithm combined with parallel tempering.

[ascl:1611.016]
Carpet: Adaptive Mesh Refinement for the Cactus Framework

Carpet is an adaptive mesh refinement and multi-patch driver for the Cactus Framework (ascl:1102.013). Cactus is a software framework for solving time-dependent partial differential equations on block-structured grids, and Carpet acts as driver layer providing adaptive mesh refinement, multi-patch capability, as well as parallelization and efficient I/O.

[ascl:2005.007]
Carpyncho: VVV Catalog browser toolkit

Carpyncho browses catalogs to search for and characterize time variable data of the Vista Variables in the Via Lactea (VVV) Survey. The stacked pawprint data from the Cambridge Astronomical Science Unit's (CASU) Vista Data Flow System (VDFS) v>= 1.3 catalogs have been crossed matched with the VDFS CASU v1.3 tile catalogs into Parquet files, allowing detection and classification of periodic variables within this dataset.

[ascl:2103.021]
Carsus: Atomic database for astronomy

Kerzendorf, Wolfgang; Mishin, Mikhail; Pássaro, Ezequiel; Eweis, Youssef; Selsing, Jonatan; Sim, Stuart

Carsus manages atomic datasets. It requires Chianti (ascl:9911.004), and can read data from a variety of sources and output them to file formats readable by radiative transfer codes such as TARDIS (ascl:1402.018).

[ascl:2103.031]
CARTA: Cube Analysis and Rendering Tool for Astronomy

Comrie, Angus; Wang, Kuo-Song; Hsu, Shou-Chieh; Moraghan, Anthony; Harris, Pamela; Pang, Qi; Pińska, Adrianna; Chiang, Cheng-Chin; Simmonds, Rob; Chang, Tien-Hao; Jan, Hengtai; Lin, Ming-Yi

CARTA (Cube Analysis and Rendering Tool for Astronomy) is a image visualization and analysis tool designed for the ALMA, VLA, SKA pathfinders, and the ngVLA. If offers catalog support, shared region analytics, profile smoothing, and spectral line query, and more. CARTA adopts a client-server architecture suitable for visualizing images with large file sizes (GB to TB) easily obtained from ALMA, VLA, or SKA pathfinder observations; computation and data storage are handled by remote enterprise-class servers or clusters with high performance storage, while processed products are sent to clients only for visualization with modern web features, such as GPU-accelerated rendering. This architecture also enables users to interact with the ALMA and VLA science archives by using CARTA as an interface. CARTA provides a desktop version and a server version. The former is suitable for single-user usage with a laptop, a desktop, or a remote server in the "remote" execution mode. The latter is suitable for institution-wide deployment to support multiple users with user authentication and additional server-side features.

[ascl:2207.025]
casa_cube: Display and analyze astronomical data cubes

casa_cube provides an interface to data cubes generated by CASA (ascl:1107.013) or Gildas (ascl:1305.010). It performs simple tasks such as plotting given channel maps, moment maps, and line profile in various units, and also corrects for cloud extinction, reconvolves with a beam taper, and permits quick and easy comparisons with models.

[ascl:1107.013]
CASA: Common Astronomy Software Applications

CASA, the Common Astronomy Software Applications package, is being developed with the primary goal of supporting the data post-processing needs of the next generation of radio astronomical telescopes such as ALMA and EVLA. The package can process both interferometric and single dish data. The CASA infrastructure consists of a set of C++ tools bundled together under an iPython interface as a set of data reduction tasks. This structure provides flexibility to process the data via task interface or as a python script. In addition to the data reduction tasks, many post-processing tools are available for even more flexibility and special purpose reduction needs.

[ascl:1912.002]
casacore: Suite of C++ libraries for radio astronomy data processing

The casacore package contains the core libraries of the old AIPS++/CASA (ascl:1107.013) package. This split was made to get a better separation of core libraries and applications. CASA is now built on top of Casacore. The system consists of a set of layered libraries (packages) and includes a library (using Boost-Python) that converts the basic Casacore types (e.g., Array, Record) to and from Python. Casacore includes the casa package for core functionality and data types like Array and Record; a scimath package for N-dim functions with auto-differentiation and linear or non-linear fitting; and a tables package for the table data system supporting N-dim arrays with advanced querying. It also includes the measures package to manage values in astronomical reference frames using physical units (Quanta) and the MeasurementSets for storing data in the UV-domain, and also the images package for N-dim images in world coordinates with various analysis operations.

[ascl:1905.023]
CASI-2D: Convolutional Approach to Shell Identification - 2D

CASI-2D (Convolutional Approach to Shell Identification) identifies stellar feedback signatures using data from magneto-hydrodynamic simulations of turbulent molecular clouds with embedded stellar sources and deep learning techniques. Specifically, a deep neural network is applied to dense regression and segmentation on simulated density and synthetic 12 CO observations to identify shells, sometimes referred to as "bubbles," and other structures of interest in molecular cloud data.

[ascl:2009.005]
CASI-3D: Convolutional Approach to Structure Identification-3D

CASI-3D identifies signatures of stellar feedback in molecular line spectra, such as 12CO and 13CO, using deep learning. The code is developed from CASI-2D (ascl:1905.023) and exploits the full 3D spectral information.

[ascl:1402.013]
CASSIS: Interactive spectrum analysis

CASSIS (Centre d'Analyse Scientifique de Spectres Infrarouges et Submillimetriques), written in Java, is suited for broad-band spectral surveys to speed up the scientific analysis of high spectral resolution observations. It uses a local spectroscopic database made of the two molecular spectroscopic databases JPL and CDMS, as well as the atomic spectroscopic database NIST. Its tools include a LTE model and the RADEX (ascl:1010.075) model connected to the LAMDA (ascl:1010.077) molecular collisional database. CASSIS can build a line list fitting the various transitions of a given species and to directly produce rotational diagrams from these lists. CASSIS is fully integrated into HIPE (ascl:1111.001), the Herschel Interactive Processing Environment, as a plug-in.

[ascl:1105.010]
CASTRO: Multi-dimensional Eulerian AMR Radiation-hydrodynamics Code

Almgren, A. S.; Beckner, V. E.; Bell, J. B.; Day, M. S.; Howell, L. H.; Katz, M; Lijewski, M. J.; Malone, C.; Nonaka, A.; Singer, M.; Zhang, W; Zingale, M.

CASTRO is a multi-dimensional Eulerian AMR radiation-hydrodynamics code that includes stellar equations of state, nuclear reaction networks, and self-gravity. Initial target applications for CASTRO include Type Ia and Type II supernovae. CASTRO supports calculations in 1-d, 2-d and 3-d Cartesian coordinates, as well as 1-d spherical and 2-d cylindrical (r-z) coordinate systems. Time integration of the hydrodynamics equations is based on an unsplit version of the piecewise parabolic method (PPM) with new limiters that avoid reducing the accuracy of the scheme at smooth extrema. CASTRO can follow an arbitrary number of isotopes or elements. The atomic weights and amounts of these elements are used to calculate the mean molecular weight of the gas required by the equation of state. CASTRO supports several different approaches to solving for self-gravity. The most general is a full Poisson solve for the gravitational potential. CASTRO also supports a monopole approximation for gravity, and a constant gravity option is also available. The CASTRO software is written in C++ and Fortran, and is based on the BoxLib software framework developed by CCSE.

[ascl:1804.013]
CAT-PUMA: CME Arrival Time Prediction Using Machine learning Algorithms

CAT-PUMA (CME Arrival Time Prediction Using Machine learning Algorithms) quickly and accurately predicts the arrival of Coronal Mass Ejections (CMEs) of CME arrival time. The software was trained via detailed analysis of CME features and solar wind parameters using 182 previously observed geo-effective partial-/full-halo CMEs and uses algorithms of the Support Vector Machine (SVM) to make its predictions, which can be made within minutes of providing the necessary input parameters of a CME.

[ascl:2108.008]
CatBoost: High performance gradient boosting on decision trees library

CatBoost is a machine learning method based on gradient boosting over decision trees and can be used for ranking, classification, regression and other machine learning tasks for Python, R, Java, C++. It supports both numerical and categorical features and computation on CPU and GPU, and is fast and scalable. Visualization tools are also included in CatBoost.

[ascl:1206.008]
Catena: Ensemble of stars orbit integration

Catena integrates the orbits of an ensemble of stars using the chain-regularization method (Mikkola & Aarseth) with an embedded Runge-Kutta integration method of 9(8)th order (Prince & Dormand).

[ascl:2007.024]
CaTffs: Calcium triplet indexes

CaTffs predicts the strength of calcium triplet indices (CaT*, PaT and CaT) on the basis of empirical fitting functions and performs required interpolations between the different local functions. Together with the indices predictions, the program also computes the random errors associated to such predictions resulting from the covariance matrices of the fits (for the indices CaT* and PaT). This ensures a reliable error index estimation for any combination of input atmospheric parameters.

[ascl:1810.013]
catsHTM: Catalog cross-matching tool

The catsHTM package quickly accesses and cross-matches large astronomical catalogs that have been reformatted into the HDF5-based file format. It performs efficient cone searches at resolutions from a few arc-seconds to degrees within a few milliseconds time, cross-match numerous catalogs, and can do general searches.

[ascl:2108.007]
catwoman: Transit modeling Python package for asymmetric light curves

catwoman models asymmetric transit lightcurves. Written in Python, it calculates light curves for any radially symmetric stellar limb darkening law, and where planets are modeled as two semi-circles of different radii. Catwoman is built on the batman library (ascl:1510.002) and uses its integration algorithm.

[submitted]
Caustic Mass Estimator for Galaxy Clusters

The caustic technique is a powerful method to infer cluster mass profiles to clustrocentric distances well beyond the virial radius. It relies in the measure of the escape velocity of the sistem using only galaxy redshift information. This method was introduced by Diaferio & Geller (1997) and Diaferio (1999). This code allows the caustic mass estimation for galaxy clusters, as well as outlier identification as a side effect. However, a pre-cleaning of interlopers is recommended, using e.g., the shifting-gapper technique.

[ascl:1904.012]
CausticFrog: 1D Lagrangian Simulation Package

CausticFrog models the reaction of a system of orbiting particles to instantaneous mass loss. It applies to any spherically symmetric potential, and follows the radial evolution of shells of mass. CausticFrog tracks the inner and outer edge of each shell, whose radius evolves as a test particle. The amount of mass in each shell is fixed but multiple shells can overlap leading to higher densities.

[ascl:2404.001]
cbeam: Coupled-mode propagator for slowly-varying waveguides

cbeam models the propagation of guided light through slowly-varying few-mode waveguides using the coupled-mode theory (CMT). When compared with more general numerical methods for waveguide simulation, such as the finite-differences beam propagation method (FD-BPM), numerical implementations of the CMT can be much more computationally efficient. Written in Python and Julia, the package provides a Pythonic class structure to define waveguides, with simple classes for directional couplers and photonic lanterns already provided. cbeam also doubles as a finite-element eigenmode solver.

[ascl:2402.004]
CCBH-Numerics: Cosmologically-coupled-black-holes formation mass numerics

CCBH-Numerics (previously called CCBH-PLPP) computes the probability of the existence of a single cosmologically coupled black hole (BH) with a formation mass below a specified threshold for given observational data of binary black holes (BBHs) from gravitational waves. The code uses the unbiased population of BBHs, as given by the power-law-plus-peak (PLPP) profile, as the observational input, and assumes that the detected BBHs are formed from stellar evolution, not primordial BHs. CCBH-Numerics also works with individual data from BBHs and for NSBH pairs as well.

[ascl:2206.020]
CCDLAB: FITS image viewer and data reducer

CCDLAB provides graphical user interface functionality for FITS image viewing and data reduction based on the JPFITS FITS-file interface. It can view, manipulate, and save FITS primary image data and image extensions, view and manipulate FITS image headers, and view FITS Bintable extensions. The code enables batch processing, viewing, and saving of FITS images and searching FITS files on disk. CCDLAB also provides general image reduction techniques, source detection and characterization, and can create World Coordinate Solutions automatically or manually for FITS images.

[ascl:1403.021]
CCDPACK: CCD Data Reduction Package

CCDPACK contains programs to debias, remove dark current, flatfield, register, resample and normalize data from single- or multiple-CCD instruments. The basic reduction stages can be set up using an X based GUI that controls an automated reduction system so one can to start working without any detailed knowledge of the package (or indeed of CCD reduction). Registration is performed using graphical, script based or automated techniques that keep the amount of work to a minimum. CCDPACK uses the Starlink environment (ascl:1110.012).

[ascl:1510.007]
ccdproc: CCD data reduction software

Craig, M. W.; Crawford, S. M.; Deil, Christoph; Gomez, Carlos; Günther, Hans Moritz; Heidt, Nathan; Horton, Anthony; Karr, Jennifer; Nelson, Stefan; Ninan, Joe Phillip; Pattnaik, Punyaslok; Rol, Evert; Schoenell, William; Seifert, Michael; Singh, Sourav; Sipocz, Brigitta; Stotts, Connor; Streicher, Ole; Tollerud, Erik; Walker, Nathan; ccdproc contributors

Ccdproc is an affiliated package for the AstroPy package for basic data reductions of CCD images. The ccdproc package provides many of the necessary tools for processing of ccd images built on a framework to provide error propagation and bad pixel tracking throughout the reduction process.

[ascl:1511.013]
CCDtoRGB: RGB image production from three-band astronomical images

Lupton, Robert; Blanton, Michael R.; Fekete, George; Hogg, David W.; O'Mullane, Wil; Szalay, Alex; Wherry, Nicholas

CCDtoRGB produces red‐green‐blue (RGB) composites from three‐band astronomical images, ensuring an object with a specified astronomical color has a unique color in the RGB image rather than burnt‐out white stars. Use of an arcsinh stretch shows faint objects while simultaneously preserving the structure of brighter objects in the field, such as the spiral arms of large galaxies.

[ascl:1707.004]
CCFpams: Atmospheric stellar parameters from cross-correlation functions

CCFpams allows the measurement of stellar temperature, metallicity and gravity within a few seconds and in a completely automated fashion. Rather than performing comparisons with spectral libraries, the technique is based on the determination of several cross-correlation functions (CCFs) obtained by including spectral features with different sensitivity to the photospheric parameters. Literature stellar parameters of high signal-to-noise (SNR) and high-resolution HARPS spectra of FGK Main Sequence stars are used to calibrate the stellar parameters as a function of CCF areas.

[ascl:1901.003]
CCL: Core Cosmology Library

Chisari, Nora Elisa; Alonso, David; Krause, Elisabeth; Leonard, C. Daniellle; Bull, Philip; Neveu, Jérémy; Villarreal, Antonia Sierra; Singh, Sukhdeep; McClintock, Thomas; Ellison, John; Du, Zilong; Zuntz, Joe; Mead, Alexander; Joudaki, Shahab; Lorenz, Christiane S.; Troester, Tilman; Sanchez, Javier; Lanusse, Francois; Ishak, Mustapha; Hlozek, Renée; Blazek, Jonathan; Campagne, Jean-Eric; Almoubayyed, Husni; Eifler, Tim; Kirby, Matthew; Kirkby, David; Plaszczynski, Stéphane; Slosar, Anze; Vrastil, Michal; Wagoner, Erika L.

The Core Cosmology Library (CCL) computes basic cosmological observables and provides predictions for many cosmological quantities, including distances, angular power spectra, correlation functions, halo bias and the halo mass function through state-of-the-art modeling prescriptions. Fiducial specifications for the expected galaxy distributions for the Large Synoptic Survey Telescope (LSST) are also included, together with the capability of computing redshift distributions for a user-defined photometric redshift model. Predictions for correlation functions of galaxy clustering, galaxy-galaxy lensing and cosmic shear are within a fraction of the expected statistical uncertainty of the observables for the models and in the range of scales of interest to LSST. CCL is written in C and has a python interface.

[ascl:1208.006]
ccogs: Cosmological Calculations on the GPU

This suite contains two packages for computing cosmological quantities on the GPU: aperture_mass, which calculates the aperture mass map for a given dataset using the filter proposed by Schirmer et al (2007) (an NFW profile with exponential cut-offs at zero and large radii), and angular_correlation, which calculates the 2-pt angular correlation function using data and a flat distribution of randomly generated galaxies. A particular estimator is chosen, but the user has the flexibility to explore other estimators.

[ascl:1604.009]
CCSNMultivar: Core-Collapse Supernova Gravitational Waves

CCSNMultivar aids the analysis of core-collapse supernova gravitational waves. It includes multivariate regression of Fourier transformed or time domain waveforms, hypothesis testing for measuring the influence of physical parameters, and the Abdikamalov et. al. catalog for example use. CCSNMultivar can optionally incorporate additional uncertainty due to detector noise and approximate waveforms from anywhere within the parameter space.

[ascl:1904.006]
CDAWeb: Coordinated Data Analysis Web

CDAWeb (Coordinated Data Analysis Workshop Web) enables viewing essentially any data produced in Common Data Format/CDF with the ISTP/IACG Guidelines and supports interactive plotting of variables from multiple instruments on multiple investigations simultaneously on arbitrary, user-defined time-scales. It also supports data retrieval in both CDF or ASCII format. NASA's GSFC Space Physics Data Facility maintains a publicly available database that includes approximately 600 data variables from Geotail, Wind, Interball, Polar, SOHO, ancilliary spacecraft and ground-based investigations. CDAWeb includes high resolution digital data products that support event correlative science. The system combines the client-server user interface technology of the Web with a powerful set of customized routines based in the COTS Interactive Data Language (IDL) package to leverage the data format standards.

[ascl:2005.017]
cdetools: Tools for Conditional Density Estimates

cdetools provides tools for evaluating conditional density estimates and has applications to photometric redshift estimation and likelihood-free cosmological inference. Available in R and Python, it provides functions for computing a so-called CDE loss function for tuning and assessing the quality of individual probability density functions (PDFs) and diagnostic functions that probe the population-level performance of the PDFs.

[ascl:2305.025]
CELEBI: Precision localizations and polarimetric data for fast radio bursts

Scott, D. R.; Cho, H.; Day, C. K.; Deller, A. T.; Glowacki, M.; Gourdji, K.; Bannister, K. W.; Bera, A.; Bhandari, S.; James, C. W.; Shannon, R. M.

The Australian Square Kilometre Array Pathfinder (ASKAP) has been enabled by the Commensal Real-time ASKAP Fast Transients Collaboration (CRAFT) to detect Fast Radio Bursts (FRBs) in real-time and save raw antenna voltages containing FRB detections. CELEBI, the CRAFT Effortless Localization and Enhanced Burst Inspection pipeline, extends CRAFT’s existing software to process ASKAP voltages to produce sub-arcsecond precision localizations and polarimetric data at time resolutions as fine as 3 ns of FRB events. CELEBI uses Nextflow (ascl:2305.024) to link together Bash and Python code to perform software correlation, interferometric imaging, and beamforming, thereby making use of common astronomical software packages.

[ascl:1709.008]
celerite: Scalable 1D Gaussian Processes in C++, Python, and Julia

celerite provides fast and scalable Gaussian Process (GP) Regression in one dimension and is implemented in C++, Python, and Julia. The celerite API is designed to be familiar to users of george and, like george, celerite is designed to efficiently evaluate the marginalized likelihood of a dataset under a GP model. This is then be used alongside a non-linear optimization or posterior inference library for the best results.

celerite has been superceded by celerite2 (ascl:2310.001).

[ascl:2310.001]
celerite2: Fast and scalable Gaussian Processes in one dimension

celerite2 is a re-write of celerite (ascl:1709.008), an algorithm for fast and scalable Gaussian Process (GP) Regression in one dimension. celerite2 improves numerical stability and integration with various machine learning frameworks. The implementation includes interfaces in Python and C++, with full support for PyMC (ascl:1610.016) and JAX (ascl:2111.002).

[ascl:1602.011]
Celestial: Common astronomical conversion routines and functions

The R package Celestial contains common astronomy conversion routines, particularly the HMS and degrees schemes, and a large range of functions for calculating properties of different cosmologies (as used by the cosmocalc website). This includes distances, ages, growth rate/factor and densities (e.g., Omega evolution and critical energy density). It also includes functions for calculating thermal properties of the CMB and Planck's equations and virial properties of halos in different cosmologies, and standard NFW and weak-lensing formulas and low level orbital routines for calculating Roche properties, Vis-Viva and free-fall times.

[ascl:1612.016]
CELib: Software library for simulations of chemical evolution

CELib (Chemical Evolution Library) simulates chemical evolution of galaxy formation under the simple stellar population (SSP) approximation and can be used by any simulation code that uses the SSP approximation, such as particle-base and mesh codes as well as semi-analytical models. Initial mass functions, stellar lifetimes, yields from type II and Ia supernovae, asymptotic giant branch stars, and neutron star mergers components are included and a variety of models are available for use. The library allows comparisons of the impact of individual models on the chemical evolution of galaxies by changing control flags and parameters of the library.

[ascl:2302.005]
celmech: Sandbox for celestial mechanics calculations

celmech provides a variety of analytical and semianalytical tools for celestial mechanics and dynamical astronomy. The package interfaces closely with the REBOUND N-body integrator (ascl:1110.016), thus facilitating comparisons between calculation results and direct N-body integrations. celmech can isolate the contribution of particular resonances to a system's dynamical evolution, and can develop simple analytical models with the minimum number of terms required to capture a particular dynamical phenomenon.

[ascl:1906.021]
centerRadon: Center determination code in stellar images

centerRadon finds the center of stars based on Radon Transform to sub-pixel precision. For a coronagraphic image of a star, it starts from a given location, then for each sub-pixel position, it interpolates the image and sums the pixels along different angles, creating a cost function. The center of the star is expected to correspond with where the cost function maximizes. The default values are set for the STIS coronagraphic images of the Hubble Space Telescope by summing over the diagonals (i.e., 45° and 135°), but it can be generally applied to other high-contrast imaging instruments with or without Adaptive Optics systems such as HST-NICMOS, P1640, or GPI.

[ascl:1308.015]
Ceph_code: Cepheid light-curves fitting

Ceph_code fits multi-band Cepheid light-curves using templates derived from OGLE observations. The templates include short period stars (<10 day) and overtone stars.

[ascl:1610.002]
CERES: Collection of Extraction Routines for Echelle Spectra

The Collection of Extraction Routines for Echelle Spectra (CERES) constructs automated pipelines for the reduction, extraction, and analysis of echelle spectrograph data. This modular code includes tools for handling the different steps of the processing: CCD reductions, tracing of the echelle orders, optimal and simple extraction, computation of the wave-length solution, estimation of radial velocities, and rough and fast estimation of the atmospheric parameters. The standard output of pipelines constructed with CERES is a FITS cube with the optimally extracted, wavelength calibrated and instrumental drift-corrected spectrum for each of the science images. Additionally, CERES includes routines for the computation of precise radial velocities and bisector spans via the cross-correlation method, and an automated algorithm to obtain an estimate of the atmospheric parameters of the observed star.

[ascl:1010.059]
CESAM: A Free Code for Stellar Evolution Calculations

The Cesam code is a consistent set of programs and routines which perform calculations of 1D quasi-hydrostatic stellar evolution including microscopic diffusion of chemical species and diffusion of angular momentum. The solution of the quasi-static equilibrium is performed by a collocation method based on piecewise polynomials approximations projected on a B-spline basis; that allows stable and robust calculations, and the exact restitution of the solution, not only at grid points, even for the discontinuous variables. Other advantages are the monitoring by only one parameter of the accuracy and its improvement by super-convergence. An automatic mesh refinement has been designed for adjusting the localisations of grid points according to the changes of unknowns. For standard models, the evolution of the chemical composition is solved by stiffly stable schemes of orders up to four; in the convection zones mixing and evolution of chemical are simultaneous. The solution of the diffusion equation employs the Galerkin finite elements scheme; the mixing of chemicals is then performed by a strong turbulent diffusion. A precise restoration of the atmosphere is allowed for.

[ascl:2111.005]
CEvNS: Calculate Coherent Elastic Neutrino-Nucleus Scattering cross sections and recoil spectra

CEvNS calculates Coherent Elastic Neutrino-Nucleus Scattering (CEvNS) cross sections and recoil spectra. It includes (among other things) the Standard Model contribution to the CEvNS cross section, along with the contribution from Simplified Models with new vector or scalar mediators. It also covers neutrino magnetic moments and non-standard contact neutrino interactions (NSI).

[ascl:1901.001]
cFE: Core Flight Executive

The Core Flight Executive is a portable, platform-independent embedded system framework that is the basis for flight software for satellite data systems and instruments; cFE can be used on other embedded systems as well. The Core Flight Executive is written in C and depends on the software library Operating System Abstraction Layer (OSAL), which is available at https://sourceforge.net/projects/osal/.

[ascl:1010.001]
CFITSIO: A FITS File Subroutine Library

CFITSIO is a library of C and Fortran subroutines for reading and writing data files in FITS (Flexible Image Transport System) data format. CFITSIO provides simple high-level routines for reading and writing FITS files that insulate the programmer from the internal complexities of the FITS format. CFITSIO also provides many advanced features for manipulating and filtering the information in FITS files.

[ascl:2101.009]
cFS: core Flight System

cFS is a platform and project independent reusable software framework and set of reusable applications developed by NASA Goddard Space Flight Center. There are three key aspects to the cFS architecture: a dynamic run-time environment, layered software, and a component based design, making it suitable for reuse on NASA flight projects and/or embedded software systems. This framework is used as the basis for the flight software for satellite data systems and instruments, but can also be used on other embedded systems. Modules of this package are used in NICER (Neutron star Interior Composition Explorer). The modules are available as separate downloads from SourceForge through the NASA cFS website.

[ascl:1904.003]
CGS: Collisionless Galactic Simulator

CGS (Collisionless Galactic Simulator) uses Fourier techniques to solve the Possion equation ∇^{2}Φ = 4πGρ, relating the mean potential Φ of a system to the mass density ρ. The angular dependence of the force is treated exactly in terms of the single-particle Legendre polynomials, which preserves accuracy and avoids systematic errors. The density is assigned to a radial grid by means of a cloud-in-cell scheme with a linear kernel, *i.e.*, a particle contributes to the density of the two closest cells with a weight depending linearly on the distance from the center of the cell considered. The same kernel is then used to assign the force from the grid to the particle. The time step is chosen adaptively in such a way that particles are not allowed to cross more than one radial cell during one step. CGS is based on van Albada's code (1982) and is distributed in the NEMO (ascl:1010.051) Stellar Dynamics Toolbox.

[ascl:1411.024]
CGS3DR: UKIRT CGS3 data reduction software

CGS3DR is data reduction software for the UKIRT CGS3 mid-infrared grating spectrometer instrument. It includes a command-line interface and a GUI. The software, originally on VMS, was ported to Unix. It uses Starlink (ascl:1110.012) infrastructure libraries.

[ascl:1406.013]
CGS4DR: Automated reduction of data from CGS4

CGS4DR is data reduction software for the CGS4 instrument at UKIRT. The software can be used offline to reprocess CGS4 data. CGS4DR allows a wide variety of data reduction configurations, and can interlace oversampled data frames; reduce known bias, dark, flat, arc, object and sky frames; remove the sky, residual sky OH-lines (λ < 2.3 μm) and thermal emission (λ ≥ 2.3 μm) from data; and add data into groups for improved signal-to-noise. It can also extract and de-ripple a spectrum and offers a variety of ways to plot data, in addition to other useful features. CGS4DR is distributed as part of the Starlink software collection (ascl:1110.012).

[ascl:1910.017]
ChainConsumer: Corner plots, LaTeX tables and plotting walks

ChainConsumer consumes the chains output from Monte Carlo processes such as MCMC to produce plots of the posterior surface inferred from the chain distributions, to plot the chains as walks to check for mixing and convergence, and to output parameter summaries in the form of LaTeX tables. It handles multiple models (chains), allowing for model comparison using AIC, BIC or DIC metrics.

[ascl:1105.005]
ChaNGa: Charm N-body GrAvity solver

ChaNGa (Charm N-body GrAvity solver) performs collisionless N-body simulations. It can perform cosmological simulations with periodic boundary conditions in comoving coordinates or simulations of isolated stellar systems. It also can include hydrodynamics using the Smooth Particle Hydrodynamics (SPH) technique. It uses a Barnes-Hut tree to calculate gravity, with hexadecapole expansion of nodes and Ewald summation for periodic forces. Timestepping is done with a leapfrog integrator with individual timesteps for each particle.

[ascl:1703.015]
Charm: Cosmic history agnostic reconstruction method

Charm (cosmic history agnostic reconstruction method) reconstructs the cosmic expansion history in the framework of Information Field Theory. The reconstruction is performed via the iterative Wiener filter from an agnostic or from an informative prior. The charm code allows one to test the compatibility of several different data sets with the LambdaCDM model in a non-parametric way.

[ascl:2309.017]
ChEAP: Chemical Evolution Analytic Package

ChEAP (Chemical Evolution Analytic Package) implements an analytic solution for the chemical evolution model of the Galaxy that extends the instantaneous recycling approximation with the contribution of Type Ia SNe. The code works for different prescriptions of the delay time distributions (DTDs), including the single and double degenerate scenarios, and allows the inclusion of an arbitrary number of pristine gas infalls. The required functions are contained in the CheapTools.py file, which is imported as a Python library. ChEAP also includes code to illustrate, with a random-parameter chemical evolution model, the accuracy of this analytic solution compared to one using numerical integration.

[ascl:1412.002]
Cheetah: Starspot modeling code

Cheetah models starspots in photometric data (lightcurves) by calculating the modulation of a light curve due to starspots. The main parameters of the program are the linear and quadratic limb darkening coefficients, stellar inclination, spot locations and sizes, and the intensity ratio of the spots to the stellar photosphere. Cheetah uses uniform spot contrast and the minimum number of spots needed to produce a good fit and ignores bright regions for the sake of simplicity.

[ascl:2107.020]
Chem-I-Calc: Chemical Information Calculator

Chem-I-Calc evaluates the chemical information content of resolved star spectroscopy. It takes advantage of the Fisher information matrix and the Cramér-Rao inequality to quickly calculate the Cramér-Rao lower bounds (CRLBs), which give the best theoretically achievable precision from a set of observations.

[ascl:1702.011]
Chempy: A flexible chemical evolution model for abundance fitting

Chempy models Galactic chemical evolution (GCE); it is a parametrized open one-zone model within a Bayesian framework. A Chempy model is specified by a set of 5-10 parameters that describe the effective galaxy evolution along with the stellar and star-formation physics: e.g. the star-formation history (SFH), the feedback efficiency, the stellar initial mass function (IMF) and the incidence of supernova of type Ia (SN Ia). Chempy can sample the posterior probability distribution in the full model parameter space and test data-model matches for different nucleosynthetic yield sets, performing essentially as a chemical evolution fitting tool. Chempy can be used to confront predictions from stellar nucleosynthesis with complex abundance data sets and to refine the physical processes governing the chemical evolution of stellar systems.

ChempyMulti (ascl:1909.006) is available as an update to the ChempyScoring package.

[ascl:1909.006]
ChempyMulti: Multi-star Bayesian inference with Chempy

Rybizki, Jan; Philcox, Oliver H. E.; Rix, Hans-Walter; Just, Andreas; Fouesneau, Morgan; Sandford, Nathan

ChempyMulti is an update to Chempy (ascl:1702.011) and provides yield table scoring and multi-star Bayesian inference. This replaces the ChempyScoring package in Chempy. Chempy is a flexible one-zone open-box chemical evolution model, incorporating abundance fitting and stellar feedback calculations. It includes routines for parameter optimization for simulations and observational data and yield table scoring.

[ascl:2108.016]
Chemulator: Thermochemical emulator for hydrodynamical modeling

The neural network-based emulator Chemulator advances the gas temperature and chemical abundances of a single position in an astrophysical gas. It is accurate on a single timestep and stable over many iterations with decreased accuracy, though performs less well at low visual extinctions. The code is useful for applications such as large scale ISM modeling; by retraining the emulator for a given parameter space, Chemulator could also perform more specialized applications such as planetary atmosphere modeling.

[ascl:9911.004]
CHIANTI: A database for astrophysical emission line spectroscopy

CHIANTI consists of a critically evaluated set of atomic data necessary to calculate the emission line spectrum of astrophysical plasmas. The data consists of atomic energy levels, atomic radiative data such as wavelengths, weighted oscillator strengths and A values, and electron collisional excitation rates. A set of programs that use these data to calculate the spectrum in a desired wavelength range as a function of temperature and density are also provided. These programs have been written in Interactive Data Language (IDL) and descriptions of these various programs are provided on the website.

[ascl:1308.017]
ChiantiPy: Python package for the CHIANTI atomic database

ChiantiPy is an object-orient Python package for calculating astrophysical spectra using the CHIANTI atomic database for astrophysical spectroscopy. It provides access to the database and the ability to calculate various physical quantities for the interpretation of astrophysical spectra.

[ascl:1504.005]
chimenea: Multi-epoch radio-synthesis data imaging

Chimenea implements an heuristic algorithm for automated imaging of multi-epoch radio-synthesis data. It generates a deep image via an iterative Clean subroutine performed on the concatenated visibility set and locates steady sources in the field of view. The code then uses this information to apply constrained and then unconstrained (*i.e.*, masked/open-box) Cleans to the single-epoch observations. This obtains better results than if the single-epoch data had been processed independently without prior knowledge of the sky-model. The chimenea pipeline is built upon CASA (ascl:1107.013) subroutines, interacting with the CASA environment via the drive-casa (ascl:1504.006) interface layer.

[ascl:1403.006]
CHIMERA: Core-collapse supernovae simulation code

Mezzacappa, Anthony; Hix, Raph; Messor, Bronson; Lentz, Eric; Chertkow, Merek Austin; Parete-Koon, Suzanne; Lingerfelt, Eric

CHIMERA simulates core collapse supernovas; it is three-dimensional and accounts for the differing energies of neutrinos. This massively parallel multiphysics code conserves total energy (gravitational, internal, kinetic, and neutrino) to within 0.5 B, given a conservative gravitational potential. CHIMERA has three main components: a hydro component, a neutrino transport component, and a nuclear reaction network component. It also includes a Poisson solver for the gravitational potential and a sophisticated equation of state.

[ascl:1602.017]
CHIP: Caltech High-res IRS Pipeline

CHIP (Caltech High-res IRS Pipeline) reduces high signal-to-noise short-high and long-high Spitzer-IRS spectra, especially that taken with dedicated background exposures. Written in IDL, it is independent of other Spitzer reduction tools except IRSFRINGE (ascl:1602.016).

[ascl:2306.046]
CHIPS: Circumstellar matter and light curves of interaction-powered transients simulator

CHIPS (Complete History of Interaction-Powered Supernovae) simulates the circumstellar matter and light curves of interaction-powered transients. Coupled with MESA (ascl:1010.083), the combined codes can obtain the circumstellar matter profile and light curves of the interaction-powered supernovae. CHIPS generates a realistic CSM from a model-agnostic mass eruption calculation, which can serve as a reference for observers to compare with various observations of the CSM. The code can also generate bolometric light curves from CSM interaction, which can be compared with observed light curves. The calculation of mass eruption and light curve typically takes respectively half a day and half an hour on modern CPUs.

[ascl:1104.012]
CHIWEI: A Code of Goodness of Fit Tests for Weighted and Unweighed Histograms

A self-contained Fortran-77 program for goodness of fit tests for histograms with weighted entries as well as with unweighted entries is presented. The code calculates test statistic for case of histogram with normalized weights of events and for case of unnormalized weights of events.

[ascl:1409.008]
CHLOE: A tool for automatic detection of peculiar galaxies

CHLOE is an image analysis unsupervised learning algorithm that detects peculiar galaxies in datasets of galaxy images. The algorithm first computes a large set of numerical descriptors reflecting different aspects of the visual content, and then weighs them based on the standard deviation of the values computed from the galaxy images. The weighted Euclidean distance of each galaxy image from the median is measured, and the peculiarity of each galaxy is determined based on that distance.

[ascl:1607.006]
Cholla: 3D GPU-based hydrodynamics code for astrophysical simulation

Cholla (Computational Hydrodynamics On ParaLLel Architectures) models the Euler equations on a static mesh and evolves the fluid properties of thousands of cells simultaneously using GPUs. It can update over ten million cells per GPU-second while using an exact Riemann solver and PPM reconstruction, allowing computation of astrophysical simulations with physically interesting grid resolutions (>256^3) on a single device; calculations can be extended onto multiple devices with nearly ideal scaling beyond 64 GPUs.

[ascl:1202.008]
Chombo: Adaptive Solutions of Partial Differential Equations

Chombo provides a set of tools for implementing finite difference methods for the solution of partial differential equations on block-structured adaptively refined rectangular grids. Both elliptic and time-dependent modules are included. Chombo supports calculations in complex geometries with both embedded boundaries and mapped grids, and also supports particle methods. Most parallel platforms are supported, and cross-platform self-describing file formats are included.

The Chombo package is a product of the community of Collaborators working with the Applied Numerical Algorithms Group (ANAG), part of the Computational Research Division at LBNL.

[ascl:1209.004]
CHORIZOS: CHi-square cOde for parameterRized modeling and characterIZation of phOtometry and Spectrophotmetry

CHORIZOS is a multi-purpose Bayesian code developed in IDL to compare photometric data with model spectral energy distributions (SEDs). The user can select the SED family (e.g. Kurucz) and choose the behavior of each parameter (e.g. Teff) to be fixed, constrained to a given range, or unconstrained. The code calculates the likelihood for the full specified parameter ranges, thus allowing for the identification of multiple solutions and the evaluation of the full correlation matrix for the derived parameters of a single solution.

[ascl:1011.018]
chrom_exact: Transit of a Spherical Planet of a Stellar Chromosphere which is Geometrically Thin

Transit light curves for stellar continua have only one minimum and a "U" shape. By contrast, transit curves for optically thin chromospheric emission lines can have a "W" shape because of stellar limb-brightening. We calculate light curves for an optically thin shell of emission and fit these models to time-resolved observations of Si IV absorption by the planet HD209458b. We find that the best fit Si IV absorption model has R_p,SIV/R_*= 0.34+0.07-0.12, similar to the Roche lobe of the planet. While the large radius is only at the limit of statistical significance, we develop formulae applicable to transits of all optically thin chromospheric emission lines.

[ascl:1804.007]
chroma: Chromatic effects for LSST weak lensing

Chroma investigates biases originating from two chromatic effects in the atmosphere: differential chromatic refraction (DCR), and wavelength dependence of seeing. These biases arise when using the point spread function (PSF) measured with stars to estimate the shapes of galaxies with different spectral energy distributions (SEDs) than the stars.

[ascl:1701.008]
ChromaStar (formerly GrayStar): Web-based pedagogical stellar modeling

ChromaStar (formerly GrayStar) is a web-based pedagogical stellar model. It approximates stellar atmospheric and spectral line modeling in JavaScript with visualization in HTML. It is suitable for a wide range of education and public outreach levels depending on which optional plots and print-outs are turned on. All plots and renderings are pure basic HTML and the plotting module contains original HTML procedures for automatically scaling and graduating x- and y-axes.

[ascl:1701.009]
ChromaStarServer (formerly GrayStarServer): Stellar atmospheric modeling and spectrum synthesis

ChromaStarServer (formerly GrayStarServer) is a stellar atmospheric modeling and spectrum synthesis code of pedagogical accuracy that is accessible in any web browser on commonplace computational devices and that runs on a timescale of a few seconds.

[ascl:2009.021]
Chrono: Multi-physics simulation engine

Tasora, Alessandro; Serban, Radu; Mazhar, Hammad; Pazouki, Arman; Melanz, Daniel; Fleischmann, Jonathan; Taylor, Michael; Sugiyama, Hiroyuki; Negrut, Dan

Chrono is a physics-based modelling and simulation infrastructure implemented in C++. It can handle multibody dynamics, collision detection, and granular flows, among many other physical processes. Though the applications for which Chrono has been used most often are vehicle dynamics, robotics, and machine design, it has been used to simulate asteroid aggregation and granular systems for astrophysics research. Chrono is written in C++; a Python version, PyChrono, is also available.

[ascl:1311.006]
CIAO: Chandra Interactive Analysis of Observations

CIAO is a data analysis system written for the needs of users of the Chandra X-ray Observatory. Because Chandra data is 4-dimensional (2 spatial, time, energy) and each dimension has many independent elements, CIAO was built to handle N-dimensional data without concern about which particular axes were being analyzed. Apart from a few Chandra instrument tools, CIAO is mission independent. CIAO tools read and write several formats, including FITS images and tables (which includes event files) and IRAF imh files. CIAO is a powerful system for the analysis of many types of data.

[ascl:1803.002]
CIFOG: Cosmological Ionization Fields frOm Galaxies

CIFOG is a versatile MPI-parallelised semi-numerical tool to perform simulations of the Epoch of Reionization. From a set of evolving cosmological gas density and ionizing emissivity fields, it computes the time and spatially dependent ionization of neutral hydrogen (HI), neutral (HeI) and singly ionized helium (HeII) in the intergalactic medium (IGM). The code accounts for HII, HeII, HeIII recombinations, and provides different descriptions for the photoionization rate that are used to calculate the residual HI fraction in ionized regions. This tool has been designed to be coupled to semi-analytic galaxy formation models or hydrodynamical simulations. The modular fashion of the code allows the user to easily introduce new descriptions for recombinations and the photoionization rate.

[ascl:1111.004]
CIGALE: Code Investigating GALaxy Emission

The CIGALE code has been developed to study the evolution of galaxies by comparing modelled galaxy spectral energy distributions (SEDs) to observed ones from the far ultraviolet to the far infrared. It extends the SED fitting algorithm written by Burgarella et al. (2005, MNRAS 360, 1411). While the previous code was designed to fit SEDs in the optical and near infrared, CIGALE is able to fit SEDs up to the far infrared using Dale & Helou (2002, ApJ 576, 159). CIGALE Bayesian and CIGALE Monte Carlo Markov Chain are available.

[ascl:1708.002]
CINE: Comet INfrared Excitation

CINE calculates infrared pumping efficiencies that can be applied to the most common molecules found in cometary comae such as water, hydrogen cyanide or methanol. One of the main mechanisms for molecular excitation in comets is the fluorescence by the solar radiation followed by radiative decay to the ground vibrational state. This command-line tool calculates the effective pumping rates for rotational levels in the ground vibrational state scaled by the heliocentric distance of the comet. Fluorescence coefficients are useful for modeling rotational emission lines observed in cometary spectra at sub-millimeter wavelengths. Combined with computational methods to solve the radiative transfer equations based, e.g., on the Monte Carlo algorithm, this model can retrieve production rates and rotational temperatures from the observed emission spectrum.

[ascl:2206.007]
CircleCraters: Crater-counting plugin for QGIS

CircleCraters is a projection independent crater counting plugin for QGIS. It has the flexibility to crater count in a GIS environment on Windows, OS X, or Linux, and uses three-click input to define crater rims as a circle.

[ascl:1202.001]
CISM_DX: Visualization and analysis tool

CISM_DX is a community-developed suite of integrated data, models, and data and model explorers, for research and education. The data and model explorers are based on code written for OpenDX and Octave; OpenDX provides the visualization infrastructures as well as the process for creating user interfaces to the model and data, and Octave allows for extensive data manipulation and reduction operations. The CISM-DX package extends the capabilities of the core software programs to meet the needs of space physics researchers.

[ascl:2202.014]
Citlalicue: Create and manipulate stellar light curves

Citlalicue allows you to create synthetic stellar light curves (transits, stellar variability and white noise) and detrend light curves using Gaussian Processes (GPs). Transits are implemented using PyTransit (ascl:1505.024). Python notebooks are provided to demonstrate using Citlalicue for both functions.

[ascl:1312.013]
CJAM: First and second velocity moments calculations

CJAM calculates first and second velocity moments using the Jeans Anisotropic MGE (JAM) models of Cappellari (2008) and Cappellari (2012). These models have been extended to calculate all three (x, y, z) first moments and all six (xx, yy, zz, xy, xz, yz) second moments. CJAM, written in C, is based on the IDL implementation of the line-of-sight calculations by Michele Cappellari.

[ascl:2210.028]
CK: Cloud modeling and removal

Teinturier, Lucas; Vieira, Nicholas; Jacquet, Elisa; Geoffrion, Juliette; Bestavros, Youssef; Keating, Dylan; Cowan, Nicolas B.

Cloud Killer recovers surface albedo maps by using reflected light photometry to map the clouds and surface of unresolved exoplanets. For light curves with negligible photometric uncertainties, the minimal top-of-atmosphere albedo at a location is a good estimate of its surface albedo. On synthetic data, it shows little bias, good precision, and accuracy, but slightly underestimated uncertainties; exoplanets with large, changing cloud structures observed near quadrature phases are good candidates for Cloud Killer cloud removal.

[ascl:2105.018]
ClaRAN: Classifying Radio sources Automatically with Neural networks

Wu, Chen; Wong, Oiwei Ivy; Rudnick, Lawrence; Shabala, Stanislav S.; Alger, Matthew J.; Banfield, Julie K.; Ong, Cheng Soon; White, Sarah V.; Garon, Avery F.; Norris, Ray P.; Andernach, Heinz; Tate, Jean; Lukic, Vesna; Tang, Hongming; Schawinski, Kevin; Diakogiannis, Foivos I.

ClaRAN (Classifying Radio sources Automatically with Neural networks) classifies radio source morphology based upon the Faster Region-based Convolutional Neutral Network (Faster R-CNN). It is capable of associating discrete and extended components of radio sources in an automated fashion. ClaRAN demonstrates the feasibility of applying deep learning methods for cross-matching complex radio sources of multiple components with infrared maps. The promising results from ClaRAN have implications for the further development of efficient cross-wavelength source identification, matching, and morphology classifications for future radio surveys.

[ascl:2403.015]
CLASS-PT: Nonlinear perturbation theory extension of the Boltzmann code CLASS

CLASS-PT modifies the CLASS (ascl:1106.020) code to compute the non-linear power spectra of dark matter and biased tracers in one-loop cosmological perturbation theory, for both Gaussian and non-Gaussian initial conditions. CLASS-PT can be interfaced with the MCMC sampler MontePython (ascl:1805.027) using the (new and improved) custom-built likelihoods found here.

[ascl:1106.020]
CLASS: Cosmic Linear Anisotropy Solving System

Boltzmann codes are used extensively by several groups for constraining cosmological parameters with Cosmic Microwave Background and Large Scale Structure data. This activity is computationally expensive, since a typical project requires from 10'000 to 100'000 Boltzmann code executions. The code CLASS (Cosmic Linear Anisotropy Solving System) incorporates improved approximation schemes leading to a simultaneous gain in speed and precision. We describe here the three approximations used by CLASS for basic LambdaCDM models, namely: a baryon-photon tight-coupling approximation which can be set to first order, second order or to a compromise between the two; an ultra-relativistic fluid approximation which had not been implemented in public distributions before; and finally a radiation streaming approximation taking reionisation into account.

[ascl:1807.013]
CLASSgal: Relativistic cosmological large scale structure code

CLASSgal computes large scale structure observables; it includes all relativistic corrections and computes both the power spectrum *C _{l}*(

Would you like to view a random code?