[submitted]
hierArc - hierarchical analysis of strong gravitational lenses

The software enables to hierarchically infer strong lensing mass density profiles and the cosmological parameters, in particular the Hubble constant. The software supports lenses with imaging data and kinematics, and optionally time delays. The kinematics modeling is performed in conjunction with lenstronomy (ascl:1804.012). The software was introduced by the paper TDCOSMO IV: Hierarchical time-delay cosmography -- joint inference of the Hubble constant and galaxy density profiles (Birrer et al. 2020).

[ascl:2006.023]
deepSIP: deep learning of Supernova Ia Parameters

deepSIP (deep learning of Supernova Ia Parameters) measures the phase and light-curve shape of a Type Ia Supernova (SN Ia) from an optical spectrum. The package contains a set of three trained Convolutional Neural Networks (CNNs) for the aforementioned purposes, but tools for preprocessing spectra, modifying the neural architecture, training models, and sweeping through hyperparameters are also included.

[ascl:2006.022]
MCSED: Spectral energy distribution fitting package for galactic systems

MCSED models the optical, near-infrared and infrared spectral energy distribution (SED) of galactic systems. Its modularity and options make it flexible and able to address the varying physical properties of galaxies over cosmic time and environment and adjust to changes in understanding of stellar evolution, the details of mass loss, and the products of binary evolution through substitution or addition of new datasets or algorithms. MCSED is built to fit a galaxy’s full SED, from the far-UV to the far-IR. Among other physical processes, it can model continuum emission from stars, continuum and line-emission from ionized gas, attenuation from dust, and mid- and far-IR emission from dust and polycyclic aromatic hydrocarbons (PAHs). MCSED performs its calculations by creating a complex stellar population (CSP) out of a linear combination of simple-stellar populations (SSPs) using an efficient Markov Chain Monte Carlo algorithm. It is very quick, and takes advantage of parallel processing.

[ascl:2006.021]
FAMED: Extraction and mode identification of oscillation frequencies for solar-like pulsators

The FAMED (Fast and AutoMated pEak bagging with Diamonds) pipeline is a multi-platform parallelized software that performs and automates extraction and mode identification of oscillation frequencies for solar-like pulsators. The pipeline can be applied to a large variety of stars, ranging from hot F-type main sequence, up to stars evolving along the red giant branch, settled into the core-Helium-burning main sequence, and even evolved beyond towards the early asymptotic giant branch. FAMED is based on DIAMONDS (ascl:1410.001), a Bayesian parameter estimation and model comparison by means of the nested sampling Monte Carlo (NSMC) algorithm.

[ascl:2006.020]
GenetIC: Initial conditions generator for cosmological simulations

GenetIC generates initial conditions for cosmological simulations, especially for zoom simulations of galaxies. It provides support for "genetic modifications" of specific attributes of simulations to allow study of the impact of such modifications on the outcomes; the code can also produce constrained initial conditions.

[ascl:2006.019]
TATOO: Tidal-chronology Age TOOl

TATOO (Tidal-chronology Age TOOl) estimates the age of massive close-in planetary systems, even those subject to tidal spin-up, using the systems' observed properties: the mass of the planet and the star, stellar rotational, and planetary orbital periods. It can also be used as a classical gyrochronological tool and offers first order correction of the impact of tidal interaction on gyrochronology.

[ascl:2006.018]
Powderday: Dust radiative transfer package

Narayanan, Desika; Turk, Matthew J.; Robitaille, Thomas; Kelly, Ashley J.; Connor McClellan, B.; Sharma, Ray S.; Garg, Prerak; Abruzzo, Matthew; Choi, Ena; Conroy, Charlie; Johnson, Benjamin D.; Kimock, Benjamin; Li, Qi; Lovell, Christopher C.; Lower, Sidney; Privon, George C.; Roberts, Jonathan; Sethuram, Snigdaa; Snyder, Gregory F.; Thompson, Robert; Wise, John H.

The dust radiative transfer software Powderday interfaces with galaxy formation simulations to produce spectral energy distributions and images. The code uses fsps (ascl:1010.043) and its Python bindings python-fsps for stellar SEDs, Hyperion (ascl:1207.004) for dust radiative transfer, and works with a variety of packages, including Arepo (ascl:1909.010), Changa (ascl:1105.005), Gasoline (ascl:1710.019), and Gizmo (ascl:1410.003); threaded throughout is yt (ascl:1011.022).

[ascl:2006.017]
AstroCatR: Time series reconstruction of large-scale astronomical catalogs

AstroCatR reconstructs celestial objects' time series data for astronomical catalogs. It is a command-line program running on the Linux platform and is implemented in C and Python; AstroCatR's capabilities are based on specialized sky partitioning and MPI parallel programming. The package contains three parts: ETL (extract-transform-load) pre-processing, TS-matching calculation, and time series data retrieval. Once the user obtains the original catalogs, running ETL pre-processing generates a sky zoning file. The TS-matching module marks celestial objects, and finally, running the Query program searches celestial objects from the time series datasets which matched with the target.

[ascl:2006.016]
SPISEA: Stellar Population Interface for Stellar Evolution and Atmospheres

SPISEA (Stellar Population Interface for Stellar Evolution and Atmospheres) generates single-age, single-metallicity populations (*i.e.*, star clusters). The software (formerly called PyPopStar) provides control over different parameters, including cluster characteristics (age, metallicity, mass, distance); total extinction, differential extinction, and extinction law; stellar evolution and atmosphere models; stellar multiplicity and Initial Mass Function; and photometric filters. SPISEA can be used to create a cluster isochrone in many filters using different stellar models, generate a star cluster at any age with an unusual IMF and unresolved multiplicity, and make a spectrum of a star cluster in integrated light.

[ascl:2006.015]
ARCHI: Add-on pipeline module for background star analysis from CHEOPS data

Silva, André M.; Sousa, Sérgio G.; Santos, Nuno; Demangeon, Olivier D. S.; Silva, Pedro; Hoyer, S.; Guterman, P.; Deleuil, Magali; Ehrenreich, David

The CHaracterizing ExOPlanet Satellite (CHEOPS) mission pipeline provides photometry for the central star in its field; ARCHI takes in data from the CHEOPS mission pipeline, analyzes the background stars, and determines the photometry of these stars, thus creating the possibility of producing photometric time-series of several close-by targets at once, in addition to using different stars in the image to calibrate systematic errors.

[ascl:2006.014]
CARACal: Containerized Automated Radio Astronomy Calibration pipeline

Józsa, Gyula I. G.; White, Sarah V.; Thorat, Kshitij; Smirnov, Oleg M.; Serra, Paolo; Ramatsoku, Mpati; Ramaila, Athanaseus J. T.; Perkins, Simon J.; Molnár, Dániel Cs.; Makhathini, Sphesihle; Maccagni, Filippo M.; Kleiner, Dane; Kamphuis, Peter; Hugo, Benjamin V.; de Blok, W. J. G.; Andati, Lexy A. L.

CARACal (Containerized Automated Radio Astronomy Calibration, formerly MeerKATHI) reduces radio-interferometric data. Developed originally as an end-to-end continuum- and line imaging pipeline for MeerKAT, it can also be used with other radio telescopes. CARACal reduces large data sets and produces high-dynamic-range continuum images and spectroscopic data cubes. The pipeline is platform-independent and delivers imaging quality metrics to efficiently assess the data quality.

[ascl:2006.013]
JoXSZ: Joint X-ray and SZ fitting for galaxy clusters in Python

JoXSZ jointly fits the thermodynamic profiles of galaxy clusters from both SZ and X-ray data using a Markov chain Monte Carlo fitting algorithm. It is an enhanced version of preprofit (ascl:1910.002), which fits only SZ data. JoXSZ parameterizes the pressure and electron density profile of a galaxy cluster with a given center and derives the temperature profile as the ratio of these quantities through the ideal gas law. The X-ray and SZ-based temperatures can be similar or different, which allows study of the cluster elongation along line of sight, gas clumping, or calibration uncertainties.

[ascl:2006.012]
pxf_kin_err: Radial velocity and velocity dispersion uncertainties estimator

pxf_kin_err estimates the radial velocity and velocity dispersion uncertainties based solely on the shape of a template spectrum used in the fitting procedure and signal-to-noise information. This method can be used for exposure time calculators, in the design of observational programs and estimates on expected uncertainties for spectral surveys of galaxies and star clusters, and as an accurate substitute for Monte-Carlo simulations when running them for large samples of thousands of spectra is unfeasible.

[ascl:2006.011]
SERVAL: SpEctrum Radial Velocity AnaLyser

Zechmeister, M.; Reiners, A.; Amado, P. J.; Azzaro, M.; Bauer, F. F.; Béjar, V. J. S.; Caballero, J. A.; Guenther, E. W.; Hagen, H. -J.; Jeffers, S. V.; Kaminski, A.; Kürster, M.; Launhardt, R.; Montes, D.; Morales, J. C.; Quirrenbach, A.; Reffert, S.; Ribas, I.; Seifert, W.; Tal-Or, L. Wolthoff, V.

SERVAL calculates radial velocities (RVs) from stellar spectra. The code uses least-squares fitting algorithms to derive the RVs and additional spectral diagnostics. Forward modeling in pixel space is used to properly weight pixel errors, and the stellar templates are reconstructed from the observations themselves to make optimal use of the RV information inherent in the stellar spectra.

[ascl:2006.010]
PRISim: Precision Radio Interferometer Simulator

PRISim is a modular radio interferometer array simulator, including the radio sky and instrumental effects, and generates a transit dataset in HD5 format.

[ascl:2006.009]
AxionNS: Ray-tracing in neutron stars

AxionNS computes radio light curves resulting from the resonant conversion of Axion dark matter into photons within the magnetosphere of a neutron star. Photon trajectories are traced from the observer to the magnetosphere where a root finding algorithm identifies the regions of resonant conversion. Given the modeling of the axion dark matter distribution and conversion probability, one can compute the photon flux emitted from these regions. The individual contributions from all the trajectories is then summed to obtain the radiated photon power per unit solid angle.

[ascl:2006.008]
DeepSphere: Graph-based spherical convolutional neural network for cosmology

DeepSphere implements a generalization of Convolutional Neural Networks (CNNs) to the sphere. It models the discretized sphere as a graph of connected pixels. The resulting convolution is more efficient (especially when data doesn't span the whole sphere) and mostly equivariant to rotation (small distortions are due to the non-existence of a regular sampling of the sphere). The pooling strategy exploits a hierarchical pixelization of the sphere (HEALPix) to analyze the data at multiple scales. The graph neural network model is based on ChebNet and its TensorFlow implementation.

[ascl:2006.007]
TATTER: Two-sAmple TesT EstimatoR

TATTER (Two-sAmple TesT EstimatoR) performs two-sample hypothesis test. The two-sample hypothesis test is concerned with whether distributions p(x) and q(x) are different on the basis of finite samples drawn from each of them. This ubiquitous problem appears in a legion of applications, ranging from data mining to data analysis and inference. This implementation can perform the Kolmogorov-Smirnov test (for one-dimensional data only), Kullback-Leibler divergence, and Maximum Mean Discrepancy (MMD) test. The module performs a bootstrap algorithm to estimate the null distribution and compute p-value.

[ascl:2006.006]
CosmoLike: Cosmological Likelihood analyses

CosmoLike analyzes cosmological data sets and forecasts future missions. It has been used in the analysis of the Dark Energy Survey and to optimize the Large Synoptic Survey Telescope and the Wide-Field Infrared Survey Telescope, and is useful for innovative theory projects that test new concepts and methods to enhance the constraining power of cosmological analyses.

[ascl:2006.005]
CosmoCov: Configuration space covariances for projected galaxy 2-point statistics

CosmoCov computes configuration space covariances for projected galaxy 2-point statistics based on the CosmoLike (ascl:2006.006) framework. The package provides a flat sky covariance module, computed with the 2D-FFTLog (ascl:2006.004) algorithm, and a curved sky covariance module.

[ascl:2006.004]
2D-FFTLog: Generalized FFTLog algorithm for non-Gaussian covariance matrices

2D-FFTLog takes the FFTLog algorithm for 1D Hankel transforms and generalizes it for 2D Hankel transforms. The algorithm is useful for efficiently computing non-Gaussian covariance matrices of cosmological 2-point statistics in configuration space from Fourier space covariances. Fast bin-averaging method is also developed for both the logarithmic binning and general binning choices. C and Python versions of the code are available.

[ascl:2006.003]
KinMS: Three-dimensional kinematic modeling of arbitrary gas distributions

The KinMS (KINematic Molecular Simulation) package simulates observations of arbitrary molecular/atomic cold gas distributions from interferometers and line observations from integral field units. This modeling tool is optimized for situations where one has analytic forms for *e.g.* the rotation curve and/or surface brightness profiles (and may want to fit the parameters of these parametric models). It can, however, also be used as a tilted-ring modelling code. The routines are flexible and have been used in various different applications, including investigating the kinematics of molecular gas in early-type galaxies and determining supermassive black-hole masses from CO interferometric observations. They are also useful for creating mock observations from hydrodynamic simulations, and input data-cubes for further simulation in, for example, CASA's (ascl:1107.013) sim_observe tool. Interactive Data Language (IDL) and Python versions of the code are available.

[ascl:2006.002]
PRIISM: Python module for Radio Interferometry Imaging with Sparse Modeling

PRIISM images radio interferometry data using the sparse modeling technique. In addition to generating an image, PRIISM can choose the best image from a range of processing parameters using cross validation. User can obtain statistically optimal images by providing the visibility data with some configuration parameters. The software is implemented as a Python module.

[ascl:2006.001]
HEARSAY: Simulations for the probability of alien contact

HEARSAY computes simulations of the causal contacts between emitters in the Galaxy. It implements the Stochastic Constrained Causal Contact Network (SC3Net) model and explores the parameter space of the model for the emergence of communicating nodes through Monte Carlo simulations and analyzes their causal connections. This model for the abundance and duration of civilizations is based on minimal assumptions and three free parameters, with focus on the statistical properties of empirical models instead of an interpretable model with variables to be determined by observation.

[ascl:2005.020]
HIPSTER: HIgh-k Power Spectrum EstimatoR

HIPSTER (HIgh-k Power Spectrum EstimatoR) computes small-scale power spectra and isotropic bispectra for cosmological simulations and galaxy surveys of arbitrary shape. The code computes the Legendre multipoles of the power spectrum, *P _{ℓ}(k)*, or bispectrum

[ascl:2005.019]
MCRaT: Monte Carlo Radiation Transfer

MCRaT (Monte Carlo Radiation Transfer) analyzes the radiation signature expected from astrophysical outflows. MCRaT injects photons in a FLASH (ascl:1010.082) simulation and individually propagates and compton scatters the photons through the fluid until the end of the simulation. This process of injection and propagating occurs for a user specified number of times until there are no more photons to be injected. Users can then construct light curves and spectra with the MCRaT calculated results. The hydrodynamic simulations used with this version of MCRaT must be in 2D; however, the photon propagation and scattering is done in 3D by assuming cylindrical symmetry. Additionally, MCRaT uses the full Klein–Nishina cross section including the effects of polarization, which can be fully simulated in the code. MCRaT works with FLASH hydrodynamic simulations and PLUTO (ascl:1010.045) AMR simulations, with both 2D spherical (r, equation) and 2D cartesian ((x,y) and (r,z)).

[ascl:2005.018]
RFCDE: Random Forests for Conditional Density Estimation

RFCDE provides an implementation of random forests designed for conditional density estimation. It computes a kernel density estimate of y with nearest neighbor weightings defined by the location of the evaluation point x relative to the leaves in the random forest.

[ascl:2005.017]
cdetools: Tools for Conditional Density Estimates

cdetools provides tools for evaluating conditional density estimates and has applications to photometric redshift estimation and likelihood-free cosmological inference. Available in R and Python, it provides functions for computing a so-called CDE loss function for tuning and assessing the quality of individual probability density functions (PDFs) and diagnostic functions that probe the population-level performance of the PDFs.

[ascl:2005.016]
RAPP: Robust Automated Photometry Pipeline

RAPP is a robust automated photometry pipeline for blurred images. RAPP requires that the observed images contain at least three stars and applies bias, dark, and flat field correction on blurred observational raw data; it also uses the median of adjacent pixels to eliminate outliers. It also uses star enhancement and robust image matching, extracts stars, and performs aperture photometry to extract information from blurred images.

[ascl:2005.015]
AMPEL: Alert Management, Photometry, and Evaluation of Light curves

Nordin, J.; Brinnel, V.; van Santen, J.; Bulla, M.; Feindt, U.; Franckowiak, A.; Fremling, C.; Gal-Yam, A.; Giomi, M.; Kowalski, M.; Mahabal, A.; Miranda, N.; Rauch, L.; Reusch, S.; Rigault, M.; Schulze, S.; Sollerman, J.; Stein, R.; Yaron, O.; van Velzen, S.; Ward, C.

AMPEL provides an analysis framework for high-throughput surveys and is suited for streamed data. The package combines the functionality of an alert broker with a generic framework capable of hosting user-contributed code; it encourages provenance and keeps track of the varying information states that a transient displays. The latter concept includes information gathered over time and data policies such as access or calibration levels.

[ascl:2005.014]
FETCH: Fast Extragalactic Transient Candidate Hunter

FETCH (Fast Extragalactic Transient Candidate Hunter) provides real-time classification of candidates from single pulse search pipelines. The package takes in a candidate file of frequency-time and DM-time data and, for each candidate and choice of model, provides the probability that the candidate is an FRB. FETCH also provides a framework for fine-tuning the models to further improve its performance for particular backends.

[ascl:2005.013]
qubefit: MCMC kinematic modeling

qubefit fits an observed data cube to generate a model cube from a user-defined emission model. The model cube is convolved with the observed beam, after which residuals between the convolved model and the observed data cube are minimized using a Markov chain Monte Carlo approach. qubefit also determines estimates of the uncertainty for each parameter of the model.

[ascl:2005.012]
2DBAT: 2D Bayesian Automated Tilted-ring fitter

2DBAT implements Bayesian fits of 2D tilted-ring models to derive rotation curves of galaxies. It performs 2D tilted-ring analysis based on a Bayesian Markov Chain Monte Carlo (MCMC) technique, thus quantifying the kinematic geometry of galaxy discs, and deriving high-quality rotation curves that can be used for mass modeling of baryons and dark matter halos.

[ascl:2005.011]
gotetra: Cosmic velocity fields tracking through the use of tetrahedra

gotetra uses phase-space tesselation techniques to extract information about cosmological N-body simulations. The key applications of this Go-based code are the measurement of splashback shells around halos and the generation of high resolution images of density fields. The package includes routines to generates 3D and 2D (projected) density fields from a particle snapshot generated by a cosmological N-body simulation, measure density along lines of sight from the center of halos, and compresse the position space data from cosmological N-body simulations. Included are two helper libraries with functions for calculating cosmological quantities and computing a number of useful mathematical functions.

[ascl:2005.010]
NNKCDE: Nearest Neighbor Kernel Conditional Density Estimation

NNKCDE is a simple and easily interpretable Conditional Density Estimation (CDE) method. It computes a kernel density estimate of y using the k nearest neighbors of the evaluation point x. The model has only two tuning parameters: the number of nearest neighbors k and the bandwidth h of the smoothing kernel in y-space. Both tuning parameters are chosen in a principled way by minimizing the CDE loss on validation data.

[ascl:2005.009]
s3PCF: Compute the 3-point correlation function in the squeezed limit

s3PCF computes the 3-point correlation function (3PCF) in the squeezed limit given galaxy positions and pair positions. The code is currently written specifically for the Abacus simulations, but the main functionalities can be also easily adapted for other galaxy catalogs with the appropriate properties.

[ascl:2005.008]
HiFLEx: Echelle data reduction pipeline

Errmann, Ronny; Cook, Neil; Anglada-Escudé, Guillem; Sithajan, Sirinrat; Mkrtichian, David; Semenko, Eugene; Martin, William; Tanvir, Tabassum S.; Feng, Fabo; Collett, James L.; Jones, Hugh R. A.

HiFLEx reduces echelle data taken with a single or bifurcated fiber input. It takes a FITS image file (*i.e.*, a CCD image) and runs data reduction steps, extracts out orders from an Echelle spectrograph (regardless of separation and curvature, as long as orders are distinguishable from one-another), applies the wavelength correction, measures the radial velocity, and performs further calibration steps.

[ascl:2005.007]
Carpyncho: VVV Catalog browser toolkit

Carpyncho browses catalogs to search for and characterize time variable data of the Vista Variables in the Via Lactea (VVV) Survey. The stacked pawprint data from the Cambridge Astronomical Science Unit's (CASU) Vista Data Flow System (VDFS) v>= 1.3 catalogs have been crossed matched with the VDFS CASU v1.3 tile catalogs into Parquet files, allowing detection and classification of periodic variables within this dataset.

[ascl:2005.006]
FFANCY: Fast Folding Algorithm for pulsar searching

FFANCY uses the Fast Folding Algorithm (FFA) on a distributed-computing framework to search for pulsars in time-domain series data. This enables the algorithm to be applied to all-sky blind pulsar surveys. The package runs an implementation of the FFA on real or simulated pulsar time series data in either SIGPROC (ascl:1107.016) or PRETSO (ascl:1107.017) format with a choice of additional algorithms to be used in the evaluation of each folded profile and outputs a periodogram along with other output threads used for testing. It also contains routines that convert the periodogram output into a list of pulsar candidates with options for candidate grouping and harmonic matching, generate simulated pulsar profiles for use in testing profile evaluation algorithms independent of the FFA, provide basic statistics for the folded profiles produced by progeny, test individual profiles using profiles produced by progeny, and other complementary functions.

[ascl:2005.005]
RoLo: Calculate radius and potential of the Roche Lobe

RoLo (Roche Lobe) calculates the radius and potential of the Roche Lobe for any specified direction, and also gives some other commonly used quantities (such as the Lagrange points). The calculator is valid for any mass ratio q between 0.01 and 100. The coordinates are spherical-polar (R, theta, phi) centered on one star (M1), with the x-axis (theta=pi/2, phi=0) pointing towards the other star (M2). The mass ratio is defined as q=M2/M1. Distances are given in units of the binary separation, a. A circular orbit is assumed.

[ascl:2005.004]
REDFIT: Red-noise spectra directly from unevenly spaced time series

Time series are commonly unevenly spaced in time make it difficult to obtain an accurate estimate of their typical red-noise spectrum. REDFIT overcomes this problem by fitting a first-order autoregressive (AR1) process directly to unevenly spaced time series. Hence, interpolation in the time domain and its inevitable bias can be avoided. The program can be used to test if peaks in the spectrum of a time series are significant against the red-noise background from an AR1 process.

[ascl:2005.003]
RM-Tools: Rotation measure (RM) synthesis and Stokes QU-fitting

RM-Tools analyzes radio polarization data, specifically the use of Faraday rotation measure synthesis and Stokes QU model fitting. It contains routines for both single-pixel 1D polarized spectra as well as 3D polarization cubes. RM-Tools is intended to serve as a toolkit for studies of polarized radio sources and measurements of their Faraday rotation. RM-Tools is the core package for the pipelines used for the POlarized Sky Survey of the Universe's Magnetism (POSSUM) and the polarization component of the Very Large Array Sky Survey (VLASS). The package is maintained by the Canadian Initiative for Radio Astronomy Data Analysis (CIRADA; cirada.org).

[ascl:2005.002]
michi2: SED and SLED fitting tool

michi2 fits combinations of arbitrary numbers of libraries/components to a given observational data. Written in C++ and Python, this chi-square fitting tool can fit a galaxy's spectral energy distribution (SED) with stellar, active galactic nuclear, dust and radio SED templates, and fit a galaxy's spectral line energy distribution (SLED) with one or more gas components using radiative transfer LVG model grid libraries.

michi2 first samples the high-dimensional parameter space (N1*N2*N3*..., where N is the number of independent templates in each library, and 1/2/3 is the ID of components) in an optimized way for a few thousand or tens of thousand times to compute the chi-square to the input observational data, then uses Python scripts to analyze the chi-square distribution and derive the best-fit, median, lower and higher 1-sigma values for each parameter in each library/component. This tool is useful for fitting larger number of templates and arbitrary combinations of libraries/components, including some constraining of one library/component onto another.

[ascl:2005.001]
RID: Relativistic Image Doubling in water Cherenkov detectors

RID (Relativistic Image Doubling in water Cherenkov detectors) uses Monte Carlo simulations to find the relative fraction of charged, relativistic particles entering a HAWC-like Water Cherenkov Detector that can cause a Relativistic Image Doubling (RID) effect visible to at least one of the internal detectors. The technique is available in C++ and Fortran; RID also includes python code for the horizontal incidence of the muon inside the tank.

[submitted]
sslf: A simple spectral-line finder

sslf is designed to be a dead-simple, effective and useful spectral line finder for 1D data. It utilizes the continuous wavelet transform from scipy, which is a productive way to find even weak spectral lines.

[ascl:2004.016]
PRECISION: Astronomical infrared observations data reduction

PRECISION reduces astronomical IR imaging data. Written with SPHERE data in mind, it provides a fast and easy reduction of bright sources suitable for science. While it may not extract the absolute maximum amount of science, the objective is to provide a means to get science-ready data with minimal computing time or human interaction.

[ascl:2004.015]
IRDAP: SPHERE-IRDIS polarimetric data reduction pipeline

van Holstein, R. G.; Girard, J. H.; de Boer, J.; Snik, F.; Milli, J.; Stam, D. M.; Ginski, C.; Mouillet, D.; Wahhaj, Z.; Schmid, H. M.; Keller, C. U.; Langlois, M.; Dohlen, K.; Vigan, A.; Pohl, A.; Carbillet, M.; Fantinel, D.; Maurel, D.; Origné, A.; Petit, C. Ramos, J.; Rigal, F.; Sevin, A.; Boccaletti, A.; Le Coroller, H.; Dominik, C.; Henning, T.; Lagadec, E.; Ménard, F.; Turatto, M.; Udry, S.; Chauvin, G.; Feldt, M.; Beuzit, J. -L.

IRDAP (IRDIS Data reduction for Accurate Polarimetry) accurately reduces SPHERE-IRDIS polarimetric data. It is a highly-automated end-to-end pipeline; its core feature is model-based correction of the instrumental polarization effects. IRDAP handles data taken both in field- and pupil-tracking mode and using the broadband filters Y, J, H and Ks. Data taken with the narrowband filters can be reduced as well, although with a somewhat worse accuracy. For pupil-tracking observations IRDAP can additionally apply angular differential imaging.

[ascl:2004.014]
PyKat: Python interface and tools for Finesse

Brown, Daniel D.; Jones, Philip; Rowlinson, Samuel; Freise, Andreas; Leavey, Sean; Green, Anna C.; Toyra, Daniel

The Python wrapper PyKat extends the optical interferometer modeling software Finesse (ascl:2004.013). It provides an efficient GUI for conducting complex numerical simulations and manipulating and viewing simulation setups, and enables the use of Python's extensive scientific software ecosystem.

[ascl:2004.013]
Finesse: Frequency domain INterfErometer Simulation SoftwarE

Finesse is a numeric simulation for laser interferometers and models parametric instabilities, easily providing the required mechanical-to-optical transfer functions in imperfect and arbitrary interferometer configurations using Hermite-Gaussian beams. The code has been used to apply limits to the number and type of higher order modes used in simulation and investigate the potential use of higher order Laguerre-Gauss modes to reduce thermal noise in future gravitational wave detector designs. The PyKat wrapper (ascl:2004.014) helps automate complex Finesse tasks.

[ascl:2004.012]
ArviZ: Exploratory analysis of Bayesian models

ArviZ provides backend-agnostic tools for diagnostics and visualizations of Bayesian inference by first converting inference data into xarray objects. It includes functions for posterior analysis, model checking, comparison and diagnostics. ArviZ’s functions work with NumPy arrays, dictionaries of arrays, xarray datasets, and have built-in support for PyMC3 (ascl:1610.016), PyStan, CmdStanPy, Pyro (ascl:1507.018), NumPyro, emcee (ascl:1303.002), and TensorFlow Probability objects. A Julia wrapper is also available.

[ascl:2004.011]
FUNDPAR: Deriving FUNDamental PARameters from equivalent widths

FUNDPAR determines fundamental parameters of solar-type stars, by using as input the Equivalent Widths of Fe I,II lines. The code uses solar-scaled ATLAS9 model atmospheres with NEWODF opacities, together with the 2009 version of the MOOG (ascl:1202.009) program. Parameter files control different details, such as the mixing-length parameter, the overshooting, and the damping of the lines. FUNDPAR also derives the uncertainties of the parameters.

[ascl:2004.010]
kombine: Kernel-density-based parallel ensemble sampler

kombine is an ensemble sampler built for efficiently exploring multimodal distributions. By using estimates of ensemble’s instantaneous distribution as a proposal, it achieves very fast burnin, followed by sampling with very short autocorrelation times.

[ascl:2004.009]
stardate: Measure precise stellar ages

stardate measures precise stellar ages by combining isochrone fitting with gyrochronology (rotation-based ages) to increase the precision of stellar ages on the main sequence. The best possible ages provided by stardate will be for stars with rotation periods, though ages can also be predicted for stars without rotation periods. stardate is an extension to isochrones that incorporates gyrochronology and the code reverts back to isochrones when no rotation period is provided.

[ascl:2004.008]
PPMAP: Column density mapping with extra dimensions

PPMAP provides column density mapping with extra dimensions (temperature and dust opacity index); it generate image cubes of differential column density as a function of (x,y) sky position and temperature for diffuse dusty structures. The code incorporates parallel processing using OpenMP for some of the more CPU-intensive steps. It is currently configured for the "Raven" cluster at Cardiff University and runs in a mode in which the computations are split between 16 separate nodes, each of which uses 16 cores with OpenMP.

[ascl:2004.007]
PyCosmo: Multi-purpose cosmology calculation tool

Tarsitano, Federica; Schmitt, Uwe; Refregier, Alexandre; Akeret, Joel; Amara, Adam; Gamper, Lukas; Nicola, Andrina

PyCosmo provides accurate predictions for cosmological observables including background quantities, power spectra and Limber and beyond-Limber angular power spectra. The software is designed to be interactive and user-friendly. It is available for download and is also offered on an interactive platform (PyCosmo Hub), which allows users to perform their own computations using Jupyter Notebooks without installing any software.

[ascl:2004.006]
ASTRAEUS: Semi-analytical semi-numerical galaxy evolution and reionization code

ASTRAEUS (semi-numerical rAdiative tranSfer coupling of galaxy formaTion and Reionization in n-body dArk mattEr simUlationS) self-consistently derives the evolution of galaxies and the reionization of the IGM based on the merger trees and density fields of a DM-only N-body simulation. It models gas accretion, star formation, SN feedback, the time and spatial evolution of the ionized regions, accounting for recombinations, HI fractions and photoionization rates within ionized regions, and radiative feedback. ASTRAEUS is for studying the galaxy-reionization interplay in the first billion years. The underlying code is written in C and is MPI-parallelized; its modular design allows new physical processes and galaxy properties to be added easily. ASTRAEUS can be run on a tree-branch-by-tree-branch basis (*i.e.,* fully vertical) or on a redshift-by-redshift basis (*i.e.,* fully horizontal) when evolving the galaxies by using local horizontal merger trees.

[ascl:2004.005]
PyWD2015: Wilson-Devinney code GUI

PyWD2015 provides a modern graphical user interface (GUI) for the 2015 version of the Wilson-Devinney (WD) code (ascl:2004.004). The GUI is written in Python 2.7 and uses the Qt4 interface framework. At its core, PyWD2015 generates lcin and dcin files from user inputs and sends them to WD, then reads and visualizes the output in a user-friendly way. It also includes tools that make the technical aspects of the modeling process significantly easier.

[ascl:2004.004]
WD: Wilson-Devinney binary star modeling

Wilson-Devinney binary star modeling code (WD) is a complete package for modeling binary stars and their eclipes and consists of two main modules. The LC module generates light and radial velocity curves, spectral line profiles, images, conjunction times, and timing residuals; the DC module handles differential corrections, performing parameter adjustment of light curves, velocity curves, and eclipse timings by the Least Squares criterion. WD handles eccentric orbits and asynchronous rotation, and can compute velocity curves (with proximity and eclipse effects). It offers options for detailed reflection and nonlinear (logarithmic law) limb darkening, adjustment of spot parameters, an optional provision for spots to drift over the surface, and can follow light curve development over large numbers of orbits. Absolute flux solution allow Direct Distance Estimation (DDE) and there are improved solutions for ellipsoidal variables and for eclipsing binaries (EBs) with very shallow eclipses. Absolute flux solutions also can estimate temperatures of both EB components under suitable circumstances.

[ascl:2004.003]
IllinoisGRMHD: GRMHD code for dynamical spacetimes

IllinoisGRMHD is an open-source, highly-extensible rewrite of the original closed-source general relativistic (ideal) magnetohydrodynamics (GRMHD) code of the Illinois Numerical Relativity (ILNR) Group. Reducing the learning curve was the primary focus of this rewrite, with the goal of facilitating community involvement in the code's use and development, as well as reducing the human effort necessary to generate new science. IllinoisGRMHD also saves computer time, generating roundoff-precision identical output to the original code on adaptive-mesh grids while being nearly twice as fast at scales of hundreds to thousands of cores.

[ascl:2004.002]
Tangra: Software for video photometry and astrometry

Tangra performs scientific grade data reduction of GPS time-tagged video observations, including reduction of stellar occultation light curves and astrometry of close flybys of Near Earth Objects. It offers Dark and Flat frame image correction, PSF and aperture photometry, multiple methods for deriving a background as well as extensibility via add-ins. Tangra is actively developed for Windows and the current version of the software supports UCAC2, UCAC3, UCAC4, NOMAD, PPMXL and Gaia DR2 star catalogues for astrometry. The software can perform motion-fitting for fast objects and derive a mini-normal astrometric positions. The supported video file formats are AVI, SER, ADV and AAV. Tangra can be also used with observations recorded as a sequence of FITS files. There are also versions for Linux and OS-X with more limited functionality.

[ascl:2004.001]
Locus: Optimized differential photometry

Locus implements the Locus Algorithm, which maximizes the performance of differential photometry systems by optimizing the number and quality of reference stars in the Field of View with the target.

[submitted]
vlt-sphere: Automatic VLT/SPHERE data reduction and analysis

The high-contrast imager SPHERE at the Very Large Telescope is a new generation instrument combining extreme adaptive optics and coronagraphy to directly image exoplanets in the near-infrared. The vlt-sphere package enables easy reduction of the data coming from IRDIS and IFS, the two near-infrared subsystems of SPHERE. The package relies on the official ESO pipeline (ascl:1402.010) which must be installed separately.

[ascl:2003.014]
Torch: Coupled gas and N-body dynamics simulator

Wall, Joshua E.; Tran, Aaron; Mac Low, Mordecai-Mark; McMillan, Stephen L. W.; Klessen, Ralf S.; Portegies Zwart, Simon; Pellegrino, Andrew

Torch simulates coupled gas and N-body dynamics in astrophysical systems such as newly forming star clusters. It combines the FLASH (ascl:1010.082) code for gas dynamics and the ph4 code for direct N-body evolution via the AMUSE framework.

[ascl:2003.013]
AstroHOG: Analysis correlations using the Histograms of Oriented Gradients

AstroHOG compares extended spectral-line observations (PPV cubes); the histogram of oriented gradients (HOG) technique takes as input two PPV cubes and provides an estimate of their spatial correlation across velocity channels to study spatial correlation between different tracers of the ISM.

[ascl:2003.012]
PYSOLATOR: Remove orbital modulation from a binary pulsar and/or its companion

PYSOLATOR removes the orbital modulation from a binary pulsar and/or its companion. In essence, it subtracts the predicted Roemer delay for the given orbit and then resamples the time series so as to make the signal appear as if it were emitted from the barycenter of the binary system, making the search for pulses easier and faster.

[ascl:2003.011]
HOMER: A Bayesian inverse modeling code

HOMER (Helper Of My Eternal Retrievals) is a machine-learning-accelerated Bayesian inverse modeling code. Given some data and uncertainties, the code determines the posterior distribution of a model. HOMER uses MC^{3} (ascl:1610.013) for its MCMC; its forward model is a neural network surrogate model trained by MARGE (ascl:2003.010). The code produces plots of the 1D marginalized posteriors, 2D pairwise posteriors, and parameter history traces, and can also overplot the 1D and 2D posteriors for comparison with another posterior. HOMER computes the Bhattacharyya coefficient to compare the similarity of two 1D marginalized posteriors.

[ascl:2003.010]
MARGE: Machine learning Algorithm for Radiative transfer of Generated Exoplanets

MARGE (Machine learning Algorithm for Radiative transfer of Generated Exoplanets) generates exoplanet spectra across a defined parameter space, processes the output, and trains, validates, and tests machine learning models as a fast approximation to radiative transfer. It uses BART (ascl:1608.004) for spectra generation and modifies BART’s Bayesian sampler (MC^{3}, ascl:1610.013) with a random uniform sampler to propose models within a defined parameter space. More generally, MARGE provides a framework for training neural network models to approximate a forward, deterministic process.

[ascl:2003.009]
TOASTER: Times-Of-Arrival Tracker

Lazarus, Patrick; Bassa, Cees; Desvignes, Greg; Hessels, Jason; Verbiest, Joris; Karuppusamy, Ramesh

TOASTER is a pulse times-of-arrival (TOA) tracker. It stores reduced/folded observations, meta data, templates, parfiles, TOAs, and timefiles in an organized manner using an SQL database. TOASTER also provides a full-featured python toolkit for reliably interacting with the data and database, and provides scripts that, for example, list and summarize the TOAs in the data base, and generate TOA files in multiple formats. The framework can also be used to generate TOAs from observations using flexible and reproducible plugins referred to as "manipulators".

[ascl:2003.008]
CoastGuard: Automated timing data reduction pipeline

Lazarus, P.; Karuppusamy, R.; Graikou, E.; Caballero, R. N.; Champion, D. J.; Lee, K. J.; Verbiest, J. P. W.; Kramer, M

CoastGuard reduces Effelsberg data; it is written in python and based on PSRCHIVE (ascl:1105.014). Though primarily designed for Effelsberg PSRIX data, it contains components sufficiently general for use with psrchive-compatible data files from other observing systems. In particular, the radio frequency interference (RFI) removal algorithm has been applied to data from the Parkes Telescope and has also been adopted by the LOFAR pulsar timing data reduction pipeline.

[ascl:2003.007]
RAPID: Real-time Automated Photometric IDentification

RAPID (Real-time Automated Photometric IDentification) classifies multiband photometric light curves into several different transient classes. It uses a deep recurrent neural network to produce time-varying classifications, and because it does not rely on deriving computationally expensive features from the data, it is well suited for processing alerts that wide-field surveys such as the Zwicky Transient Facility (ZTF) and the Large Synoptic Survey Telescope (LSST) will produce.

[ascl:2003.006]
PORTAL: POlarized Radiative Transfer Adapted to Lines

PORTAL (POlarized Radiative Transfer Adapted to Lines), a 3D polarized radiative transfer code, simulates the emergence of polarization in the emission of atomic or molecular (sub-)millimeter lines. Written in Fortran90, PORTAL can be used in standalone mode or can process the output of other 3D radiative transfer codes

[ascl:2003.005]
RHT: Rolling Hough Transform

The RHT (Rolling Hough Transform) measures linear intensity as a function of orientation in images. This machine vision algorithm works on any image-space (2D) data, and quantifies the presence of linear structure as a function of orientation. The RHT can be used to identify linear features in images, to quantify the orientation of structure in images, and to map image intensity from 2D x-y space to 3D x-y-orientation space. An option in the code allows the user to quantify intensity as a function of direction (modulo 2pi) rather than orientation (modulo pi). The RHT was first used to discover that filamentary structures in neutral hydrogen emission are aligned with the ambient magnetic field.

[ascl:2003.004]
scousepy: Semi-automated multi-COmponent Universal Spectral-line fitting Engine

scousepy is a Python implementation of spectral line-fitting IDL code SCOUSE (ascl:1601.003). It fits a large amount of complex astronomical spectral line data in a systematic way.

[ascl:2003.003]
acorns: Agglomerative Clustering for ORganising Nested Structures

acorns generates a hierarchical system of clusters within discrete data by using an n-dimensional unsupervised machine-learning algorithm that clusters spectroscopic position-position-velocity data. The algorithm is based on a technique known as hierarchical agglomerative clustering. Although acorns was designed with the analysis of discrete spectroscopic position-position-velocity (PPV) data in mind (rather than uniformly spaced data cubes), clustering can be performed in n-dimensions and the algorithm can be readily applied to other data sets in addition to PPV measurements.

[ascl:2003.002]
MAGNETAR: Histogram of relative orientation calculator for MHD observations

MAGNETAR is a set of tools for the study of the magnetic field in simulations of MHD turbulence and polarization observations. It calculates the histogram of relative orientation between density structure in the magnetic field in data cubes from simulations of MHD turbulence and observations of polarization using the method of histogram of relative orientations (HRO).

[ascl:2003.001]
TESS-Point: High precision TESS pointing tool

Burke, Christopher J.; Levine, Alan; Fausnaugh, Michael; Vanderspek, Roland; Barclay, Thomas; Libby-Roberts, Jessica E.; Morris, Brett; Sipocz, Brigitta; Owens, Martin; Feinstein, Adina D.; Camacho, Joao

TESS-Point converts astronomical target coordinates given in right ascension and declination to detector pixel coordinates for the MIT-led NASA Transiting Exoplanet Survey Satellite (TESS) spacecraft. The program can also provide detector pixel coordinates for a star by TESS input catalog identifier number and common astronomical name. Tess-point outputs the observing sector number, camera number, detector number, and pixel column and row.

[submitted]
Pulseblast 2

PulseBlast 2 is a mass directory scanner for Unix, Darwin and DOS/NT kernels that allows the user to analyze PSRFITS data all the way from raw time series to fully calibrated and RFI excised TOAs. Many PSRFITS software packages in the past have dealt with data on the scale of a single file and many optimization methods for pulsar data currently used are used in PulseBlast 2 for individual file analysis. However, what sets PulseBlast 2 apart is its optimization across many directories spread out across the OS. The program is entirely written in Python 3, with a scope to change string operations into Perl scripts.

[ascl:2002.022]
DISKMODs: Accretion Disk Radial Structure Models

DISKMODs provides radial structure models of accretion disk solutions. The following models are included: Novikov-Thorne thin disk model and Sadowski polytropic slim disk model. Each model implements a common interface that gives the radial dependence of selected geometrical, physical and thermodynamic quantities of the accretion flow. The model interpolates through a set of tabulated numerical solutions. These solutions are computed for a reference mass M=10 M_{sun}. The model can rescale the disk structure to any mass, with masses in the range of 5-20 M_{sun} giving reasonably good results.

[ascl:2002.021]
CR-SISTEM: Symplectic integrator for lunar core-mantle and orbital dynamics

CR-SISTEM models lunar orbital and rotational dynamics, taking into account the effects of a liquid core. Orbits of the Moon and Earth are fully integrated, and other planets (or additional point-mass satellites) may be included in the integration. Lunar and solar tides on Earth, eccentricity and obliquity tides on the Moon, and lunar core-mantle friction are included. The integrator is one file (crsistem5.for) written in FORTRAN 90, uses seven input files (settings.in, planets.in, moons.in, tidal.in, lunar.in, precess.in and core.in), and has at least eight output files (planet101.out, moon101.out, pole.out, spin_orb.out, spin_ecl.out, cspin_ecl.out, long.out and clong.out); additional moons and planets would add more output. The input files provided with the code set up a 1 Myr simulation of a slow-spinning Moon on an orbit of 40 Earth radii, which will then dynamically relax to the lowest-energy state (in this case it is a synchronous rotation with a core spinning separately from the mantle).

[ascl:2002.020]
ExoCAM: Exoplanet Community Atmospheric Model

ExoCAM adapts the NCAR Community Earth System Model (CESM) for planetary and exoplanetary applications. The system files, source code, initial conditions files, and namelists provided do not run standalone. ExoCAM is a patch to be used with standard distributions of CESM version 1.2.1 (http://www.cesm.ucar.edu/models/current.html), and is also intended to be run with ExoRT (ascl:2002.019), a correlated-k radiative transfer package.

[ascl:2002.019]
ExoRT: Two-stream radiative transfer code

ExoRT is a flexible, two-stream radiative transfer code that interfaces with CAM/CESM (http://www.cesm.ucar.edu/models/current.html) or 1D offline; it is also used with ExoCAM (ascl:2002.020). Quadrature is used for shortwave and hemispheric mean is used for longwave. The gas phase optical depths are calculate using a correlated K-distribution method, with overlapping bands treated using an amount weighted scheme. Cloud optics are treated using mie scattering for both liquid and ice clouds, and cloud overlap is treated using Monte Carlo Independent Column Approximation.

[ascl:2002.018]
Bayesfit: Command-line program for combining Tempo2 and MultiNest components

Bayesfit pulls together Tempo2 (ascl:1210.015) and MultiNest (ascl:1109.006) components to provide additional functionality such as the specification of priors; Nelder–Mead optimization of the maximum-posterior point; and the capability of computing the partially marginalized likelihood for a given subset of timing-model parameters. Bayesfit is a single python command-line application.

[ascl:2002.017]
libstempo: Python wrapper for Tempo2

libstempo uses the Tempo2 library (ascl:1210.015) to load a pulsar's tim/par files, providing Python access to the TOAs, the residuals, the timing-model parameters, the fit procedure, and more.

[ascl:2002.016]
Cobra: Bayesian pulsar searching

Cobra uses single pulse time series data to search for and time pulsars, performing a fully phase coherent timing analysis. The GPU-accelerated Bayesian analysis package, written in Python, incorporates models for both isolated and accelerated systems, as well as both Keplerian and relativistic binaries. Cobra builds a model pulse train that incorporates effects such as aliasing, scattering and binary motion and a simple Gaussian profile and compares this directly to the data; the software can thus combine data over multiple frequencies, epochs, or even across telescopes.

[ascl:2002.015]
GizmoAnalysis: Read and analyze Gizmo simulations

GizmoAnalysis reads and analyzes N-body simulations run with the Gizmo code (ascl:1410.003). Written in Python, it was developed primarily to analyze FIRE simulations, though it is usable with any Gizmo snapshot files. It offers the following functionality: reads snapshot files and converts particle data to physical units; provides a flexible dictionary class to store particle data and compute derived quantities on the fly; plots images and properties of particles; and generates region files for input to MUSIC (ascl:1311.011) to generate cosmological zoom-in initial conditions. GizmoAnalysis also computes rates of supernovae and stellar winds, including their nucleosynthetic yields, as used in FIRE simulations. The software package includes a tutorial in a Jupyter notebook.

[ascl:2002.014]
HaloAnalysis: Read and analyze halo catalogs and merger trees

HaloAnalysis reads and analyzes halo/galaxy catalogs, generated from Rockstar (ascl:1210.008) or AHF (ascl:1102.009), and merger trees generated from Consistent Trees (ascl:1210.011). Written in Python, it offers the following functionalities: reads halo/galaxy/tree catalogs from multiple file formats; assigns baryonic particles and galaxy properties to dark-matter halos; combines and re-generates halo/galaxy/tree files in hdf5 format; analyzes properties of halos/galaxies; and selects halos to generate zoom-in initial conditions. The code includes a tutorial in the form of a Jupyter notebook.

[ascl:2002.013]
GWecc: Calculator for pulsar timing array signals due to eccentric supermassive binaries

GWecc computes the pulsar timing array (PTA) signals induced by eccentric supermassive binaries. Written in C++, it computes the plus/cross polarizations as well as Earth and pulsar terms of the PTA signal given the binary parameters and the sky locations of the binary and the pulsar. A python wrapper is included through which GWecc can be used to simulate, search for and constrain gravitational wave-emitting eccentric supermassive binaries using packages such as ENTERPRISE (ascl:1912.015) and libstempo (ascl:2002.017).

[ascl:2002.012]
DMRadon: Radon Transform calculation tools

DMRadon calculates the Radon Transform for use in the analysis of Directional Dark Matter Direct Detection. The code can calculate speed distributions, velocity distribution, velocity integral (eta) and Radon Transforms or a standard Maxwell-Boltzmann distribution. DMRadon also calculates the velocity distribution averaged over different angular bins.

[ascl:2002.011]
PyHammer: Python spectral typing suite

Kesseli, Aurora; West, Andrew; Veyette, Mark; Harrison, Brandon; Feldman, Dan; Morgan, Dylan; Theissen, Chris; Robinson, Connor

PyHammer performs rapid and automatic spectral classification of stars according to the Morgan-Keenan classification system; it is a Python revision of the IDL code The Hammer (ascl:1405.003) and offers additional capabilities. Working in the range of 3,650-10,200 Angstroms, the automatic spectral typing algorithm compares important spectral lines to template spectra and determines the best matching spectral type, ranging from O to L type stars. The code can also determine a star's metallicity ([Fe/H]) and radial velocity shifts. Once the automatic classification algorithm has run, PyHammer provides the user an interface for determining spectral types visually by comparing their spectra to provided templates.

[ascl:2002.010]
Apercal: Pipeline for the Westerbork Synthesis Radio Telescope Apertif upgrade

Apercal is a dedicated, automated data reduction and analysis pipeline written for the Apertif (APERture Tile In Focus) upgrade to the Westerbork Synthesis Radio Telescope. This upgrade dramatically increases the field of view and survey speed of the telescope and is being used for survey observations that can produce 5 terabytes of data for each observation. Apercal uses existing and new tools and parallelization to provide the performance needed for the large volume of data produced Apertif surveys. The software is written entirely in Python and uses third–party astronomical software, such as AOFlagger (ascl:1010.017), CASA (ascl:1107.013), and Miriad (ascl:1106.007), for certain tasks. Apercal is modular, making it possible to run specific modules manually instead of the full pipeline, and information can be exchanged between modules because status parameters are written and read from a python pickled dictionary file. The pipeline can also run fully automatically.

[ascl:2002.009]
DASH: Deep Automated Supernova and Host classifier

DASH classifies the type, age, redshift and host for any supernova spectra based on the learned features, through use of a deep convolutional neural network to train a matching algorithm, of each supernova’s type and age. The Python library allows a user to classify spectra; the software is fast and can classify thousands of spectra in seconds. A graphical interface that enables a user to view and classify a spectrum is also available.

[ascl:2002.008]
ExoSim: Simulator for predicting signal and noise in transit spectroscopy observations

ExoSim models host star and planet transit events, simulating the temporal change in stellar flux due to the light curve. It is wavelength-dependent, using an input planet spectrum to determine the light curve depth for any given wavelength and can capture temporal effects, such as correlated noise. ExoSim's star spot simulator produces simulated observations that include spot and facula contamination. The code is flexible and can be generically applied to different instruments that simulate specific time-dependent processes.

[ascl:2002.007]
ProSpect: Spectral generation package

ProSpect generates good quality SEDs that can be used to estimate the broad band photometric properties of galaxies that have known star formation and gas metallicity histories. It allows for complex star formation and metallicity histories to be specified, and can be used in a generative or fitting (Bayesian) mode. ProSpect provides a high level interface to the BC03 (low and high resolution) and EMILES libraries, as well as the Dale 2014 dust emission templates. Its source code is available for download, and it is also available as an interactive web tool.

[ascl:2002.006]
ScamPy: Sub-halo Clustering and Abundance Matching Python interface

ScamPy "paints" an observed population of cosmological objects on top of the DM-halo/subhalo hierarchy obtained from DM-only simulations. The method combines the Halo Occupation Distribution (HOD) method with sub-halo abundance matching (SHAM); the two processes together are dubbed Sub-halo clustering and abundance matching (SCAM). The procedure requires applying the two methods in sequence; by applying the HOD scheme, the host sub-haloes are selected, and the SHAM algorithm associates an observable property of choice to each sub-halo. The provided python interface allows users to load and populate DM halos and sub-halos obtained by FoF and SUBFIND algorithms applied to DM snapshots at any redshift. The software is highly-optimized and flexible.

[ascl:2002.005]
ODUSSEAS: Observing Dwarfs Using Stellar Spectroscopic Energy-Absorption Shapes

Antoniadis-Karnavas, A.; Sousa, S. G.; Delgado-Mena, E.; Santos, N. C.; Teixeira, G. D. C.; Neves, V.

ODUSSEAS (Observing Dwarfs Using Stellar Spectroscopic Energy-Absorption Shapes) uses machine learning to derive the Teff and [Fe/H] of M dwarf stars by using their optical spectra obtained by different spectrographs with different resolutions. The software uses the measurement of the pseudo equivalent widths for more than 4000 stellar absorption lines and the machine learning Python package scikit-learn (https://scikit-learn.org/stable/) to predict the stellar parameters.

[ascl:2002.004]
triceratops: Candidate exoplanet rating tool

triceratops (Tool for Rating Interesting Candidate Exoplanets and Reliability Analysis of Transits Originating from Proximate Stars) validates planet candidates from the Transiting Exoplanet Survey Satellite (TESS). The code calculates the probabilities of a wide range of transit-producing scenarios using the primary transit of the planet candidate and preexisting knowledge of its host and nearby stars. It then uses the known properties of these stars to calculate star-specific priors for each scenario with estimates of stellar multiplicity and planet occurrence rates.

[ascl:2002.003]
ORIGIN: detectiOn and extRactIon of Galaxy emIssion liNes

ORIGIN performs blind detection of faint emitters in MUSE datacubes. The algorithm is tuned to detect faint spatial-spectral emission signatures while allowing for a stable false detection rate over the data cube, and providing in the same time an automated and reliable estimation of the purity. ORIGIN implements a nuisance removal part based on a continuum subtraction combining a Discrete Cosine Transform and an iterative Principal Component Analysis and a detection part based on the local maxima of Generalized Likelihood Ratio test statistics obtained for a set of spatial-spectral profiles of emission line emitters. In addition, it performs a purity estimation in which the proportion of true emission lines is estimated from the data itself: the distribution of the local maxima in the noise only configuration is estimated from that of the local minima.

[ascl:2002.002]
RASCAS: Resonant line transfer in AMR simulations

Michel-Dansac, Leo; Blaizot, Jeremy; Garel, Thibault; Verhamme, Anne; Kimm, Taysun; Trebitsch, Maxime

The massively parallel code RASCAS (RAdiative SCattering in Astrophysical Simulations) performs radiative transfer on an adaptive mesh with an octree structure using the Monte Carlo technique. The code features full MPI parallelization, domain decomposition, adaptive load-balancing, and a standard peeling algorithm to construct mock observations. The radiative transport of resonant line photons through different mixes of species (e.g. HI, SiII, MgII, FeII), including their interaction with dust, is implemented in a modular fashion to allow new transitions to be easily added to the code. RASCAS may also be used to propagate photons at any wavelength (e.g. stellar continuum or fluorescent lines), and has been designed to be easily customizable and to process simulations of arbitrarily large sizes on large supercomputers.

[ascl:2002.001]
SDAR: Slow-Down Algorithmic Regularization code for solving few-body problems

SDAR (Slow-Down Algorithmic Regularization) simulates the long-term evolution of few-body systems such as binaries and triples. The algorithm used provides a few orders of magnitude faster performance than the classical N-body method. The secular evolution of hierarchical systems, *e.g.* Kozai-Lidov oscillation, can be well reproduced. The code is written in the C++ language and can be used either as a stand-alone tool or a library to be plugged in other N-body codes. The high precision of the floating point to 62 digits is also supported.

[submitted]
Determination of Length of (Earth) Day [LOD] in the past geologic epochs

The protocol describes the algorithm of arriving at LOD in a given past geological Epoch. First the lunar orbital radius of the given geologic epoch has to be determined. For this the velocity of recession of Moon for the accelerated phase has to be determined. The spatial integral of the reciprocal of Velocity of recession gives the the transit time of Moon from desired orbit to the present orbit.Through several iterations the transit time is made to converge on the geologic epoch. Once we determine the desired orbital radius it has to be substituted in the LOD expression to determine the LOD in the given geologic epoch.

Would you like to view a random code?