ASCL.net

Astrophysics Source Code Library

Making codes discoverable since 1999

Browsing Codes

Results 2001-2250 of 2216 (2174 ASCL, 42 submitted)

Order
Title Date
 
Mode
Abstract Compact
Per Page
50100250All
[ascl:1907.009] Plonk: Smoothed particle hydrodynamics data analysis and visualization

Plonk analyzes and visualizes smoothed particle hydrodynamics simulation data. It is built on the scientific Python ecosystem, including NumPy, Matplotlib, Cython, h5py, SymPy, and pandas. Plock's visualization module uses Splash (ascl:1103.004) to produce images using smoothed particle hydrodynamics interpolation. The code is modular and extendible, and can be scripted or used interactively.

[ascl:1907.010] OMNICAL: Redundant calibration code for low frequency radio interferometers

OMNICAL calibrates antennas in the redundant subset of the array. The code consists of two algorithms, a logarithmic method (logcal) and a linearized method (lincal). OMNICAL makes visibilities from physically redundant baselines agree with each other and also explicitly minimizes the variance within redundant visibilities.

[ascl:1907.011] beamconv: Cosmic microwave background detector data simulator

beamconv simulates the scanning of the CMB sky while incorporating realistic beams and scan strategies. It uses (spin-)spherical harmonic representations of the (polarized) beam response and sky to generate simulated CMB detector signal timelines. Beams can be arbitrarily shaped. Pointing timelines can be read in or calculated on the fly; optionally, the results can be binned on the sphere.

[ascl:1907.012] molly: 1D astronomical spectra analyzer

molly analyzes 1D astronomical spectra. Its prime purpose is for handling large numbers of similar spectra (e.g., time series spectroscopy), but it contains many of the standard operations used for normal spectrum analysis as well. It overlaps with the various similar programs such as dipso (ascl:1405.016) and has strengths (particularly for time series spectra) and weaknesses compared to them.

[ascl:1907.013] RVSpecFit: Radial velocity and stellar atmospheric parameter fitting

RVSpecFit determines radial velocities and stellar atmospheric parameters from spectra by direct pixel fitting by interpolated stellar templates. The code doesn't require spectrum normalization and can deal with non-flux calibrated spectra. RVSpecFit is able to fit multiple spectra simultaneously.

[ascl:1907.014] sbpy: Small-body planetary astronomy

sbpy, an Astropy affiliated package, supplements functionality provided by Astropy (ascl:1304.002) with functions and methods that are frequently used for planetary astronomy with a clear focus on asteroids and comets. It offers access tools for various databases for orbital and physical data, spectroscopy analysis tools and models, photometry models for resolved and unresolved observations, ephemerides services, and other tools useful for small-body planetary astronomy.

[ascl:1907.015] TurbuStat: Turbulence statistics in spectral-line data cubes

TurbuStat implements a variety of turbulence-based statistics described in the astronomical literature and defines distance metrics for each statistic to quantitatively compare spectral-line data cubes, as well as column density, integrated intensity, or other moment maps. The software can simulate observations of fractional Brownian Motion fields, including 2-D images and optically thin H I data cubes. TurbuStat also offers multicore fast-Fourier-transform support and provides a segmented linear model for fitting lines with a break point.

[ascl:1907.016] astrodendro: Astronomical data dendrogram creator

Astrodendro, written in Python, creates dendrograms for exploring and displaying hierarchical structures in observed or simulated astronomical data. It handles noisy data by allowing specification of the minimum height of a structure and the minimum number of pixels needed for an independent structure. Astrodendro allows interactive viewing of computed dendrograms and can also produce publication-quality plots with the non-interactive plotting interface.

[ascl:1907.017] ZChecker: Zwicky Transient Facility moving target checker for short object lists

ZChecker finds, measures, and visualizes known comets in the Zwicky Transient Facility time-domain survey. Images of targets are identified using on-line ephemeris generation and survey metadata. The photometry of the targets are measured and the images are processed with temporal filtering to highlight morphological variations in time.

[ascl:1907.018] StePar: Inferring stellar atmospheric parameters using the EW method

StePar computes the stellar atmospheric parameters Teff, log g, [Fe/H], and ξ of FGK-type stars using the Equivalent Width (EW) method. The code implements a grid of MARCS model atmospheres and uses the MOOG radiative transfer code (ascl:1202.009) and TAME (ascl:1503.003). StePar uses a Downhill Simplex minimization algorithm, running it twice for any given star, to compute the stellar atmospheric parameters.

[ascl:1907.019] GaussPy: Python implementation of the Autonomous Gaussian Decomposition algorithm

GaussPy implements the Autonomous Gaussian Decomposition (AGD) algorithm, which uses computer vision and machine learning techniques to provide optimized initial guesses for the parameters of a multi-component Gaussian model automatically and efficiently. The speed and adaptability of AGD allow it to interpret large volumes of spectral data efficiently. Although it was initially designed for applications in radio astrophysics, AGD can be used to search for one-dimensional Gaussian (or any other single-peaked spectral profile)-shaped components in any data set. To determine how many Gaussian functions to include in a model and what their parameters are, AGD uses a technique called derivative spectroscopy. The derivatives of a spectrum can efficiently identify shapes within that spectrum corresponding to the underlying model, including gradients, curvature and edges.

[ascl:1907.020] GaussPy+: Gaussian decomposition package for emission line spectra

GaussPy+ is a fully automated Gaussian decomposition package for emission line spectra. It is based on GaussPy (ascl:1907.019) and offers several improvements, including automating preparatory steps and providing an accurate noise estimation, improving the fitting routine, and providing a routine to refit spectra based on neighboring fit solutions. GaussPy+ handles complex emission and low to moderate signal-to-noise values.

[ascl:1907.021] PRISM: Probabilistic Regression Instrument for Simulating Models

PRISM analyzes scientific models using the Bayes linear approach, the emulation technique, and history matching to construct an approximation ('emulator') of any given model. The software facilitates and enhances existing MCMC methods by restricting plausible regions and exploring parameter space efficiently and can be used as a standalone alternative to MCMC for model analysis, providing insight into the behavior of complex scientific models. PRISM stores results in HDF5-files and can be executed in serial or MPI on any number of processes. It accepts any type of model and comparison data and can reduce relevant parameter space by factors over 100,000 using only a few thousand model evaluations.

[ascl:1907.022] CMDPT: Color Magnitude Diagrams Plot Tool

CMD Plot Tool calculates and plots Color Magnitude Diagrams (CMDs) from astronomical photometric data, e.g. of a star cluster observed in two filter bandpasses. It handles multiple file formats (plain text, DAOPHOT .mag files, ACS Survey of Galactic Globular Clusters .zpt files) to generate professional and customized plots without a steep learning curve. It works “out of the box” and does not require any installation of development environments, additional libraries, or resetting of system paths. The tool is available as a single application/executable file with the source code. Sample data is also bundled for demonstration. CMD Plot Tool can also convert DAOPHOT magnitude files to CSV format.

[ascl:1907.023] REVOLVER: REal-space VOid Locations from suVEy Reconstruction

REVOLVER reconstructs real space positions from redshift-space tracer data by subtracting RSD through FFT-based reconstruction (optional) and applies void-finding algorithms to create a catalogue of voids in these tracers. The tracers are normally galaxies from a redshift survey but could also be halos or dark matter particles from a simulation box. Two void-finding routines are provided. The first is based on ZOBOV (ascl:1304.005) and uses Voronoi tessellation of the tracer field to estimate the local density, followed by a watershed void-finding step. The second is a voxel-based method, which uses a particle-mesh interpolation to estimate the tracer density, and then uses a similar watershed algorithm. Input data files can be in FITS format, or ASCII- or NPY-formatted data arrays.

[ascl:1907.024] Skyfield: High precision research-grade positions for planets and Earth satellites generator

Skyfield computes positions for the stars, planets, and satellites in orbit around the Earth. Its results should agree with the positions generated by the United States Naval Observatory and their Astronomical Almanac to within 0.0005 arcseconds (which equals half a “mas” or milliarcsecond). It computes geocentric coordinates or topocentric coordinates specific to your location on the Earth’s surface. Skyfield accepts AstroPy (ascl:1304.002) time objects as input and can return results in native AstroPy units but is not dependend on AstroPy nor its compiled libraries.

[ascl:1907.025] GIST: Galaxy IFU Spectroscopy Tool

GIST (Galaxy IFU Spectroscopy Tool) provides a convenient all-in-one framework for the scientific analysis of fully reduced, (integral-field) spectroscopic data, conducting all the steps from the preparation of input data to the scientific analysis and to the production of publication-quality plots. In its basic set-up, the GIST pipeline extracts stellar kinematics, performs an emission-line analysis, and derives stellar population properties from full spectral fitting and via the measurement of absorption line-strength indices by exploiting pPXF (ascl:1210.002)and GandALF routines. The pipeline is not specific to any instrument or analysis technique, and includes a dedicated visualization routine with a sophisticated graphical user interface for fully interactive plotting of all measurements, spectra, fits, and residuals, as well as star formation histories and the weight distribution of the models.

[ascl:1907.026] MCRGNet: Morphological Classification of Radio Galaxy Network

MCRGNet (Morphological Classification of Radio Galaxy Network) classifies radio galaxies of different morphologies. It is based on the Convolutional Neural Network (CNN), which is trained and applied under a three-step framework: 1.) pretraining the network unsupervisedly with unlabeled samples, 2.) fine-tuning the pretrained network parameters supervisedly with labeled samples, and 3.) classifying a new radio galaxy by the trained network. The code uses a dichotomous tree classifier composed of cascaded CNN based subclassifiers.

[ascl:1907.027] intensitypower: Spectrum multipoles modeler

intensitypower measures and models the auto- and cross-power spectrum multipoles of galaxy catalogs and radio intensity maps presented in spherical coordinates. It can also convert the multipoles to power spectrum wedges P(k,mu) and 2D power spectra P(k_perp,k_par). The code assumes the galaxy catalog is a set of discrete points and the radio intensity map is a pixelized continuous field which includes angular pixelization using healpix, binning in redshift channels, smoothing by a Gaussian telescope beam, and the addition of a Gaussian noise in each cell. The galaxy catalog and radio intensity map are transferred onto an FFT grid, and power spectrum multipoles are measured including curved-sky effects. Both maps include redshift-space distortions.

[ascl:1907.028] ROHSA: Separation of diffuse sources in hyper-spectral data

ROHSA (Regularized Optimization for Hyper-Spectral Analysis) reveals the statistical properties of interstellar gas through atomic and molecular lines. It uses a Gaussian decomposition algorithm based on a multi-resolution process from coarse to fine grid to decompose any kind of hyper-spectral observations into a sum of coherent Gaussian. Optimization is performed on the whole data cube at once to obtain a solution with spatially smooth parameters.

[ascl:1907.029] XDF-GAN: Mock astronomical survey generator

XDF-GAN generates mock galaxy surveys with a Spatial Generative Adversarial Network (SGAN)-like architecture. Mock galaxy surveys are generated from data that is preprocessed as little as possible (preprocessing is only a 99.99th percentile clipping). The outputs can also be tessellated together to create a very large survey, limited in size only by the RAM of the generation machine.

[ascl:1907.030] Wōtan: Stellar detrending methods

Wōtan provides free and open source algorithms to remove trends from time-series data automatically as an aid to to search efficiently for transits in stellar light curves from surveys. The toolkit helps determine empirically the best tool for a given job, serving as a one-stop solution for various smoothing tasks.

[ascl:1907.031] MGB: Interactive spectral classification code

MGB (Marxist Ghost Buster) attacks spectral classification by using an interactive comparison with spectral libraries. It allows the user to move along the two traditional dimensions of spectral classification (spectral subtype and luminosity classification) plus the two additional ones of rotation index and spectral peculiarities. Double-lined spectroscopic binaries can also be fitted using a combination of two standards. The code includes OB2500 v2.0, a standard grid of blue-violet R ~ 2500 spectra of O stars from the Galactic O-Star Spectroscopic Survey, but other grids can be added to MGB.

[ascl:1907.032] Astro-SCRAPPY: Speedy Cosmic Ray Annihilation Package in Python

Astro-SCRAPPY detects cosmic rays in images (numpy arrays), based on Pieter van Dokkum's L.A.Cosmic algorithm and originally adapted from cosmics.py written by Malte Tewes. This implementation is optimized for speed, resulting in slight difference from the original code, such as automatic recognition of saturated stars (rather than treating such stars as large cosmic rays, and use of a separable median filter instead of the true median filter. Astro-SCRAPPY is an AstroPy (ascl:1304.002) affiliated package.

[ascl:1908.001] QAC: Quick Array Combinations front end to CASA

QAC (Quick Array Combinations) is a front end to CASA (ascl:1107.013) and calls tools and tasks to help in combining data from a single dish and interferometer. QAC hides some of the complexity of writing CASA scripts and provide a simple interface to array combination tools and tasks in CASA. This project was conceived alongside the TP2VIS (ascl:1904.021) project, where it was used to provide an easier way to call CASA and perform regression tests.

[ascl:1908.002] Molsoft: Molonglo Telescope Observing Software

Molsoft operates, monitors and schedules observations, both through predetermined schedule files and fully dynamically, at the refurbished Molonglo Observatory Synthesis Radio Telescope (MOST). It was developed as part of the UTMOST upgrade of the facility. The software runs a large-scale pulsar timing program; the autonomous observing system and the dynamic scheduler have increased the observing efficiency by a factor of 2-3 in comparison with static scheduling.

[ascl:1908.003] ActSNClass: Active learning for supernova photometric classification

ActSNClass uses a parametric feature extraction method, Random Forest classifier and two learning strategies (uncertainty sampling and random sampling) to performs active learning for supernova photometric classification.

[ascl:1908.004] Gramsci: GRAph Made Statistics for Cosmological Information

Gramsci (GRAph Made Statistics for Cosmological Information) computes the general N-point spatial correlation functions of any discrete point set embedded within an Euclidean space of ℝ^n. It uses kd-trees and graph databases to count all possible N-tuples in binned configurations within a given length scale, e.g. all pairs of points or all triplets of points with side lengths. Gramsci can run in serial, OpenMP, MPI and hybrid parallel schemes. It is useful for performing domain decomposition of input catalogs, especially if the catalogs are large or the Rmax value is too large.

[ascl:1908.005] dips: Detrending periodic signals in timeseries

dips detrends timeseries of strictly periodic signals. It does not assume any functional form for the signal or the background or the noise; it disentangles the strictly periodic component from everything else. It has been used for detrending Kepler, K2 and TESS timeseries of periodic variable stars, eclipsing binary stars, and exoplanets.

[ascl:1908.006] GBKFIT: Galaxy kinematic modeling

GBKFIT performs galaxy kinematic modeling. It can be used to extract morphological and kinematical properties of galaxies by fitting models to spatially resolved kinematic data. The software can also take beam smearing into account by using the knowledge of the line and point spread functions. GBKFIT can take advantage of many-core and massively parallel architectures such as multi-core CPUs and Graphics Processing Units (GPUs), making it suitable for modeling large-scale surveys of thousands of galaxies within a very seasonable time frame. GBKFIT features an extensible object-oriented architecture that supports arbitrary models and optimization techniques in the form of modules; users can write custom modules without modifying GBKFIT’s source code. The software is written in C++ and conforms to the latest ISO standards.

[ascl:1908.007] MosfireDRP: MOSFIRE Data Reduction Pipeline

MosfireDRP reduces data from the MOSFIRE spectrograph of the Keck Observatory; it produces flat-fielded, wavelength calibrated, rectified, and stacked 2D spectrograms for each slit on a given mask in nearly real time. Background subtraction is performed in two states: a simple pairwise subtraction of interleaved stacks, and then fitting a 2D b-spline model to the background residuals.

[ascl:1908.008] TRISTAN-MP: TRIdimensional STANford - Massively Parallel code

TRISTAN-MP is a fully relativistic Particle-In-Cell (PIC) code for plasma physics computations and self-consistently solves the full set of Maxwell’s equations, along with the relativistic equations of motion for the charged particles. Fields are discretized on a finite 3D or 2D mesh, the computational grid; the code then uses time-centered and space-centered finite difference schemes to advance the equations in time via the Lorentz force equation, and to calculate spatial derivatives, so that the algorithm is second order accurate in space and time. The charges and currents derived from the particles' velocities and positions are then used as source terms to re-calculate the electromagnetic fields. TRISTAN-MP is based on the original TRISTAN code by O. Buneman (1999).

[ascl:1908.009] PyRADS: Python RADiation model for planetary atmosphereS

The 1D radiation code PyRADS provides line-by-line spectral resolution. For Earth-like atmospheres, PyRADS currently uses HITRAN 2016 line lists and the MTCKD continuum model. A version for shortwave radiation (scattering) is also available.

[ascl:1908.010] SNAPDRAGONS: Stellar Numbers And Parameters Determined Routinely And Generated Observing N-body Systems

SNAPDRAGONS (Stellar Numbers And Parameters Determined Routinely And Generated Observing N-body Systems) is a simplified version of the population synthesis code Galaxia (ascl:1101.007), using a different process to generate the stellar catalog. It splits each N-body particle from the galaxy simulation into an appropriate number of stellar particles to create a mock catalog of observable stars from the N-body model. SNAPDRAGON uses the same isochrones and extinction map as Galaxia.

[ascl:1908.011] NuRadioMC: Monte Carlo simulation package for radio neutrino detectors

NuRadioMC simulates ultra-high energy neutrino detectors that rely on the radio detection method, which exploits the radio emission generated in the electromagnetic component of a particle shower following a neutrino interaction. The code simulates the neutrino interaction in a medium, subsequent Askaryan radio emission, propagation of the radio signal to the detector and the detector response. NuRadioMC is a Monte Carlo framework that combines flexibility in detector design with user-friendliness. It includes an event generator, improved modeling of the radio emission, a revisited approach to signal propagation, and increased flexibility and precision in the detector simulation.

[ascl:1908.012] oscode: Oscillatory ordinary differential equation solver

oscode solves oscillatory ordinary differential equations efficiently. It is designed to deal with equations of the form x¨(t)+2γ(t)x˙(t)+ω2(t)x(t)=0, where γ(t) and ω(t) can be given as explicit functions or sequence containers (Eigen::Vectors, arrays, std::vectors, lists) in C++ or as numpy.arrays in Python. oscode makes use of an analytic approximation of x(t) embedded in a stepping procedure to skip over long regions of oscillations, giving a reduction in computing time. The approximation is valid when the frequency changes slowly relative to the timescales of integration, it is therefore worth applying when this condition holds for at least some part of the integration range.

[ascl:1908.013] BEAST: Bayesian Extinction And Stellar Tool

BEAST (Bayesian Extinction and Stellar Tool) fits the ultraviolet to near-infrared photometric SEDs of stars to extract stellar and dust extinction parameters. The stellar parameters are age (t), mass (M), metallicity (M), and distance (d). The dust extinction parameters are dust column (Av), average grain size (Rv), and mixing between type A and B extinction curves (fA).

[ascl:1908.014] Vlasiator: Hybrid-Vlasov simulation code

Vlasiator is a 6-dimensional Vlasov theory-based simulation. It simulates the entire near-Earth space at a global scale using the kinetic hybrid-Vlasov approach, to study fundamental plasma processes (reconnection, particle acceleration, shocks), and to gain a deeper understanding of space weather.

[ascl:1908.015] Analysator: Quantitative analysis of Vlasiator files

Analysator analyzes vlsv files produced by Vlasiator (ascl:1908.014). The code facilitates studies of particle paths, pitch angle distributions, velocity distributions, and more. It can read and write VLSV files and do calculations with the data, plot the real space from VLSV files with Mayavi (ascl:1205.008), and plot the velocity space (both blocks and iso surface) from VLSV files. It can also take cut-throughs, pitch angle distributions, gyrophase angle, and 3d slices, plot variables with sub plots in a clean format, and fit 1D polynomials to data.

[ascl:1908.016] DustCharge: Charge distribution for a dust grain

DustCharge calculates the equilibrium charge distribution for a dust grain of a given size and composition, depending on the local interstellar medium conditions, such as density, temperature, ionization fraction, local radiation field strength, and cosmic ray ionization fraction.

[ascl:1908.017] JPLephem: Jet Propulsion Lab ephemerides package

JPLephem loads and uses standard Jet Propulsion Laboratory (JPL) ephemerides for predicting the position and velocity of a planet or other Solar System body. It is one of the foundations of the Skyfield (ascl:1907.024) astronomy library for Python, and can also be used as a standalone package to generate raw vectors.

[ascl:1908.018] EBAI: Eclipsing Binaries with Artificial Intelligence

Eclipsing Binaries via Artificial Intelligence (EBAI) automates the process of solving light curves of eclipsing binary stars. EBAI is based on the back-propagating neural network paradigm and is highly flexible in construction of neural networks. EBAI comes in two flavors, serial (ebai) and multi-processor (ebai.mpi), and can be run in training, continued training, and recognition mode.

[ascl:1908.019] MAESTROeX: Low Mach number stellar hydrodynamics code

MAESTROeX solves the equations of low Mach number hydrodynamics for stratified atmospheres or stars with a general equation of state. It includes reactions and thermal diffusion and can be used on anything from a single core to 100,000s of processor cores with MPI + OpenMP. MAESTROeX maintains the accuracy of its predecessor MAESTRO (ascl:1010.044) while taking advantage of a simplified temporal integration scheme and leveraging the AMReX software framework for block-structured adaptive mesh refinement (AMR) applications.

[ascl:1908.020] QLF: Luminosity function analysis code

QLF derives full posterior distributions for and analyzes luminosity functions models; it also models hydrogen and helium reionization. Used with the included homogenized data, the derived luminosity functions can be easily compared with theoretical models or future data sets.

[ascl:1908.021] bias_emulator: Halo bias emulator

bias_emulator models the clustering of halos on large scales. It incorporates the cosmological dependence of the bias beyond the mapping of halo mass to peak height. Precise measurements of the halo bias in the simulations are interpolated across cosmological parameter space to obtain the halo bias at any point in parameter space within the simulation cloud. A tool to produce realizations of correlated noise for propagating the modeling uncertainty into error budgets that use the emulator is also provided.

[ascl:1908.022] YMW16: Electron-density model

YMW16 models the distribution of free electrons in the Galaxy, the Magellanic Clouds and the inter-galactic medium and can be used to estimate distances for real or simulated pulsars and fast radio bursts (FRBs) based on their position and dispersion measure. The Galactic model is based on 189 pulsars that have independently determined distances as well as dispersion measures, whereas simpler models are used for the electron density in the MC and the IGM.

[ascl:1908.023] FIRST Classifier: Automated compact and extended radio sources classifier

FIRST Classifier is an on-line system for automated classification of compact and extended radio sources. It is developed based on a trained Deep Convolutional Neural Network Model to automate the morphological classification of compact and extended radio sources observed in the FIRST radio survey. FIRST Classifier is able to predict the morphological class for a single source or for a list of sources as Compact or Extended (FRI, FRII and BENT).

[ascl:1908.024] PYSAT: Python Satellite Data Analysis Toolkit

The Python Satellite Data Analysis Toolkit (pysat) provides a simple and flexible interface for downloading, loading, cleaning, managing, processing, and analyzing space science data. The toolkit supports in situ satellite observations and many different types of ground- and space-based measurements. Its analysis routines are independent of instrument and data source.

[ascl:1908.025] FastCSWT: Fast directional Continuous Spherical Wavelet Transform

FastCSWT performs a directional continuous wavelet transform on the sphere. The transform is based on the construction of the continuous spherical wavelet transform (CSWT) developed by Antoine and Vandergheynst (1999). A fast implementation of the CSWT (based on the fast spherical convolution developed by Wandelt and Gorski 2001) is also provided.

[ascl:1909.001] Auto-multithresh: Automated masking for clean

Auto-multithresh implements an automated masking algorithm for clean. It operates on the residual image within the minor cycle of clean to identify and mask regions of significant emission. It then cascades these significant regions down to lower signal to noise. It includes features to pad the mask to avoid sharp edges and to remove small regions that are unlikely to be significant emission. The algorithm described by this code was incorporated into the tclean task within CASA as auto-multithresh.

[ascl:1909.002] MultiColorFits: Colorize and combine multiple fits images for visually aesthetic scientific plots

MultiColorFits is a tool to colorize and combine multiple fits images for making visually aesthetic scientific plots. The standard method to make color composites by combining fits images programmatically in python is to assign three images as separate red, green, and blue channels. This can produce unsatisfactory results for a variety of reasons, such as when less than three images are available, or additional images are desired to be shown. MultiColorFits breaks these limitations by allowing users to apply any color to a given image, not just red, green, or blue. Composites can then be created from an arbitrary number of images. Controls are included for stretching brightness scales with common functions.

[ascl:1909.003] SecularMultiple: Hierarchical multiple system secular evolution model

SecularMultiple computes the secular (orbit-averaged) gravitational dynamics of hierarchical multiple systems composed of nested binary orbits (simplex-type systems) with any configuration and any number of bodies. A particle can represent a binary or a body. The structure of the system is determined by linking to other particles with the attributes child1 and child2, and tidal interactions and relativistic corrections are included in an ad hoc fashion. SecularMultiple also includes routines for external perturbations such as flybys and supernovae.

[ascl:1909.004] TPI: Test Particle Integrator

TPI computes the gravitational dynamics of particles orbiting a supermassive black hole (SBH). A distinction is made to two types of particles: test particles and field particles. Field particles are assumed to move in quasi-static Keplerian orbits around the SBH that precess due to the enclosed mass (Newtonian 'mass precession') and relativistic effects. Otherwise, field-particle-field-particle interactions are neglected. Test particles are integrated in the time-dependent potential of the field particles and the SBH. Relativistic effects are included in the equations of motion (including the effects of SBH spin), and test-particle-test-particle interactions are neglected.

[ascl:1909.005] HADES: Hexadecapolar Analysis for Dust Estimation in Simulations (of CMB B-mode thermal dust emission)

HADES analyzse dust levels in simulated CMB galactic dust maps with realistic experimental noise and lensing configurations. It allows detection of dust via its anisotropy properties in CMB B-modes. It also includes techniques for computing null-tests and a rudimentary technique for dedusting.

[ascl:1909.006] ChempyMulti: Multi-star Bayesian inference with Chempy

ChempyMulti is an update to Chempy (ascl:1702.011) and provides yield table scoring and multi-star Bayesian inference. This replaces the ChempyScoring package in Chempy. Chempy is a flexible one-zone open-box chemical evolution model, incorporating abundance fitting and stellar feedback calculations. It includes routines for parameter optimization for simulations and observational data and yield table scoring.

[ascl:1909.007] EBHLIGHT: General relativistic radiation magnetohydrodynamics with Monte Carlo transport

EBHLIGHT solves the equations of general relativistic radiation magnetohydrodynamics in stationary spacetimes. Fluid integration is performed with the second order shock-capturing scheme HARM (ascl:1209.005) and frequency-dependent radiation transport is performed with the second order Monte Carlo code grmonty (ascl:1306.002). Fluid and radiation exchange four-momentum in an explicit first-order operator-split fashion.

[ascl:1909.008] RascalC: Fast code for galaxy covariance matrix estimation

RascalC quickly estimates covariance matrices from two- or three-point galaxy correlation functions. Given an input set of random particle locations and a two-point correlation function (or input set of galaxy positions), RascalC produces an estimate of the associated covariance for a given binning strategy, with non-Gaussianities approximated by a ‘shot-noise-rescaling’ parameter. For the 2PCF, the rescaling parameter can be calibrated by dividing the particles into jackknife regions and comparing sample to theoretical jackknife covariance. RascalC can also be used to compute Legendre-binned covariances and cross-covariances between different two-point correlation functions.

[ascl:1909.009] CLOVER: Convolutional neural network spectra identifier and kinematics predictor

CLOVER (Convnet Line-fitting Of Velocities in Emission-line Regions) is a convolutional neural network (ConvNet) trained to identify spectra with two velocity components along the line of sight and predict their kinematics. It works with Gaussian emission lines (e.g., CO) and lines with hyperfine structure (e.g., NH3). CLOVER has two prediction steps, classification and parameter prediction. For the first step, CLOVER segments the pixels in an input data cube into one of three classes: noise (i.e., no emission), one-component (emission line with single velocity component), and two-component (emission line with two velocity components). For the pixels identified as two-components in the first step, a second regression ConvNet is used to predict centroid velocity, velocity dispersion, and peak intensity for each velocity component.

[ascl:1909.010] AREPO: Cosmological magnetohydrodynamical moving-mesh simulation code

AREPO is a massively parallel gravity and magnetohydrodynamics code for astrophysics, designed for problems of large dynamic range. It employs a finite-volume approach to discretize the equations of hydrodynamics on a moving Voronoi mesh, and a tree-particle-mesh method for gravitational interactions. AREPO is originally optimized for cosmological simulations of structure formation, but has also been used in many other applications in astrophysics.

[ascl:1909.011] WVTICs: SPH initial conditions using Weighted Voronoi Tesselations

WVTICs generates glass-like initial conditions for Smoothed Particle Hydrodynamics. Relaxation of the particle distribution is done using an algorithm based on Weighted Voronoi Tesselations; additional particle reshuffling can be enabled to improve over- and undersampled maxima/minima. The WBTICs package includes a full suite of analytical test problems.

[ascl:1909.012] HISS: HI spectra stacker

HISS stacks HI (emission and absorption) spectra in a consistent and reliable manner to enable statistical analysis of average HI properties. It provides plots of the stacked spectrum and reference spectrum with any fitted function, of the stacked noise response, and of the distribution of the integrated fluxes when calculating the uncertainties. It also produces a table containing the integrated flux calculated from the fitted functions and the stacked spectrum, among other output files.

[ascl:1909.013] EPOS: Exoplanet Population Observation Simulator

EPOS (Exoplanet Population Observation Simulator) simulates observations of exoplanet populations. It provides an interface between planet formation simulations and exoplanet surveys such as Kepler. EPOS can also be used to estimate planet occurrence rates and the orbital architectures of planetary systems.

[ascl:1909.014] fgivenx: Functional posterior plotter

fgivenx plots a predictive posterior of a function, dependent on sampled parameters, for a Bayesian posterior Post(theta|D,M) described by a set of posterior samples {theta_i}~Post. If there is a function parameterized by theta y=f(x;theta), this script produces a contour plot of the conditional posterior P(y|x,D,M) in the (x,y) plane.

[submitted] Magnetizer: computing magnetic fields of evolving galaxies

Computes time and radial dependent magnetic fields for a sample of galaxies in the output of a semi-analytic model of galaxy formation. The magnetic field is obtained by numerically solving the galactic dynamo equations throughout history of each galaxy. Stokes parameters and Faraday rotation measure can also be computed along a random line-of-sight for each galaxy.

[ascl:1910.001] PINK: Parallelized rotation and flipping INvariant Kohonen maps

Morphological classification is one of the most demanding challenges in astronomy. With the advent of all-sky surveys, an enormous amount of imaging data is publicly available, and are typically analyzed by experts or encouraged amateur volunteers. For upcoming surveys with billions of objects, however, such an approach is not feasible anymore. PINK (Parallelized rotation and flipping INvariant Kohonen maps) is a simple yet effective variant of a rotation-invariant self-organizing map that is suitable for many analysis tasks in astronomy. The code reduces the computational complexity via modern GPUs and applies the resulting framework to galaxy data for morphological analysis.

[ascl:1910.002] PreProFit: Pressure Profile Fitter for galaxy clusters in Python

PreProFit fits the pressure profile of galaxy clusters using Markov chain Monte Carlo (MCMC). The software can analyze data from different sources and offers flexible parametrization for the pressure profile. PreProFit accounts for Abel integral, beam smearing, and transfer function filtering when fitting data and returns χ2, model parameters and uncertainties in addition to marginal and joint probability contours, diagnostic plots, and surface brightness radial profiles. The code can be used for analytic approximations for the beam and transfer functions for feasibility studies.

[ascl:1910.003] a3cosmos-gas-evolution: Galaxy cold molecular gas evolution functions

a3cosmos-gas-evolution calculates galaxies' cold molecular gas properties using gas scaling functions derived from the A3COSMOS project. By known galaxies' redshifts or cosmic age, stellar masses, and star formation enhancement to galaxies' star-forming main sequence (Delta MS), the gas scaling functions predict their stellar mass ratio (gas fraction) and gas depletion time.

[ascl:1910.004] DM_phase: Algorithm for correcting dispersion of radio signals

DM_phase maximizes the coherent power of a radio signal instead of its intensity to calculate the best dispersion measure (DM) for a burst such as those emitted by pulsars and fast radio bursts (FRBs). It is robust to complex burst structures and interference, thus mitigating the limitations of traditional methods that search for the best DM value of a source by maximizing the signal-to-noise ratio (S/N) of the detected signal.

[ascl:1910.005] exoplanet: Probabilistic modeling of transit or radial velocity observations of exoplanets

exoplanet is a toolkit for probabilistic modeling of transit and/or radial velocity observations of exoplanets and other astronomical time series using PyMC3 (ascl:1610.016), a flexible and high-performance model building language and inference engine. exoplanet extends PyMC3's language to support many of the custom functions and distributions required when fitting exoplanet datasets. These features include a fast and robust solver for Kepler's equation; scalable Gaussian processes using celerite (ascl:1709.008); and fast and accurate limb darkened light curves using the code starry (ascl:1810.005). It also offers common reparameterizations for limb darkening parameters, and planet radius and impact parameters.

[ascl:1910.006] EMERGE: Empirical ModEl for the foRmation of GalaxiEs

Emerge (Empirical ModEl for the foRmation of GalaxiEs) populates dark matter halo merger trees with galaxies using simple empirical relations between galaxy and halo properties. For each model represented by a set of parameters, it computes a mock universe, which it then compares to observed statistical data to obtain a likelihood. Parameter space can be explored with several advanced stochastic algorithms such as MCMC to find the models that are in agreement with the observations.

[ascl:1910.007] TLS: Transit Least Squares

TLS is an optimized transit-fitting algorithm to search for periodic transits of small planets. In contrast to BLS: Box Least Squares (ascl:1607.008), which searches for rectangular signals in stellar light curves, TLS searches for transit-like features with stellar limb-darkening and including the effects of planetary ingress and egress. TLS also analyses the entire, unbinned data of the phase-folded light curve. TLS yields a ~10% higher detection efficiency (and similar false alarm rates) compared to BLS though has a higher computational load. This load is partly compensated for by applying an optimized period sampling and transit duration sampling constrained to the physically plausible range.

[ascl:1910.008] ECLIPS3D: Linear wave and circulation calculations

ECLIPS3D (Eigenvectors, Circulation, and Linear Instabilities for Planetary Science in 3 Dimensions) calculates a posteriori energy equations for the study of linear processes in planetary atmospheres with an arbitrary steady state, and provides both increased robustness and physical meaning to the obtained eigenmodes. It was developed originally for planetary atmospheres and includes python scripts for data analysis. ECLIPS3D can be used to study the initial spin up of superrotation of GCM simulations of hot Jupiters in addition to being applied to other problems.

[ascl:1910.009] orbitize: Orbit-fitting for directly imaged objects

orbitize fits the orbits of directly-imaged objects by packaging the Orbits for the Impatient (OFTI) algorithm and a parallel-tempered Markov Chain Monte Carlo (MCMC) algorithm into a consistent API. It accepts observations in three measurement formats, which can be mixed in the same input file, generates orbits, and plots the computed orbital parameters. orbitize offers numerous ways to visualize the data, including histograms, corner plots, and orbit plots. Generated orbits can be saved in HDF5 format for future use and analysis.

[ascl:1910.010] PEXO: Precise EXOplanetology

PEXO provides a global modeling framework for ns timing, μas astrometry, and μm/s radial velocities. It can account for binary motion and stellar reflex motions induced by planetary companions and also treat various relativistic effects both in the Solar System and in the target system (Roemer, Shapiro, and Einstein delays). PEXO is able to model timing to a precision of 1 ns, astrometry to a precision of 1 μas, and radial velocity to a precision of 1 μm/s.

[ascl:1910.011] LEO-Py: Likelihood Estimation of Observational data with Python

LEO-Py uses a novel technique to compute the likelihood function for data sets with uncertain, missing, censored, and correlated values. It uses Gaussian copulas to decouple the correlation structure of variables and their marginal distributions to compute likelihood functions, thus mitigating inconsistent parameter estimates and accounting for non-normal distributions in variables of interest or their errors.

[ascl:1910.012] AOTOOLS: Reduce IR images from Adaptive Optics

AOTOOLS reduces IR images from adaptive optics. It uses effective dithering, either sky subtraction or dark-subtration, and flat-fielding techniques to determine the effect of the instrument on an image of an object. It also performs bad pixel masking, degrades an AO on-axis PSF due to effects of anisoplanicity, and corrects an AO on-axis PSF due to effects of seeing.

[ascl:1910.013] E0102-VR: Virtual Reality application to visualize the optical ejecta in SNR 1E 0102.2-7219

E0102-VR facilitates the characterization of the 3D structure of the oxygen-rich optical ejecta in the young supernova remnant 1E 0102.2-7219 in the Small Magellanic Cloud. This room-scale Virtual Reality application written for the HTC Vive contributes to the exploration of the scientific potential of this technology for the field of observational astrophysics.

[ascl:1910.014] ANNz2: Estimating photometric redshift and probability density functions using machine learning methods

ANNz2, a newer implementation of ANNz (ascl:1209.009), utilizes multiple machine learning methods such as artificial neural networks, boosted decision/regression trees and k-nearest neighbors to measure photo-zs based on limited spectral data. The code dynamically optimizes the performance of the photo-z estimation and properly derives the associated uncertainties. In addition to single-value solutions, ANNz2 also generates full probability density functions (PDFs) in two different ways. In addition, estimators are incorporated to mitigate possible problems of spectroscopic training samples which are not representative or are incomplete. ANNz2 is also adapted to provide optimized solutions to general classification problems, such as star/galaxy separation.

[ascl:1910.015] MarsLux: Illumination Mars maps generator

MarsLux generates illumination maps of Mars from Digital Terrain Model (DTM), permitting users to investigate in detail the illumination conditions on Mars based on its topography and the relative position of the Sun. MarsLux consists of two Python codes, SolaPar and MarsLux. SolaPar calculates the matrix with solar parameters for one date or a range between the two. The Marslux code generates the illumination maps using the same DTM and the files generated by SolaPar. The resulting illumination maps show areas that are fully illuminated, areas in total shadow, and areas with partial shade, and can be used for geomorphological studies to examine gullies, thermal weathering, or mass wasting processes as well as for producing energy budget maps for future exploration missions.

[ascl:1910.016] MiSTree: Construct and analyze Minimum Spanning Tree graphs

MiSTree quickly constructs minimum spanning tree graphs for various coordinate systems, including Celestial coordinates, by using a k-nearest neighbor graph (k NN, rather than a matrix of pairwise distances) which is then fed to Kruskal's algorithm to create the graph. MiSTree bins the MST statistics into histograms and plots the distributions; enabling the inclusion of high-order statistics information from the cosmic web to provide additional information that improves cosmological parameter constraints. Though MiSTree was designed for use in cosmology, it can be used in any field requiring extracting non-Gaussian information from point distributions.

[ascl:1910.017] ChainConsumer: Corner plots, LaTeX tables and plotting walks

ChainConsumer consumes the chains output from Monte Carlo processes such as MCMC to produce plots of the posterior surface inferred from the chain distributions, to plot the chains as walks to check for mixing and convergence, and to output parameter summaries in the form of LaTeX tables. It handles multiple models (chains), allowing for model comparison using AIC, BIC or DIC metrics.

[ascl:1910.018] GetDist: Monte Carlo sample analyzer

GetDist analyzes Monte Carlo samples, including correlated samples from Markov Chain Monte Carlo (MCMC). It offers a point and click GUI for selecting chain files, viewing plots, marginalized constraints, and LaTeX tables, and includes a plotting library for making custom publication-ready 1D, 2D, 3D-scatter, triangle and other plots. Its convergence diagnostics include correlation length and diagonalized Gelman-Rubin statistics, and the optimized kernel density estimation provides an automated optimal bandwidth choice for 1D and 2D densities with boundary and bias correction. It is available as a standalong package and with CosmoMC (ascl:1106.025).

[ascl:1910.019] Cobaya: Bayesian analysis in cosmology

Cobaya (Code for BAYesian Analysis) provides a framework for sampling and statistical modeling and enables exploration of an arbitrary prior or posterior using a range of Monte Carlo samplers, including the advanced MCMC sampler from CosmoMC (ascl:1106.025) and the advanced nested sampler PolyChord (ascl:1502.011). The results of the sampling can be analyzed with GetDist (ascl:1910.018). It supports MPI parallelization and is highly extensible, allowing the user to define priors and likelihoods and create new parameters as functions of other parameters.

It includes interfaces to the cosmological theory codes CAMB (ascl:1102.026) and CLASS (ascl:1106.020) and likelihoods of cosmological experiments, such as Planck, Bicep-Keck, and SDSS. Automatic installers are included for those external modules; Cobaya can also be used as a wrapper for cosmological models and likelihoods, and integrated it in other samplers and pipelines. The interfaces to most cosmological likelihoods are agnostic as to which theory code is used to compute the observables, which facilitates comparison between those codes. Those interfaces are also parameter-agnostic, allowing use of modified versions of theory codes and likelihoods without additional editing of Cobaya’s source.

[ascl:1910.020] OCD: O'Connell Effect Detector using push-pull learning

OCD (O'Connell Effect Detector) detects eclipsing binaries that demonstrate the O'Connell Effect. This time-domain signature extraction methodology uses a supporting supervised pattern detection algorithm. The methodology maps stellar variable observations (time-domain data) to a new representation known as Distribution Fields (DF), the properties of which enable efficient handling of issues such as irregular sampling and multiple values per time instance. Using this representation, the code applies a metric learning technique directly on the DF space capable of specifically identifying the stars of interest; the metric is tuned on a set of labeled eclipsing binary data from the Kepler survey, targeting particular systems exhibiting the O’Connell Effect. This code is useful for large-scale data volumes such as that expected from next generation telescopes such as LSST.

[ascl:1910.021] AOtools: Adaptive optics modeling and analysis toolkit

The AOtools package offers generic adaptive optics processing tools in addition to astronomy-specific methods; among these are analyzing data in the pupil plane, images and point spread functions in the focal plane, wavefront sensors, modeling of atmospheric turbulence, physical optical propagation of wavefronts, and conversion functions to convert stellar brightness into photon flux for a given waveband. The software also calculates integrated atmospheric parameters, such as coherence time and isoplanatic angle from atmospheric turbulence and wind speed profile.

[ascl:1910.022] qnm: Kerr quasinormal modes, separation constants, and spherical-spheroidal mixing coefficients calculator

qnm computes the Kerr quasinormal mode frequencies, angular separation constants, and spherical-spheroidal mixing coefficients. The qnm package includes a Leaver solver with the Cook-Zalutskiy spectral approach to the angular sector, and a caching mechanism to avoid repeating calculations. A large cache of low ℓ, m, n modes is available for download and can be installed with a single function call and interpolated to provide good initial guess for root-polishing at new values of spin.

[submitted] Network Flux Transport Demonstration

We have developed a method to efficiently simulate the dynamics of the magnetic flux in the solar network. We call this method Network Flux Transport (NFT). Implemented using a Spherical Centroidal Voronoi Tessellation (SCVT) based network model, magnetic flux is advected by photospheric plasma velocity fields according to the geometry of the SCVT model. We test NFT by simulating the magnetism of the Solar poles. The poles of the sun above 55 deg latitude are free from flux emergence from active regions or ephemeral regions. As such, they are ideal targets for a simplified simulation that relies on the strengths of the NFT model. This simulation method reproduces the magnetic and spatial distributions for the solar poles over two full solar cycles.

[ascl:1911.001] PLAN: A Clump-finder for Planetesimal Formation Simulations

PLAN (PLanetesimal ANalyzer) identifies and characterizes planetesimals produced in numerical simulations of the Streaming Instability that includes particle self-gravity with code Athena (ascl:1010.014). PLAN works with the 3D particle output of Athena and finds gravitationally bound clumps robustly and efficiently. PLAN — written in C++ with OpenMP/MPI — is massively parallelized, memory-efficient, and scalable to analyze billions of particles and multiple snapshots simultaneously. The approach of PLAN is based on the dark matter halo finder HOP (ascl:1102.019), but with many customizations for planetesimal formation. PLAN can be easily adapted to analyze other object formation simulations that use Lagrangian particles (e.g., Athena++ simulations). PLAN is also equipped with a toolkit to analyze the grid-based hydro data (VTK dumps of primitive variables) from Athena, which requires the Boost MultiDimensional Array Library.

[ascl:1911.002] uvplot: Interferometric visibilities plotter

uvplot makes nice plots of deprojected interferometric visibilities (often called uvplots). It implements plotting functionality, handles MS tables with spectral windows with different number of channels, and can import visibilities from ASCII to MS Table. It also allows export of specific channels. uvplot can be installed inside the NRAO CASA package (ascl:1107.013).

[ascl:1911.003] OpenSPH: Astrophysical SPH and N-body simulations and interactive visualization tools

OpenSPH runs hydrodynamical and N-body simulations and was written for asteroid collisions and subsequent gravitational evolution. The code offers SPH and N-body solvers with several different equations of state and material rheologies. It is written in C++14 with a modular object-oriented design, focused on extensibility and maintainability, and it can be used either as a library or as a standalone graphical program that allows to set up the problem in a convenient graphical node editor. The graphical program further allows real-time visualization of the simulation, diagnostics and tools for analysis of the results.

[ascl:1911.004] PypeIt: Python spectroscopic data reduction pipeline

PypeIt reduces data from echelle and low-resolution spectrometers; the code can be run in several modes of reduction that demark the level of sophistication (e.g. quick and dirty vs. MonteCarlo) and also the amount of output written to disk. It also generates numerous data products, including 1D and 2D spectra; calibration images, fits, and meta files; and quality assurance figures.

[ascl:1911.005] MARTINI: Mock spatially resolved spectral line observations of simulated galaxies

MARTINI (Mock APERTIF-like Radio Telescope Interferometry of the Neutal ISM) creates synthetic resolved HI line observations (data cubes) of smoothed-particle hydrodynamics simulations of galaxies. The various aspects of the mock-observing process are divided logically into sub-modules handling the data cube, source, beam, noise, spectral model and SPH kernel. MARTINI is object-oriented: each sub-module provides a class (or classes) which can be configured as desired. For most sub-modules, base classes are provided to allow for straightforward customization. Instances of each sub-module class are then given as parameters to the Martini class. A mock observation is then constructed by calling a handful of functions to execute the desired steps in the mock-observing process.

[ascl:1911.006] ATHOS: A Tool for HOmogenizing Stellar parameters

ATHOS provides on-the-fly stellar parameter determination of FGK stars based on flux ratios from optical spectra. Once configured properly, it will measure flux ratios in the input spectra and deduce the stellar parameters effective temperature, iron abundance (a.k.a [Fe/H]), and surface gravity by employing pre-defined analytical relations. ATHOS can be configured to run in parallel in an arbitrary number of threads, thus enabling the fast and efficient analysis of huge datasets.

[ascl:1911.007] planetplanet: General photodynamical code for exoplanet light curves

planetplanet models exoplanet transits, secondary eclipses, phase curves, and exomoons, as well as eclipsing binaries, circumbinary planets, and more. The code was originally developed to model planet-planet occultation (PPO) light curves for the TRAPPIST-1 system, but it is generally applicable to any exoplanet system. During a PPO, a planet occults (transits) the disk of another planet in the same planetary system, blocking its thermal (and reflected) light, which can be measured photometrically by a distant observer. planetplanet is coded in C and wrapped in a user-friendly Python interface.

[ascl:1911.008] HeatingRate: Radioactive heating rate and macronova (kilonova) light curve

HeatingRate calculates the nuclear heating rates [erg/s/g] of beta-decay, alpha-decay, and spontaneous fission of r-process nuclei, taking into account for thermalization of gamma-rays and charged decay products in r-process ejecta. It uses the half-lives and injection energy spectra from an evaluated nuclear data library (ENDF/B-VII.1). Each heating rate is computed for given abundances, ejecta mass, velocity, and density profile. HeatingRate also computes the bolometric light curve and the evolution of the effective temperature for given abundances, ejecta mass, velocity, and density profile assuming opacities independent of the wavelength.

[ascl:1911.009] frbpoppy: Fast radio burst population synthesis in Python

frbpoppy conducts fast radio burst population synthesis and continues the work of PSRPOP (ascl:1107.019) and PsrPopPy (ascl:1501.006) in the realm of FRBs. The code replicates observed FRB detection rates and FRB distributions in three steps. It first simulates a cosmic population of one-off FRBs and allows the user to select options such as models for source number density, cosmology, DM host/IGM/Milky Way, luminosity functions, and emission bands as well as maximum redshift and size of the FRB population. The code then generates a survey by adopting a beam pattern using various survey parameters, among them telescope gain, sampling time, receiver temperature, central frequency, channel bandwidth, number of polarizations, and survey region limits. Finally, frbpoppy convolves the generated intrinsic population with the generated survey to simulate an observed FRB population.

[ascl:1911.010] Fruitbat: Fast radio burst redshift estimation

Fruitbat estimates the redshift of Fast Radio Bursts (FRB) from their dispersion measure. The code combines various dispersion measure (DM) and redshift relations with the YMW16 galactic dispersion measure model into a single easy to use API.

[ascl:1911.011] IDG: Image Domain Gridding

IDG (Image Domain Gridding) is an imager that makes w-term corrections and a-term corrections computationally very cheap. It works with WSClean (ascl:1408.023) and supports the same cleaning and data selections options that WSClean offers in normal mode (such as cotton-schwab, multi-frequency multi-scale cleaning, and auto-masking). IDG also allows gridding with a time-variable beam including the LOFAR, AARTFAAC and MWA beam and can perform full beam or differential correction. The code requires measurement sets with four polarizations (e.g. XX/XY/YX/YY), can apply a spatially varying time-variable TEC term that can additionally be different for different antennas and output channels, and performs extremely well on GPUs.

[ascl:1911.012] Zeltron: Explicit 3D relativistic electromagnetic Particle-In-Cell code

Zeltron is an explicit 3D relativistic electromagnetic Particle-In-Cell code suited for modeling particle acceleration in astrophysical plasmas. The code is efficiently parallelized with the Message Passing Interface, and can be run on a laptop computer or on multiple cores on current supercomputers. Zeltron takes into account the effect of the radiation reaction force on the motion of the particles; it assigns variable weights to the macro-particles to model particle density gradients, and does not strictly conserve the total energy. The code uses linear interpolation to deposit the charges and currents generated by each particle at the nodes of the computational grid, and computes the charge and current densities for Maxwell's equations. Zeltron contains a large set of analysis tools, including plasma density, particle spectrum, optically thin synchrotron and inverse Compton spectra, angular distributions, and stress-energy tensor.

[ascl:1911.013] ATLAS: Turning Dopplergram images into frequency shift measurements

ATLAS performs the tracking, projecting, power-spectrum-making, and ring-fitting needed to turn a set of Dopplergram images into a set of frequency shift measurements. This code is essentially a combination of three codes, FRACK (FORTRAN Tracking), PSPEC (Power SPECtrum), and MRF (Multi-Ridge Fitting), included in the ATLAS package. ATLAS reads in a list of longitude/latitude coordinates corresponding to the desired tile centers and a set of full-disk Dopplergram images and outputs frequency shift measurements from each wave mode of each tile. The code relies on both distributed-memory (MPI) and shared-memory (OpenMP) parallelism to scale up to around 1000 processes. Due to the immense volume of data produced by the tracking and projecting steps, the intermediate data products (tiles, power spectra) are never written out.

[ascl:1911.014] MORDI: Massively-Overlapped Ring-Diagram Inversion

MORDI (Massively-Overlapped Ring-Diagram Inversion) performs three-dimensional ring-diagram inversions. The code reads in frequency shift measurements and their associated sensitivity kernels and outputs two-dimensional slices of the subsurface flow field at a constant depth and (optionally) the associated averaging kernels. It relies on both distributed-memory (MPI) and shared-memory (OpenMP) parallelism to scale efficiently up to a few thousand processors, but can also run reasonably well on small machines (1-4 cpus). The actions of the code are modified by command-line parameters, which enable a significant amount of flexibility when setting up an inversion.

[ascl:1911.015] LATTICEEASY: Lattice simulator for evolving interacting scalar fields in an expanding universe

LATTICEEASY creates lattice simulations of the evolution of interacting scalar fields in an expanding universe. The program can do runs with different parameters and new models can be easily introduced for evaluation. Simulations can be done in one, two, or three dimensions by resetting a single variable. Mathematica notebooks for plotting the output and a range of models are also available for download; a parallel processing version of LATTICEEASY called CLUSTEREASY (ascl:1911.016) is also available.

[ascl:1911.016] CLUSTEREASY: Lattice simulator for evolving interacting scalar fields in an expanding universe on parallel computing clusters

CLUSTEREASY is a parallel programming extension of the simulation program LATTICEEASY (ascl:1911.015); running the program in parallel greatly extends the range of scales and times that can be simulated. The program is particularly useful for the study of reheating and thermalization after inflation.

[ascl:1911.017] HLattice: Scalar fields and gravity simulator for the early universe

HLattice simulates scalar fields and gravity in the early universe. The code allows the user to select between symplectic integrators, descretization schemes, and metrics such as Minkowski or FRW backgrounds and adaptice schemes in an "all-in-one" configuration file.

[ascl:1911.018] WhereWolf: Galaxy/(sub)Halo ghosting tool for N-body simulations

WhereWolf tracks (sub)haloes even if they have been lost by a halo finder in cosmological simulations and supplements halo catalogs such as VELOCIraptor (ascl:1911.020) with these recovered (sub)haloes. The code can improve measurements of the subhalo/halo mass function and present estimates of the distribution of radii at which subhaloes merge.

[ascl:1911.019] OrbWeaver: Galaxy/(sub)halo orbital processing tool

OrbWeaver extracts orbits from halo catalogs, enabling large statistical studies of their orbital parameters. The code is run in two stages. For the first run, a configuration file is used to modify orbit host selection and the region around orbit host used for the superset of orbiting halos. Each orbit host has a orbit forest (containing halos that passed within the region of interest); the code generates a pre-processed catalog which contains a superset of orbiting halo for each identified orbit host. The second run uses the file list generated in the first stage for the creation of the orbit catalog, which is the final output.

[ascl:1911.020] VELOCIraptor-STF: Six-dimensional Friends-of-Friends phase space halo finder

VELOCIraptor-STF, formerly STructure Finder (ascl:1306.009), is a 6-Dimensional Friends-of-Friends (6D-FoF) phase space halo finder and constructs halo catalogs. The code uses using MPI and OpenMP APIs and can be compiled as a library for on-the-fly halo finding within an N-body/hydrodynamnical code. There is an associated halo merger tree code TreeFrog (ascl:1911.021).

[ascl:1911.021] TreeFrog: Construct halo merger trees and compare halo catalogs

TreeFrog reads in particle IDs information between various structure catalogs and cross matches catalogs, assuming that particle IDs are unique and constant across snapshots. Though it is built as a cross correlator (in that it can match particles across several different catalogs), its principle use is as halo merger tree builder. TreeFrog produces links between objects found at different snapshots (or catalogs) and uses several possible functions to evaluate the merit of a link between one object at a given snapshot (or in a given catalog) to another object in a previous snapshot (or different catalog). It can also produce a full graph. The code utilizes MPI and OpenMP. It is optimzed for reading VELOCIraptor (ascl:1911.020) output but can also read output from other structure finders such as AHF (ascl:1102.009).

[ascl:1911.022] FFTLog-and-beyond: Generalized FFTLog algorithm

FFTLog-and-beyond takes the FFTLog algorithm for single-Bessel integrals and generalizes it for integrals containing a derivative of the Bessel function to solve the non-Limber integrals. The full non-Limber angular power spectrum integral is simplified by noting the small contribution from unequal-time nonlinear terms; this significantly reduces the computation and avoids the double-Bessel integral. The original FFTLog algorithm is also extended to compute integrals containing derivatives of Bessel functions, which can be used to efficiently compute angular power spectra including redshift-space distortions (RSD) and Doppler effects. C and Python versions of the code are available.

[ascl:1911.023] miluphcuda: Smooth particle hydrodynamics code

miluphcuda is the CUDA port of the original miluph code; it runs on single Nvidia GPUs with compute capability 5.0 and higher and provides fast and efficient computation. The code can be used for hydrodynamical simulations and collision and impact physics, and features self-gravity via Barnes-Hut trees and porosity models such as P-alpha and epsilon-alpha. It can model solid bodies, including ductile and brittle materials, as well as non-viscous fluids, granular media, and porous continua.

[ascl:1911.024] comb: Spectral line data reduction and analysis package

comb is a single-dish radio astronomy spectral line data reduction and analysis package developed at AT&T Bell labs and was used for data reduction for many single-dish telescopes, including Bell Labs 7-m, NRAO 12-m, DSN network, FCRAO 14-m, Arecibo, AST/RO, SEST, BIMA, and in 2011-2012, the Stratospheric Terahertz Observatory. A cookbook for the code is available.

[ascl:1912.001] Polyspectrum: Computing polyspectra using an FFT estimator

Polyspectrum computes the polyspectrum from 3D grids using a fast Fourier transformation (FFT) estimator. The code, written in C and MPI-parallelized, support the computation of power- and bispectra; it also supports higher-order polyspectra, but streamlining the input data is required.

[ascl:1912.002] casacore: Suite of C++ libraries for radio astronomy data processing

The casacore package contains the core libraries of the old AIPS++/CASA (ascl:1107.013) package. This split was made to get a better separation of core libraries and applications. CASA is now built on top of Casacore. The system consists of a set of layered libraries (packages) and includes a library (using Boost-Python) that converts the basic Casacore types (e.g., Array, Record) to and from Python.
Casacore includes the casa package for core functionality and data types like Array and Record; a scimath package for N-dim functions with auto-differentiation and linear or non-linear fitting; and a
tables package for the table data system supporting N-dim arrays with advanced querying. It also includes the measures package to manage values in astronomical reference frames using physical units (Quanta) and the MeasurementSets for storing data in the UV-domain, and also the images package for N-dim images in world coordinates with various analysis operations.

[ascl:1912.003] ASKAPsoft: ASKAP science data processor software

ASKAPsoft provides data processing functionality for Australian Square Kilometre Array Pathfinder, including calibration, spectral line imaging, continuum imaging, source detection and generation of source catalogs, and transient detection. The MPI-based package is the primary software for storing and processing raw data, and initiating the archiving of resulting science data products into the data archive (CASDA). The processing pipelines within ASKAPsoft are largely written in C++ built on top of casacore (ascl:1912.002) and other third party libraries.

[ascl:1912.004] DALiuGE: Data Activated Liu Graph Engine

DALiuGE provides a distributed data management platform and a scalable pipeline execution environment to support continuous, soft real-time, data-intensive processing for producing radio astronomy data products; it originated from a prototyping activity as part of the SKA SDP Consortium called Data Flow Management System (DFMS). Though the development of DALiuGE is largely based on radio astronomy processing requirements, it has adopted a generic, data-driven framework architecture potentially applicable to many other data-intensive applications.

[ascl:1912.005] Athena++: Radiation GR magnetohydrodynamics code

Athena++ is a complete re-write of the Athena astrophysical magnetohydrodynamics (MHD) code (ascl:1010.014) in C++. Compared to earlier versions, the Athena++ code has much more flexible coordinate and grid options and supports new physics. It also offers significantly improved performance and scalability, and improved source code clarity and modularity. Athena++ supports compressible hydrodynamics and MHD in 1D, 2D, and 3D, and special and general relativistic hydrodynamics and MHD. In addition, it supports Cartesian, cylindrical, or spherical polar coordinates; static or adaptive mesh refinement in any coordinate system; mixed parallelization with both OpenMP and MPI; and a task-based execution model for improved load balancing, scalability and modularity.

[ascl:1912.006] HSIM: HARMONI simulation pipeline

HSIM simulates observations with HARMONI on the Extremely Large Telescope. HSIM takes high spectral and spatial resolution input data cubes, encoding physical descriptions of astrophysical sources, and generates mock observed data cubes. The simulations incorporate detailed models of the sky, telescope, instrument, and detectors to produce realistic mock data. HSIM performs in-depth simulations for several key science cases as part of the design and development of the HARMONI integral field spectrograph, including the ELT AO performance, atmospheric effects and realistic detector statistics.

[ascl:1912.007] anesthetic: Nested sampling visualization

anesthetic brings together tools for processing nested sampling chains, leveraging standard scientific python libraries. The code provides computation of Bayesian evidences, Kullback-Liebler divergences and Bayesian model dimensionalities, marginalized 1d and 2d plots, and dynamic replaying of nested sampling. anesthetic was designed primarily for use with nested sampling outputs, although it can be used for normal MCMC chains.

[ascl:1912.008] PopSyCLE: Population Synthesis for Compact object Lensing Events

PopSyCLE performs compact object population synthesis while taking photometric and astrometric microlensing effects into consideration. It uses Galaxia (ascl:1101.007) to produces a synthetic survey, injects compact objects into the resulting survey, and then produces a list of microlensing events, enabling the discovery of black holes with microlensing. It can be used to examine historical microlensing events from photometric surveys to statistically constrain the abundance of black holes in our galaxy, and to forward model microlensing survey results to constrain, for example, the properties of compact objects, Galactic structure, and the initial-final mass relation.

[submitted] PyFOSC: a pipeline toolbox for BFOSC/YFOSC long-slit spectroscopy data reduction

PyFOSC is a pipeline toolbox for long-slit spectroscopy data reduction written in Python. It can be used for FOSC (Faint Object Spectrograph and Camera) data from Xinglong/Lijiang 2-meter telescopes in China. This pipeline privodes a neat way for data pre-processing, including updating missing header fileds for BFOSC data, reducing fits file extension for YFOSC data, etc. And it makes the data reduction procedure efficient by using previously identified lamp spectra as re-identification references during wavelength calibration, and applying multiprocessing in some modules. PyFOSC also enables customization for any other long-slit spectroscopy data.

[submitted] Exo-MerCat: a merged exoplanet catalog with Virtual Observatory connection

The heterogeneity of papers dealing with the discovery and characterization of exoplanets makes every attempt to maintain a uniform exoplanet catalog almost impossible. Four sources currently available online (NASA Exoplanet Archive, Exoplanet Orbit Database, Exoplanet Encyclopaedia, and Open Exoplanet Catalogue) are commonly used by the community, but they can hardly be compared, due to discrepancies in notations and selection criteria.
Exo-MerCat is a Python code that collects and selects the most precise measurement for all interesting planetary and orbital parameters contained in the four databases, accounting for the presence of multiple aliases for the same target. It can download information about the host star as well by the use of Virtual Observatory ConeSearch connections to the major archives such as SIMBAD and those available in VizieR. A Graphical User Interface is provided to filter data based on the user's constraints and generate automatic plots that are commonly used in the exoplanetary community.
With Exo-MerCat, we retrieved a unique catalog that merges information from the four main databases, standardizing the output and handling notation differences issues. Exo-MerCat can correct as many issues that prevent a direct correspondence between multiple items in the four databases as possible, with the available data. The catalog is available as a VO resource for everyone to use and it is periodically updated, according to the update rates of the source catalogs.

[ascl:1912.009] FORSTAND: Flexible ORbit Superposition Toolbox for ANalyzing Dynamical models

FORSTAND constructs dynamical models of galaxies using the Schwarzschild orbit-superposition method; the method is available as part of the AGAMA (ascl:1805.008) framework. The models created are constrained by line-of-sight kinematic observations and are applicable to galaxies of all morphological types, including disks and triaxial rotating bars.

[ascl:1912.010] AstroAccelerate: Accelerated software package for processing time-domain radio astronomy data

AstroAccelerate processes time-domain radio astronomy data. It offers a standalone code that can be used to process filterbank data and a library that performs GPU-accelerated single pulse processing (SPS), Fourier Domain Acceleration Searching (FDAS) and dedispersion in real-time on very large data-sets comparable to those that will be produced by next-generation radio telescopes such as the SKA. AstroAccelerate uses NVIDIAR GPUs, and is configurable, stable, and easily maintained.

[ascl:1912.011] QSOSIM: Simulated Quasar Spectrum Generator

QSOSIM realistically simulates high-resolution quasar spectra using a set of basic parameters (magnitude, redshift, and spectral index). The simulated spectra include physical effects seen in the real data: the power-law quasar continuum, the narrow and broad emission lines, absorption by neutral hydrogen (HI) in the Lyman alpha forest, and heavy element transitions along the line of sight. The code uses empirical HI column density, redshift, and b-parameter distributions to simulate absorption in the Lyman alpha forest. All absorbers with column densities larger than log [N(HI)/cm2]>17 have heavy element absorption, for which the column densities are calculated using the plasma simulation code CLOUDY (ascl:9910.001) and the radiative transfer code CUBA. The code also simulates the clustering of the intergalactic medium along the line of sight, the proximity effect of the quasar, and the effect of the cosmic ultraviolet background. Each simulated spectrum is saved in a single FITS file in as a noiseless R=100000 spectrum, as well as a spectrum convolved with Sloan Digital Sky Survey resolution (R=10000) and realistic noise.

[ascl:1912.012] GAME: GAlaxy Machine learning for Emission lines

GAME infers different ISM physical properties by analyzing the emission line intensities in a galaxy spectrum. The code is trained with a large library of synthetic spectra spanning many different ISM phases, including HII (ionized) regions, PDRs, and neutral regions. GAME is based on a Supervised Machine Learning algorithm called AdaBoost with Decision Trees as base learner. Given a set of input lines in a spectrum, the code performs a training on the library and then evaluates the line intensities to give a determination of the physical properties. The errors on the input emission line intensities and the uncertainties on the physical properties determinations are also taken into account. GAME infers gas density, column density, far-ultraviolet (FUV, 6–13.6 eV) flux, ionization parameter, metallicity, escape fraction, and visual extinction. A web interface for using the code is available.

[ascl:1912.013] GriSPy: Fixed-radius nearest neighbors grid search in Python

GriSPy (Grid Search in Python) uses a regular grid search algorithm for quick fixed-radius nearest-neighbor lookup. It indexes a set of k-dimensional points in a regular grid providing a fast approach for nearest neighbors queries. Optional periodic boundary conditions can be provided for each axis individually. GriSPy implements three types of queries: bubble, shell and the nth-nearest, and offers three different metrics of interest in astronomy: the Euclidean and two distance functions in spherical coordinates of varying precision, haversine and Vincenty. It also provides a custom distance function. GriSPy is particularly useful for large datasets where a brute-force search is not practical.

[ascl:1912.014] HARMPI: 3D massively parallel general relativictic MHD code

HARMPI is a parallel, 3D version of HARM (ascl:1209.005), which solves hyperbolic partial differential equations in conservative form using high-resolution shock-capturing techniques. The code is parallelized using MPI and is fully operational in 3D. HARMPI, like HARM, is capable of using non-uniform grids and solves the relativistic magnetohydrodynamic equations of motion on a stationary black hole spacetime in Kerr-Schild coordinates to evolve an accretion disk model.

[ascl:1912.015] ENTERPRISE: Enhanced Numerical Toolbox Enabling a Robust PulsaR Inference SuitE

ENTERPRISE (Enhanced Numerical Toolbox Enabling a Robust PulsaR Inference SuitE) is a pulsar-timing analysis code which performs noise analysis, gravitational-wave searches, and timing model analysis. It uses Tempo2 (ascl:1210.015) to find the maximum-likelihood fit for the timing parameters and the basis of the fit for the red noise parameters if they are significant.

[ascl:1912.016] GWpy: Python package for studying data from gravitational-wave detectors

The Python package GWpy analyzes and characterizes gravitational wave data. It provides a user-friendly, intuitive interface to the common time-domain and frequency-domain data produced by the LIGO and Virgo observatories and their analyses. The core Python infrastructure is influenced by, and extends the functionality of, the Astropy (ascl:1304.002) package, and its methodology has been derived from, and augmented by, the LIGO Algorithm Library Suite (LALSuite), a large collection of primarily C99 routines for analysis and manipulation of data from gravitational-wave detectors. These packages use the SWIG program to produce Python wrappings for all C modules, allowing the GWpy package to leverage both the completeness, and the speed, of these libraries.

[ascl:1912.017] PTMCMCSampler: Parallel tempering MCMC sampler package written in Python

PTMCMCSampler performs MCMC sampling using advanced techniques. The code implements a variety of proposal schemes, including adaptive Metropolis, differential evolution, and parallel tempering, which can be used together in the same run.

[ascl:1912.018] Tangos: Framework and web interface for database-driven analysis of numerical structure formation simulations

Tangos builds databases (along the lines of Eagle or MultiDark) for cosmological and zoom simulations. Its
modular system generates and queries databases. It is designed to store and manage results from a user's own analysis code, provides web and python interfaces, and allows users to construct science-focused queries, including across entire merger trees, without requiring knowledge of SQL. Tangos manages the process of populating the database with science data, including auto-parallelizing the analysis. It can be customized to work with multiple python modules such as pynbody (ascl:1305.002) or yt (ascl:1011.022) to process raw simulation data; it defaults to using SQLite, but allows use of other databases as the underlying store through the use of SQLAlchemy.

[ascl:1912.019] STACKER: Stack sources in interferometric data

STACKER stacks sources in interferometric data, i.e., averaging emission from different sources. The library allows stacking to be done directly on visibility data as well as in the image domain. The code is in format of a CASA (ascl:1107.013) task and implements uv- and image-stacking algorithms; it also provides several useful tasks for stacking related data processing. It allows introduction and stacking of random sources to estimate bias and noise, and also allows removal of a model of bright sources from the data.

[ascl:1912.020] MRExo: Non-parametric mass-radius relationship for exoplanets

MRExo performs non-parametric fitting and analysis of the mass-radius (M-R) relationship for exoplanets. Written in Python, it offers tools for fitting the M-R relationship to a given data set and also includes predicting (M->R, and R->M) and plotting functions.

[submitted] K2CE: Kepler-K2 Cadence Events

Since early 2018, the Kepler/K2 project has been performing a uniform global reprocessing of data from K2 Campaigns 0 through 14. Subsequent K2 campaigns (C15-C19) are being processed using the same processing pipeline. One of the major benefits of the reprocessing effort is that, for the first time, short-cadence (1-min) light curves are produced in addition to the standard long-cadence (30-min) light curves. Users have been cautioned that the Kepler pipeline detrending module (PDC), developed for use on original Kepler data, has not been tailored for use on short-cadence K2 observations. Systematics due to events on fast timescales, such as thruster firings, are sometimes poorly corrected for many short-cadence targets. A Python data visualization and manipulation tool, called Kepler-K2 Cadence Events, has been developed that identifies and removes cadences associated with problematic thruster events, thus producing better light curves. Kepler-K2 Cadence Events can be used to visualize and manipulate light curve files and target pixel files from the Kepler, K2, and TESS missions. This software is available at the following NASA GitHub repository https://github.com/nasa/K2CE .

[submitted] SDSS Dual Active Nuclei Galaxy Detection Pipeline

Dual Active Nuclei Galaxies (DAGNs) are rare occurrences in the sky. Until now, most AGNs have been described to be found serendipitously, or by manual observation. In recent years, there has been an increasing interest in such dual AGNs and their astrophysical properties. Their study is important to the understanding of galaxy formation, star formation and these objects are the precursors to Gravitational Wave Sources.

Hence, we have devised a pipeline, that along with systematic data collection, can detect such dual AGN candidates. A novel algorithm 'Graph-Boosted Gradient Ascent' has been devised to detect whether an R-band image of a galaxy is a potential candidate for a DAGN or not. The pipeline can be cloned to a user's machine, and by joining the AstrIRG_DAGN group on SciServer, astronomers can collectively contribute to the mining of DAGNs.

[submitted] JetSeT: numerical modeling and SED fitting tool for relativistic jets

JetSeT is an open source C/Python framework to reproduce radiative and accelerative processes acting in relativistic jets, allowing to fit the numerical models to observed data. The main features of this framework are:

- Handling observed data: re-binning, definition of data sets, bindings to astropy tables and quantities
definition of complex numerical radiative scenarios: Synchrotron Self-Compton (SSC), external Compton
(EC) and EC against the CMB

- Constraining of the model in the pre-fitting stage, based on accurate and already published
phenomenological trends. In particular, starting from phenomenological parameters, such as spectral indices,
peak fluxes and frequencies, and spectral curvatures, that the code evaluates automatically, the pre-fitting
algorithm is able to provide a good starting model, following the phenomenological trends that I have
implemented. fitting of multiwavelength SEDs using both frequentist approach (iminuit) and bayesian MCMC
sampling (emcee)

- Self-consistent temporal evolution of the plasma under the effect of radiative and accelerative processes,
both first order and second order (stochastic acceleration) processes.

[submitted] Time-domain astronomy sandbox

Time-domain astronomy sandbox consists in a series of classes to simulate and process time-domain astronomy data products in Python. The code was originally developed to model Fast Radio Burst (FRB) and Radio Frequency Interference (RFI), and evaluate different RFI mitigation methods and their effect on FRB search.

[submitted] amber_meta

amber_meta integrates a few routines to launch AMBER (the “Apertif Monitor for Bursts Encountered in Real-time”, https://github.com/AA-ALERT/AMBER) in a systematic manner. To avoid typing a string in the command line manually with all parameters required to launch AMBER, amber_meta generates the command from configuration files, and can directly launch AMBER instances.

[ascl:2001.001] Min-CaLM: Mineral compositional analysis on debris disk spectra

Min-CaLM performs automated mineral compositional analysis on debris disk spectra. The user inputs the debris disk spectrum, and using Min-CaLM's built-in mineralogical library, Min-CaLM calculates the relative mineral abundances within the disk. To do this calculation, Min-CaLM converts the debris disk spectrum and the mineralogical library spectra into a system of linear equations, which it then solves using non-negative least square minimization. This code comes with a GitHub tutorial on how to use the Min-CaLM package.

[ascl:2001.002] TRANSPHERE: 1-D spherical continuum radiative transfer

TRANSPHERE is a simple dust continuum radiative transfer code for spherically symmetric circumstellar envelopes. It handles absorption and re-emission and computes the dust temperature self-consistently; it does not, however, deal with scattering. TRANSPHERE uses a variable eddington factor method for the radiative transfer. The RADMD code (ascl:1108.016) is more versatile, but for a spherically symmetric problem for which scattering is of much concern, it may be easier to use a simple code such as TRANSPHERE.

Please note that this code has not been updated since 2006.

[ascl:2001.003] sf3dmodels: Star-forming regions 3D modelling package

sf3dmodels models star-forming regions; it brings together analytical models in order to compute their physical properties in a 3-dimensional grid. The package can couple different models in a single grid to recreate complex star forming systems such as those being revealed by current instruments. The output data can be read with LIME (ascl:1107.012) or RADMC-3D (ascl:1108.016) to carry out radiative transfer calculations of the modeled region.

[ascl:2001.004] FragMent: Fragmentation techniques for studying filaments

FragMent studies fragmentation in filaments by collating a number of different techniques, including nearest neighbour separations, minimum spanning tree, two-point correlation function, and Fourier power spectrum. It also performs model selection using a frequentist and Bayesian approach to find the best descriptor of a filament's fragmentation. While the code was designed to investigate filament fragmentation, the functions are general and may be used for any set of 2D points to study more general cases of fragmentation.

[ascl:2001.005] FAKEOBS: Model visibilities generator

The CASA (1107.013) task FAKEOBS generates model visibilities from already-existing measurement sets. This task can be used to substitute all the visibilities of the target with simulations computed from any model image. The measurement can either be with real or simulated data, the target can have been observed in mosaic mode, and there can be several sources (e.g., bandpass calibrator, flux/phase calibrator, and target).

[ascl:2001.006] Protostellar Evolution: Stellar evolution simulator

Protostellar Evolution simulates the evolution of stellar stellar radius and luminosity from the bound core stage through to the core hydrogen ignition as a zero-age main-sequence (ZAMS) star and beyond. Written in Fortran 90, the code is implemented as a module of the FLASH astrophysical fluid dynamics code (ascl:1010.082).

[ascl:2001.007] BTS: Behind The Spectrum

Behind The Spectrum (BTS) is a fully-automated multiple-component fitter for optically-thin spectra. Written as a python module, the routine uses the first, second and third derivatives to determine thenumber of components in the spectrum. A least-squared fitting routine then determines the best fit with that number of components, checking for over-fitting and over-lapping velocity centroids.

[ascl:2001.008] DebrisDiskFM: Debris Disk Forward Modeling

DebrisDiskFM provides forward modeling for circumstellar debris disks in scattered light using the MCFOST disk modeling software to generate disk model images using given input parameters and emcee (ascl:1303.002) to obtain the posterior distributions for these parameters.

[ascl:2001.009] ORCS: Analysis engine for SITELLE spectral cubes

ORCS (Outils de Réduction de Cubes Spectraux) is an analysis engine for SITELLE spectral cubes. The software extracts integrated spectra, fits the sinc emission lines, and recalibrates data in wavelength, astrometry and flux. ORCS offers a choice between a Bayesian or a classical fitting algorithm
, and also provides automatic source detection and radial velocity correction.

[ascl:2001.010] CosMOPED: Compressed Planck likelihood

CosMOPED (Cosmological MOPED) uses the MOPED (Multiple/Massively Optimised Parameter Estimation and Data compression) compression scheme to compress the Planck power spectrum. This convenient and lightweight compressed likelihood code is implemented in Python. To compute the likelihood for the LambdaCDM model using CosMOPED, one needs only six compression vectors, one for each parameter, and six numbers from compressing the Planck data using the six compression vectors. Using these, the likelihood of a theory power spectrum given the Planck data is the product of six one-dimensional Gaussians. Extended cosmological models require computing extra compression vectors.

[ascl:2001.011] ExoTETHyS: Exoplanetary transits and eclipsing binaries modeler

ExoTETHyS models exoplanetary transits, eclipsing binaries, and related phenomena. The package calculates stellar limb-darkening coefficients down to <10 parts per million (ppm) and generates an exact transit light-curve based on the entire stellar intensity profile rather than limb-darkening coefficients.

[ascl:2001.012] MCMCI: Markov Chain Monte Carlo + Isochrones method for characterizing exoplanetary systems

MCMCI (Markov chain Monte Carlo + isochrones) characterizes a whole exoplanetary system directly by modeling the star and its planets simultaneously. The code, written in Fortran, uses light curves and basic stellar parameters with a transit analysis algorithm that interacts with stellar evolutionary models, thus using both model-dependent and empirical age indicators to characterize the system.

[ascl:2001.013] RPPPS: Re-analyzing Pipeline for Parkes Pulsar Survey

RPPPS (Re-analysing Pipeline for Parkes Pulsar Survey) uses Linux shell scripts, C language, and python code and two parallel strategies to reorganize the PRESTO (ascl:1107.017) pulsar search pipeline to run multiple processes in parallel, thus accelerating the search for pulsars. Though originally designed for reprocessing PMPS data, the code has also been successfully used with FAST (Five-hundred-meter Aperture Spherical radio Telescope) drift scan data. The pipeline is only CPU-based and can be easily and quickly deployed in computing nodes for testing purposes or data processes.

[ascl:2001.014] Peasoup: C++/CUDA GPU pulsar searching library

The NVIDIA GPU-based pipeline code peasoup provides a one-step pulsar search, including searching for pulsars with up to moderate accelerations, with only one command. Its features include dedispersion, dereddening in the Fourier domain, resampling, peak detection, and optional time series folding. peasoup's output is the candidate list.

[ascl:2001.015] gnm: The MCMC Jagger

gnm is an implementation of the affine-invariant sampler for Markov chain Monte Carlo (MCMC) that uses the Gauss-Newton-Metropolis (GNM) Algorithm. The GNM algorithm is specialized in sampling highly non-linear posterior probability distribution functions of the form exp(-||f(x)||^2/2). The code includes dynamic hyper-parameter optimization to increase performance of the sampling; other features include the Jacobian tester and an error bars creator.

[submitted] StarburstPy: Python Wrapper for Starburst99

StarburstPy is a python wrapper for Starburst99 (ascl:1104.003). The code contains methods for setting all inputs, running Starburst99, and reading output data into python dictionaries.

[submitted] pycf3 - Cosmicflows-3 Distance-Velocity Calculator client for Python

The project is a simple Python client for Cosmicflows-3 Distance-Velocity Calculator at distances less than 400 Mpc (http://edd.ifa.hawaii.edu/CF3calculator/)

Compute expectation distances or velocities based on smoothed velocity field from the Wiener filter model of https://ui.adsabs.harvard.edu/abs/2019MNRAS.488.5438G/abstract.

[submitted] MERA: Analysis Tool for Astrophysical Simulation Data in the Julia Language

MERA works with large 3D AMR/uniform-grid and N-body particle data sets from astrophysical simulations such as those produced by the hydrodynamic code RAMSES (ascl:1011.007) and is written entirely in the Julia language. The package provides essential functions for efficient and memory lightweight data loading and analysis. The core of MERA is a database framework.

[submitted] Determination of Length of (Earth) Day [LOD] in the past geologic epochs

The protocol describes the algorithm of arriving at LOD in a given past geological Epoch. First the lunar orbital radius of the given geologic epoch has to be determined. For this the velocity of recession of Moon for the accelerated phase has to be determined. The spatial integral of the reciprocal of Velocity of recession gives the the transit time of Moon from desired orbit to the present orbit.Through several iterations the transit time is made to converge on the geologic epoch. Once we determine the desired orbital radius it has to be substituted in the LOD expression to determine the LOD in the given geologic epoch.

[ascl:2002.001] SDAR: Slow-Down Algorithmic Regularization code for solving few-body problems

SDAR (Slow-Down Algorithmic Regularization) simulates the long-term evolution of few-body systems such as binaries and triples. The algorithm used provides a few orders of magnitude faster performance than the classical N-body method. The secular evolution of hierarchical systems, e.g. Kozai-Lidov oscillation, can be well reproduced. The code is written in the C++ language and can be used either as a stand-alone tool or a library to be plugged in other N-body codes. The high precision of the floating point to 62 digits is also supported.

[ascl:2002.002] RASCAS: Resonant line transfer in AMR simulations

The massively parallel code RASCAS (RAdiative SCattering in Astrophysical Simulations) performs radiative transfer on an adaptive mesh with an octree structure using the Monte Carlo technique. The code features full MPI parallelization, domain decomposition, adaptive load-balancing, and a standard peeling algorithm to construct mock observations. The radiative transport of resonant line photons through different mixes of species (e.g. HI, SiII, MgII, FeII), including their interaction with dust, is implemented in a modular fashion to allow new transitions to be easily added to the code. RASCAS may also be used to propagate photons at any wavelength (e.g. stellar continuum or fluorescent lines), and has been designed to be easily customizable and to process simulations of arbitrarily large sizes on large supercomputers.

[ascl:2002.003] ORIGIN: detectiOn and extRactIon of Galaxy emIssion liNes

ORIGIN performs blind detection of faint emitters in MUSE datacubes. The algorithm is tuned to detect faint spatial-spectral emission signatures while allowing for a stable false detection rate over the data cube, and providing in the same time an automated and reliable estimation of the purity. ORIGIN implements a nuisance removal part based on a continuum subtraction combining a Discrete Cosine Transform and an iterative Principal Component Analysis and a detection part based on the local maxima of Generalized Likelihood Ratio test statistics obtained for a set of spatial-spectral profiles of emission line emitters. In addition, it performs a purity estimation in which the proportion of true emission lines is estimated from the data itself: the distribution of the local maxima in the noise only configuration is estimated from that of the local minima.

[ascl:2002.004] triceratops: Candidate exoplanet rating tool

triceratops (Tool for Rating Interesting Candidate Exoplanets and Reliability Analysis of Transits Originating from Proximate Stars) validates planet candidates from the Transiting Exoplanet Survey Satellite (TESS). The code calculates the probabilities of a wide range of transit-producing scenarios using the primary transit of the planet candidate and preexisting knowledge of its host and nearby stars. It then uses the known properties of these stars to calculate star-specific priors for each scenario with estimates of stellar multiplicity and planet occurrence rates.

[ascl:2002.005] ODUSSEAS: Observing Dwarfs Using Stellar Spectroscopic Energy-Absorption Shapes

ODUSSEAS (Observing Dwarfs Using Stellar Spectroscopic Energy-Absorption Shapes) uses machine learning to derive the Teff and [Fe/H] of M dwarf stars by using their optical spectra obtained by different spectrographs with different resolutions. The software uses the measurement of the pseudo equivalent widths for more than 4000 stellar absorption lines and the machine learning Python package scikit-learn (https://scikit-learn.org/stable/) to predict the stellar parameters.

[ascl:2002.006] ScamPy: Sub-halo Clustering and Abundance Matching Python interface

ScamPy "paints" an observed population of cosmological objects on top of the DM-halo/subhalo hierarchy obtained from DM-only simulations. The method combines the Halo Occupation Distribution (HOD) method with sub-halo abundance matching (SHAM); the two processes together are dubbed Sub-halo clustering and abundance matching (SCAM). The procedure requires applying the two methods in sequence; by applying the HOD scheme, the host sub-haloes are selected, and the SHAM algorithm associates an observable property of choice to each sub-halo. The provided python interface allows users to load and populate DM halos and sub-halos obtained by FoF and SUBFIND algorithms applied to DM snapshots at any redshift. The software is highly-optimized and flexible.

[ascl:2002.007] ProSpect: Spectral generation package

ProSpect generates good quality SEDs that can be used to estimate the broad band photometric properties of galaxies that have known star formation and gas metallicity histories. It allows for complex star formation and metallicity histories to be specified, and can be used in a generative or fitting (Bayesian) mode. ProSpect provides a high level interface to the BC03 (low and high resolution) and EMILES libraries, as well as the Dale 2014 dust emission templates. Its source code is available for download, and it is also available as an interactive web tool.

[ascl:2002.008] ExoSim: Simulator for predicting signal and noise in transit spectroscopy observations

ExoSim models host star and planet transit events, simulating the temporal change in stellar flux due to the light curve. It is wavelength-dependent, using an input planet spectrum to determine the light curve depth for any given wavelength and can capture temporal effects, such as correlated noise. ExoSim's star spot simulator produces simulated observations that include spot and facula contamination. The code is flexible and can be generically applied to different instruments that simulate specific time-dependent processes.

[ascl:2002.009] DASH: Deep Automated Supernova and Host classifier

DASH classifies the type, age, redshift and host for any supernova spectra based on the learned features, through use of a deep convolutional neural network to train a matching algorithm, of each supernova’s type and age. The Python library allows a user to classify spectra; the software is fast and can classify thousands of spectra in seconds. A graphical interface that enables a user to view and classify a spectrum is also available.

[ascl:2002.010] Apercal: Pipeline for the Westerbork Synthesis Radio Telescope Apertif upgrade

Apercal is a dedicated, automated data reduction and analysis pipeline written for the Apertif (APERture Tile In Focus) upgrade to the Westerbork Synthesis Radio Telescope. This upgrade dramatically increases the field of view and survey speed of the telescope and is being used for survey observations that can produce 5 terabytes of data for each observation. Apercal uses existing and new tools and parallelization to provide the performance needed for the large volume of data produced Apertif surveys. The software is written entirely in Python and uses third–party astronomical software, such as AOFlagger (ascl:1010.017), CASA (ascl:1107.013), and Miriad (ascl:1106.007), for certain tasks. Apercal is modular, making it possible to run specific modules manually instead of the full pipeline, and information can be exchanged between modules because status parameters are written and read from a python pickled dictionary file. The pipeline can also run fully automatically.

[ascl:2002.011] PyHammer: Python spectral typing suite

PyHammer performs rapid and automatic spectral classification of stars according to the Morgan-Keenan classification system; it is a Python revision of the IDL code The Hammer (ascl:1405.003) and offers additional capabilities. Working in the range of 3,650-10,200 Angstroms, the automatic spectral typing algorithm compares important spectral lines to template spectra and determines the best matching spectral type, ranging from O to L type stars. The code can also determine a star's metallicity ([Fe/H]) and radial velocity shifts. Once the automatic classification algorithm has run, PyHammer provides the user an interface for determining spectral types visually by comparing their spectra to provided templates.

[ascl:2002.012] DMRadon: Radon Transform calculation tools

DMRadon calculates the Radon Transform for use in the analysis of Directional Dark Matter Direct Detection. The code can calculate speed distributions, velocity distribution, velocity integral (eta) and Radon Transforms or a standard Maxwell-Boltzmann distribution. DMRadon also calculates the velocity distribution averaged over different angular bins.

[ascl:2002.013] GWecc: Calculator for pulsar timing array signals due to eccentric supermassive binaries

GWecc computes the pulsar timing array (PTA) signals induced by eccentric supermassive binaries. Written in C++, it computes the plus/cross polarizations as well as Earth and pulsar terms of the PTA signal given the binary parameters and the sky locations of the binary and the pulsar. A python wrapper is included through which GWecc can be used to simulate, search for and constrain gravitational wave-emitting eccentric supermassive binaries using packages such as ENTERPRISE (ascl:1912.015) and libstempo (ascl:2002.017).

[ascl:2002.014] HaloAnalysis: Read and analyze halo catalogs and merger trees

HaloAnalysis reads and analyzes halo/galaxy catalogs, generated from Rockstar (ascl:1210.008) or AHF (ascl:1102.009), and merger trees generated from Consistent Trees (ascl:1210.011). Written in Python, it offers the following functionalities: reads halo/galaxy/tree catalogs from multiple file formats; assigns baryonic particles and galaxy properties to dark-matter halos; combines and re-generates halo/galaxy/tree files in hdf5 format; analyzes properties of halos/galaxies; and selects halos to generate zoom-in initial conditions. The code includes a tutorial in the form of a Jupyter notebook.

[ascl:2002.015] GizmoAnalysis: Read and analyze Gizmo simulations

GizmoAnalysis reads and analyzes N-body simulations run with the Gizmo code (ascl:1410.003). Written in Python, it was developed primarily to analyze FIRE simulations, though it is usable with any Gizmo snapshot files. It offers the following functionality: reads snapshot files and converts particle data to physical units; provides a flexible dictionary class to store particle data and compute derived quantities on the fly; plots images and properties of particles; and generates region files for input to MUSIC (ascl:1311.011) to generate cosmological zoom-in initial conditions. GizmoAnalysis also computes rates of supernovae and stellar winds, including their nucleosynthetic yields, as used in FIRE simulations. The software package includes a tutorial in a Jupyter notebook.

[ascl:2002.016] Cobra: Bayesian pulsar searching

Cobra uses single pulse time series data to search for and time pulsars, performing a fully phase coherent timing analysis. The GPU-accelerated Bayesian analysis package, written in Python, incorporates models for both isolated and accelerated systems, as well as both Keplerian and relativistic binaries. Cobra builds a model pulse train that incorporates effects such as aliasing, scattering and binary motion and a simple Gaussian profile and compares this directly to the data; the software can thus combine data over multiple frequencies, epochs, or even across telescopes.

[ascl:2002.017] libstempo: Python wrapper for Tempo2

libstempo uses the Tempo2 library (ascl:1210.015) to load a pulsar's tim/par files, providing Python access to the TOAs, the residuals, the timing-model parameters, the fit procedure, and more.

[ascl:2002.018] Bayesfit: Command-line program for combining Tempo2 and MultiNest components

Bayesfit pulls together Tempo2 (ascl:1210.015) and MultiNest (ascl:1109.006) components to provide additional functionality such as the specification of priors; Nelder–Mead optimization of the maximum-posterior point; and the capability of computing the partially marginalized likelihood for a given subset of timing-model parameters. Bayesfit is a single python command-line application.

[ascl:2002.019] ExoRT: Two-stream radiative transfer code

ExoRT is a flexible, two-stream radiative transfer code that interfaces with CAM/CESM (http://www.cesm.ucar.edu/models/current.html) or 1D offline; it is also used with ExoCAM (ascl:2002.020). Quadrature is used for shortwave and hemispheric mean is used for longwave. The gas phase optical depths are calculate using a correlated K-distribution method, with overlapping bands treated using an amount weighted scheme. Cloud optics are treated using mie scattering for both liquid and ice clouds, and cloud overlap is treated using Monte Carlo Independent Column Approximation.

[ascl:2002.020] ExoCAM: Exoplanet Community Atmospheric Model

ExoCAM adapts the NCAR Community Earth System Model (CESM) for planetary and exoplanetary applications. The system files, source code, initial conditions files, and namelists provided do not run standalone. ExoCAM is a patch to be used with standard distributions of CESM version 1.2.1 (http://www.cesm.ucar.edu/models/current.html), and is also intended to be run with ExoRT (ascl:2002.019), a correlated-k radiative transfer package.

[ascl:2002.021] CR-SISTEM: Symplectic integrator for lunar core-mantle and orbital dynamics

CR-SISTEM models lunar orbital and rotational dynamics, taking into account the effects of a liquid core. Orbits of the Moon and Earth are fully integrated, and other planets (or additional point-mass satellites) may be included in the integration. Lunar and solar tides on Earth, eccentricity and obliquity tides on the Moon, and lunar core-mantle friction are included. The integrator is one file (crsistem5.for) written in FORTRAN 90, uses seven input files (settings.in, planets.in, moons.in, tidal.in, lunar.in, precess.in and core.in), and has at least eight output files (planet101.out, moon101.out, pole.out, spin_orb.out, spin_ecl.out, cspin_ecl.out, long.out and clong.out); additional moons and planets would add more output. The input files provided with the code set up a 1 Myr simulation of a slow-spinning Moon on an orbit of 40 Earth radii, which will then dynamically relax to the lowest-energy state (in this case it is a synchronous rotation with a core spinning separately from the mantle).

[ascl:2002.022] DISKMODs: Accretion Disk Radial Structure Models

DISKMODs provides radial structure models of accretion disk solutions. The following models are included: Novikov-Thorne thin disk model and Sadowski polytropic slim disk model. Each model implements a common interface that gives the radial dependence of selected geometrical, physical and thermodynamic quantities of the accretion flow. The model interpolates through a set of tabulated numerical solutions. These solutions are computed for a reference mass M=10 Msun. The model can rescale the disk structure to any mass, with masses in the range of 5-20 Msun giving reasonably good results.

[submitted] Pulseblast 2

PulseBlast 2 is a mass directory scanner for Unix, Darwin and DOS/NT kernels that allows the user to analyze PSRFITS data all the way from raw time series to fully calibrated and RFI excised TOAs. Many PSRFITS software packages in the past have dealt with data on the scale of a single file and many optimization methods for pulsar data currently used are used in PulseBlast 2 for individual file analysis. However, what sets PulseBlast 2 apart is its optimization across many directories spread out across the OS. The program is entirely written in Python 3, with a scope to change string operations into Perl scripts.

[ascl:2003.001] TESS-Point: High precision TESS pointing tool

TESS-Point converts astronomical target coordinates given in right ascension and declination to detector pixel coordinates for the MIT-led NASA Transiting Exoplanet Survey Satellite (TESS) spacecraft. The program can also provide detector pixel coordinates for a star by TESS input catalog identifier number and common astronomical name. Tess-point outputs the observing sector number, camera number, detector number, and pixel column and row.

[ascl:2003.002] MAGNETAR: Histogram of relative orientation calculator for MHD observations

MAGNETAR is a set of tools for the study of the magnetic field in simulations of MHD turbulence and polarization observations. It calculates the histogram of relative orientation between density structure in the magnetic field in data cubes from simulations of MHD turbulence and observations of polarization using the method of histogram of relative orientations (HRO).

[ascl:2003.003] acorns: Agglomerative Clustering for ORganising Nested Structures

acorns generates a hierarchical system of clusters within discrete data by using an n-dimensional unsupervised machine-learning algorithm that clusters spectroscopic position-position-velocity data. The algorithm is based on a technique known as hierarchical agglomerative clustering. Although acorns was designed with the analysis of discrete spectroscopic position-position-velocity (PPV) data in mind (rather than uniformly spaced data cubes), clustering can be performed in n-dimensions and the algorithm can be readily applied to other data sets in addition to PPV measurements.

[ascl:2003.004] scousepy: Semi-automated multi-COmponent Universal Spectral-line fitting Engine

scousepy is a Python implementation of spectral line-fitting IDL code SCOUSE (ascl:1601.003). It fits a large amount of complex astronomical spectral line data in a systematic way.

[ascl:2003.005] RHT: Rolling Hough Transform

The RHT (Rolling Hough Transform) measures linear intensity as a function of orientation in images. This machine vision algorithm works on any image-space (2D) data, and quantifies the presence of linear structure as a function of orientation. The RHT can be used to identify linear features in images, to quantify the orientation of structure in images, and to map image intensity from 2D x-y space to 3D x-y-orientation space. An option in the code allows the user to quantify intensity as a function of direction (modulo 2pi) rather than orientation (modulo pi). The RHT was first used to discover that filamentary structures in neutral hydrogen emission are aligned with the ambient magnetic field.

[ascl:2003.006] PORTAL: POlarized Radiative Transfer Adapted to Lines

PORTAL (POlarized Radiative Transfer Adapted to Lines), a 3D polarized radiative transfer code, simulates the emergence of polarization in the emission of atomic or molecular (sub-)millimeter lines. Written in Fortran90, PORTAL can be used in standalone mode or can process the output of other 3D radiative transfer codes

[ascl:2003.007] RAPID: Real-time Automated Photometric IDentification

RAPID (Real-time Automated Photometric IDentification) classifies multiband photometric light curves into several different transient classes. It uses a deep recurrent neural network to produce time-varying classifications, and because it does not rely on deriving computationally expensive features from the data, it is well suited for processing alerts that wide-field surveys such as the Zwicky Transient Facility (ZTF) and the Large Synoptic Survey Telescope (LSST) will produce.

[ascl:2003.008] CoastGuard: Automated timing data reduction pipeline

CoastGuard reduces Effelsberg data; it is written in python and based on PSRCHIVE (ascl:1105.014). Though primarily designed for Effelsberg PSRIX data, it contains components sufficiently general for use with psrchive-compatible data files from other observing systems. In particular, the radio frequency interference (RFI) removal algorithm has been applied to data from the Parkes Telescope and has also been adopted by the LOFAR pulsar timing data reduction pipeline.

[ascl:2003.009] TOASTER: Times-Of-Arrival Tracker

TOASTER is a pulse times-of-arrival (TOA) tracker. It stores reduced/folded observations, meta data, templates, parfiles, TOAs, and timefiles in an organized manner using an SQL database. TOASTER also provides a full-featured python toolkit for reliably interacting with the data and database, and provides scripts that, for example, list and summarize the TOAs in the data base, and generate TOA files in multiple formats. The framework can also be used to generate TOAs from observations using flexible and reproducible plugins referred to as "manipulators".

[ascl:2003.010] MARGE: Machine learning Algorithm for Radiative transfer of Generated Exoplanets

MARGE (Machine learning Algorithm for Radiative transfer of Generated Exoplanets) generates exoplanet spectra across a defined parameter space, processes the output, and trains, validates, and tests machine learning models as a fast approximation to radiative transfer. It uses BART (ascl:1608.004) for spectra generation and modifies BART’s Bayesian sampler (MC3, ascl:1610.013) with a random uniform sampler to propose models within a defined parameter space. More generally, MARGE provides a framework for training neural network models to approximate a forward, deterministic process.

[ascl:2003.011] HOMER: A Bayesian inverse modeling code

HOMER (Helper Of My Eternal Retrievals) is a machine-learning-accelerated Bayesian inverse modeling code. Given some data and uncertainties, the code determines the posterior distribution of a model. HOMER uses MC3 (ascl:1610.013) for its MCMC; its forward model is a neural network surrogate model trained by MARGE (ascl:2003.010). The code produces plots of the 1D marginalized posteriors, 2D pairwise posteriors, and parameter history traces, and can also overplot the 1D and 2D posteriors for comparison with another posterior. HOMER computes the Bhattacharyya coefficient to compare the similarity of two 1D marginalized posteriors.

[ascl:2003.012] PYSOLATOR: Remove orbital modulation from a binary pulsar and/or its companion

PYSOLATOR removes the orbital modulation from a binary pulsar and/or its companion. In essence, it subtracts the predicted Roemer delay for the given orbit and then resamples the time series so as to make the signal appear as if it were emitted from the barycenter of the binary system, making the search for pulses easier and faster.

[ascl:2003.013] AstroHOG: Analysis correlations using the Histograms of Oriented Gradients

AstroHOG compares extended spectral-line observations (PPV cubes); the histogram of oriented gradients (HOG) technique takes as input two PPV cubes and provides an estimate of their spatial correlation across velocity channels to study spatial correlation between different tracers of the ISM.

[ascl:2003.014] Torch: Coupled gas and N-body dynamics simulator

Torch simulates coupled gas and N-body dynamics in astrophysical systems such as newly forming star clusters. It combines the FLASH (ascl:1010.082) code for gas dynamics and the ph4 code for direct N-body evolution via the AMUSE framework.

[submitted] vlt-sphere: a python package to automatically reduce VLT/SPHERE data

The high-contrast imager SPHERE at the Very Large Telescope is a new generation instrument combining extreme adaptive optics and coronagraphy to directly image exoplanets in the near-infrared. This python package enables easy reduction of the data coming from IRDIS and IFS, the two near-infrared subsystems of SPHERE. The package relies on the official ESO pipeline, which must be installed separately.

[submitted] michi2 SED and SLED fitting tool

We provide a C++/Python-based chi-square fitting tool for fitting combinations of arbitrary number of libraries/components to a given observational data. Our applications include the fitting of a galaxy's spectral energy distribution (SED) with stellar, active galactic nuclear, dust and radio SED templates, and the fitting of a galaxy's spectral line energy distribution (SLED) with one or more gas components using radiative transfer LVG model grid libraries. Our tool first samples the high-dimensional parameter space (N1*N2*N3*..., where N is the number of independent templates in each library, and 1/2/3 is the ID of components) in an optimised way for a few thousand or tens of thousand times to compute the chi-square to the input observational data, then using Python scripts to analyse the chi-square distribution and derive the best-fit, median, lower and higher 1-sigma values for each parameter in each library/component. This tool is useful for fitting larger number of templates and arbitrary combinations of libraries/components, including some constraining of one library/component onto another. Technical details and applications are given in Liu et al. 2020 to be submitted.

[ascl:2004.001] Locus: Optimized differential photometry

Locus implements the Locus Algorithm, which maximizes the performance of differential photometry systems by optimizing the number and quality of reference stars in the Field of View with the target.

[ascl:2004.002] Tangra: Software for video photometry and astrometry

Tangra performs scientific grade data reduction of GPS time-tagged video observations, including reduction of stellar occultation light curves and astrometry of close flybys of Near Earth Objects. It offers Dark and Flat frame image correction, PSF and aperture photometry, multiple methods for deriving a background as well as extensibility via add-ins. Tangra is actively developed for Windows and the current version of the software supports UCAC2, UCAC3, UCAC4, NOMAD, PPMXL and Gaia DR2 star catalogues for astrometry. The software can perform motion-fitting for fast objects and derive a mini-normal astrometric positions. The supported video file formats are AVI, SER, ADV and AAV. Tangra can be also used with observations recorded as a sequence of FITS files. There are also versions for Linux and OS-X with more limited functionality.

[ascl:2004.003] IllinoisGRMHD: GRMHD code for dynamical spacetimes

IllinoisGRMHD is an open-source, highly-extensible rewrite of the original closed-source general relativistic (ideal) magnetohydrodynamics (GRMHD) code of the Illinois Numerical Relativity (ILNR) Group. Reducing the learning curve was the primary focus of this rewrite, with the goal of facilitating community involvement in the code's use and development, as well as reducing the human effort necessary to generate new science. IllinoisGRMHD also saves computer time, generating roundoff-precision identical output to the original code on adaptive-mesh grids while being nearly twice as fast at scales of hundreds to thousands of cores.

[ascl:2004.004] WD: Wilson-Devinney binary star modeling

Wilson-Devinney binary star modeling code (WD) is a complete package for modeling binary stars and their eclipes and consists of two main modules. The LC module generates light and radial velocity curves, spectral line profiles, images, conjunction times, and timing residuals; the DC module handles differential corrections, performing parameter adjustment of light curves, velocity curves, and eclipse timings by the Least Squares criterion. WD handles eccentric orbits and asynchronous rotation, and can compute velocity curves (with proximity and eclipse effects). It offers options for detailed reflection and nonlinear (logarithmic law) limb darkening, adjustment of spot parameters, an optional provision for spots to drift over the surface, and can follow light curve development over large numbers of orbits. Absolute flux solution allow Direct Distance Estimation (DDE) and there are improved solutions for ellipsoidal variables and for eclipsing binaries (EBs) with very shallow eclipses. Absolute flux solutions also can estimate temperatures of both EB components under suitable circumstances.

[ascl:2004.005] PyWD2015: Wilson-Devinney code GUI

PyWD2015 provides a modern graphical user interface (GUI) for the 2015 version of the Wilson-Devinney (WD) code (ascl:2004.004). The GUI is written in Python 2.7 and uses the Qt4 interface framework. At its core, PyWD2015 generates lcin and dcin files from user inputs and sends them to WD, then reads and visualizes the output in a user-friendly way. It also includes tools that make the technical aspects of the modeling process significantly easier.

[ascl:2004.006] ASTRAEUS: Semi-analytical semi-numerical galaxy evolution and reionization code

ASTRAEUS (semi-numerical rAdiative tranSfer coupling of galaxy formaTion and Reionization in n-body dArk mattEr simUlationS) self-consistently derives the evolution of galaxies and the reionization of the IGM based on the merger trees and density fields of a DM-only N-body simulation. It models gas accretion, star formation, SN feedback, the time and spatial evolution of the ionized regions, accounting for recombinations, HI fractions and photoionization rates within ionized regions, and radiative feedback. ASTRAEUS is for studying the galaxy-reionization interplay in the first billion years. The underlying code is written in C and is MPI-parallelized; its modular design allows new physical processes and galaxy properties to be added easily. ASTRAEUS can be run on a tree-branch-by-tree-branch basis (i.e., fully vertical) or on a redshift-by-redshift basis (i.e., fully horizontal) when evolving the galaxies by using local horizontal merger trees.

[ascl:2004.007] PyCosmo: Multi-purpose cosmology calculation tool

PyCosmo provides accurate predictions for cosmological observables including background quantities, power spectra and Limber and beyond-Limber angular power spectra. The software is designed to be interactive and user-friendly. It is available for download and is also offered on an interactive platform (PyCosmo Hub), which allows users to perform their own computations using Jupyter Notebooks without installing any software.

[ascl:2004.008] PPMAP: Column density mapping with extra dimensions

PPMAP provides column density mapping with extra dimensions (temperature and dust opacity index); it generate image cubes of differential column density as a function of (x,y) sky position and temperature for diffuse dusty structures. The code incorporates parallel processing using OpenMP for some of the more CPU-intensive steps. It is currently configured for the "Raven" cluster at Cardiff University and runs in a mode in which the computations are split between 16 separate nodes, each of which uses 16 cores with OpenMP.

[ascl:2004.009] stardate: Measure precise stellar ages

stardate measures precise stellar ages by combining isochrone fitting with gyrochronology (rotation-based ages) to increase the precision of stellar ages on the main sequence. The best possible ages provided by stardate will be for stars with rotation periods, though ages can also be predicted for stars without rotation periods. stardate is an extension to isochrones that incorporates gyrochronology and the code reverts back to isochrones when no rotation period is provided.

[ascl:2004.010] kombine: Kernel-density-based parallel ensemble sampler

kombine is an ensemble sampler built for efficiently exploring multimodal distributions. By using estimates of ensemble’s instantaneous distribution as a proposal, it achieves very fast burnin, followed by sampling with very short autocorrelation times.

[submitted] Carpyncho VVV Catalog Browser toolkit

Carpyncho, is a catalog browser that we hope will be reutilized to search for and characterize time variable data of the ~PiB size VVV/VVVx[1] survey. Is being developed for the detection and classification of periodic variables. For this purpose the stacked pawprint data from the VDFS CASU v >= 1.3 catalogs have been crossed matched with the VDFS CASU v1.3 tile catalogs into a Parquet files.
The Carpyncho infrastructure https://carpyncho.gihub.io is being developed entirely in Python on top of a Custom-Framework for data processing.
Also, a companion Python library is developed to access the same dataset as a Pandas DataFrame (https://pandas.pydata.org/).

[ascl:2004.011] FUNDPAR: Deriving FUNDamental PARameters from equivalent widths

FUNDPAR determines fundamental parameters of solar-type stars, by using as input the Equivalent Widths of Fe I,II lines. The code uses solar-scaled ATLAS9 model atmospheres with NEWODF opacities, together with the 2009 version of the MOOG (ascl:1202.009) program. Parameter files control different details, such as the mixing-length parameter, the overshooting, and the damping of the lines. FUNDPAR also derives the uncertainties of the parameters.

[ascl:2004.012] ArviZ: Exploratory analysis of Bayesian models

ArviZ provides backend-agnostic tools for diagnostics and visualizations of Bayesian inference by first converting inference data into xarray objects. It includes functions for posterior analysis, model checking, comparison and diagnostics. ArviZ’s functions work with NumPy arrays, dictionaries of arrays, xarray datasets, and have built-in support for PyMC3 (ascl:1610.016), PyStan, CmdStanPy, Pyro (ascl:1507.018), NumPyro, emcee (ascl:1303.002), and TensorFlow Probability objects. A Julia wrapper is also available.

[ascl:2004.013] Finesse: Frequency domain INterfErometer Simulation SoftwarE

Finesse is a numeric simulation for laser interferometers and models parametric instabilities, easily providing the required mechanical-to-optical transfer functions in imperfect and arbitrary interferometer configurations using Hermite-Gaussian beams. The code has been used to apply limits to the number and type of higher order modes used in simulation and investigate the potential use of higher order Laguerre-Gauss modes to reduce thermal noise in future gravitational wave detector designs. The PyKat wrapper (ascl:2004.014) helps automate complex Finesse tasks.

[ascl:2004.014] PyKat: Python interface and tools for Finesse

The Python wrapper PyKat extends the optical interferometer modeling software Finesse (ascl:2004.013). It provides an efficient GUI for conducting complex numerical simulations and manipulating and viewing simulation setups, and enables the use of Python's extensive scientific software ecosystem.

[ascl:2004.015] IRDAP: SPHERE-IRDIS polarimetric data reduction pipeline

RDAP (IRDIS Data reduction for Accurate Polarimetry) accurately reduces SPHERE-IRDIS polarimetric data. It is a highly-automated end-to-end pipeline; its core feature is model-based correction of the instrumental polarization effects. IRDAP handles data taken both in field- and pupil-tracking mode and using the broadband filters Y, J, H and Ks. Data taken with the narrowband filters can be reduced as well, although with a somewhat worse accuracy. For pupil-tracking observations IRDAP can additionally apply angular differential imaging.

[ascl:2004.016] PRECISION: Astronomical infrared observations data reduction

PRECISION reduces astronomical IR imaging data. Written with SPHERE data in mind, it provides a fast and easy reduction of bright sources suitable for science. While it may not extract the absolute maximum amount of science, the objective is to provide a means to get science-ready data with minimal computing time or human interaction.

[submitted] RM-Tools

RM-Tools is a package for analysis of radio polarization data, specifically the use of Faraday rotation measure synthesis and Stokes QU model fitting. It contains routines for both single-pixel 1D polarized spectra as well as 3D polarization cubes. RM-Tools is intended to serve as a toolkit for studies of polarized radio sources and measurements of their Faraday rotation. RM-Tools is the core package for the pipelines used for the POlarized Sky Survey of the Universe's Magnetism (POSSUM) and the polarization component of the Very Large Array Sky Survey (VLASS). The package is maintained by the Canadian Initiative for Radio Astronomy Data Analysis (CIRADA; cirada.org).

[submitted] sslf: A simple spectral-line finder

sslf is designed to be a dead-simple, effective and useful spectral line finder for 1D data. It utilises the continuous wavelet transform from scipy, which is a productive way to find even weak spectral lines.

[submitted] RID: Relativistic Image Doubling in water Cherenkov detectors

Monte Carlo simulations were performed in C++ and Fortran to find the relative fraction of charged, relativistic particles entering a HAWC-like Water Cherenkov Detector that can cause a Relativistic Image Doubling (RID) effect visible to at least one of the internal detectors. Also presented is the code in python for the horizontal incidence of the muon inside the tank.

Would you like to view a random code?