ASCL.net

Astrophysics Source Code Library

Making codes discoverable since 1999

Browsing Codes

Order
Title Date
 
Mode
Abstract Compact
Per Page
50100250All
[ascl:2012.017] SLIT: Sparse Lens Inversion Technique

SLIT (Sparse Lens Inversion Technique) provides a method for inversion of lensed images in the frame of strong gravitational lensing. The code requires the input image along with lens mass profile and a PSF. The user then has to chose a maximum number of iterations after which the algorithm will stop if not converged and a image size ratio to the input image to set the resolution of the reconstructed source. Results are displayed in pyplot windows.

[ascl:2012.016] Pomegranate: Probabilistic model builder

Pomegranate builds probabilistic models in Python that is implemented in Cython for speed. The code merges the easy-to-use API of scikit-learn with the modularity of probabilistic modeling, including general mixture and hidden Markov models and Bayesian networks, to allow users to specify complicated models without the need to be concerned about implementation details. The models are built from the ground up and natively support features such as multi-threaded parallelism and out-of-core processing.

[ascl:2012.015] seaborn: Statistical data visualization

Seaborn provides a high-level interface for drawing attractive statistical graphics. Written in Python, it builds on matplotlib and integrates closely with pandas data structures. Its plotting functions operate on dataframes and arrays containing whole datasets and internally perform the necessary semantic mapping and statistical aggregation to produce informative plots. Its dataset-oriented, declarative API allows the user to focus on what the different elements of the plots mean, rather than on the details of how to draw them.

[ascl:2012.014] dolphin: Automated pipeline for lens modeling

Dolphin uniformly models large lens samples. It is a wrapper for Lenstronomy (ascl:1804.012), and features semi-automated modeling of a large sample of quasar and galaxy-galaxy lenses. Dolphin, written in Python, provides easy portability between local and MPI environments.

[ascl:2012.013] sedop: Optimize discrete versions of common SEDs

sedop is a Monte-Carlo minimization code designed to optimally construct spectral energy distributions (SEDs) for sources of ultraviolet and X-ray radiation employed in numerical simulations of reionization and radiative feedback.

[ascl:2012.012] TRAN_K2: Planetary transit search

TRAN_K2 searches for periodic transits in the photometric time series of the Kepler K2 mission. The search is made by considering stellar variability and instrumental systematics. TRAN_K2 is written in Fortran 77 and has a single input parameter file that can be edited by the user depending on the type of run and parameter ranges to be used.

[ascl:2012.011] Skye: Excess clustering of transit times detection

Skye detects a statistically significant excess clustering of transit times, indicating that there are likely systematics at specific times that cause many false positive detections, for the Kepler DR25 planet candidate catalog. The technique could be used for any survey looking to statistically cull false alarms.

[ascl:2012.010] MADLens: Differentiable lensing simulator

MADLens produces non-Gaussian cosmic shear maps at arbitrary source redshifts. A MADLens simulation with only 256^3 particles produces convergence maps whose power agree with theoretical lensing power spectra up to scales of L=10000. The code is based on a highly parallelizable particle-mesh algorithm and employs a sub-evolution scheme in the lensing projection and a machine-learning inspired sharpening step to achieve these high accuracies.

[ascl:2012.009] HydroCode1D: 1D finite volume code

HydroCode1D is a 1D finite volume code that can run any problem with 1D or 2D/3D spherical symmetry including external gravity or self-gravity. The program provides, depending on the configuration, output files that contain the midpoint position, density, velocity and pressure for each cell in the grid (in SI units). The program will by default use all available threads (as given by the environment variable OMP_NUM_THREADS). This can be overwritten by giving the desired number of threads as a command line argument to the program.

[ascl:2012.008] LIFELINE: LIne proFiles in massivE coLliding wInd biNariEs

LIFELINE (LIne proFiles in massivE coLliding wInd biNariEs) simulates the X-ray lines profile in colliding wind binaries. The code is self-consistent and computes the distribution of the wind velocity, the characterization of the wind shock region, and the line profile. In addition to perform the overall computation, LIFELINE can use a pre-computed velocity distribution to compute the shock characteristics and the line profile, or use pre-computed shock characteristics and velocity distributions to compute only the line profile.

[ascl:2012.007] EOS: Equation of State for planetary impacts

EOS is an analytical equation of state which models high pressure theory and fits well to the experimental data of ∊-Fe, SiO2, Mg2SiO4, and the Earth. The cold part of the EOS is modeled after the Varpoly EOS. The thermal part is based on a new formalism of the Gruneisen parameter, which improves behavior from earlier models and bridges the gap between elasticity and thermoelasticity. The EOS includes an expanded state model, which allows for the accurate modeling of material vapor curves.

[ascl:2012.006] Robovetter: Automatic vetting of Threshold Crossing Events (TCEs)

The DR25 Kepler Robovetter is a robotic decision-making code that dispositions each Threshold Crossing Event (TCE) from the final processing (DR 25) of the Kepler data into Planet Candidates (PCs) and False Positives (FPs). The Robovetter provides four major flags to designate each FP TCE as Not Transit-Like (NTL), a Stellar Eclipse (SS), a Centroid Offset (CO), and/or an Ephemeris Match (EM). It produces a score ranging from 0.0 to 1.0 that indicates the Robovetter's disposition confidence, where 1.0 indicates strong confidence in PC, and 0.0 indicates strong confidence in FP. Finally, the Robovetter provides comments in a text string that indicate the specific tests each FP TCE fails and provides supplemental information on all TCEs as necessary.

[ascl:2012.005] MLC_ELGs: Machine Learning Classifiers for intermediate redshift Emission Line Galaxies

MLC_EPGs classifies intermediate redshift (z = 0.3–0.8) emission line galaxies as star-forming galaxies, composite galaxies, active galactic nuclei (AGN), or low-ionization nuclear emission regions (LINERs). It uses four supervised machine learning classification algorithms: k-nearest neighbors (KNN), support vector classifier (SVC), random forest (RF), and a multi-layer perceptron (MLP) neural network. For input features, it uses properties that can be measured from optical galaxy spectra out to z < 0.8—[O III]/Hβ, [O II]/Hβ, [O III] line width, and stellar velocity dispersion—and four colors (u−g, g−r, r−i, and i−z) corrected to z = 0.1.

[ascl:2012.004] BinaryStarSolver: Orbital elements of binary stars solver

Given a series of radial velocities as a function of time for a star in a binary system, BinaryStarSolver solves for various orbital parameters. Namely, it solves for eccentricity (e), argument of periastron (ω), velocity amplitude (K), long term average radial velocity (γ), and orbital period (P). If the orbital parameters of a primary star are already known, it can also find the orbital parameters of a companion star, with only a few radial velocity data points.

[ascl:2012.003] Sengi: Interactive viewer for spectral outputs from stellar population synthesis models

Sengi enables online viewing of the spectral outputs of stellar population synthesis (SPS) codes. Typical SPS codes require significant disk space or computing resources to produce spectra for simple stellar populations with arbitrary parameters, making it difficult to present their results in an interactive, web-friendly format. Sengi uses Non-negative Matrix Factorisation (NMF) and bilinear interpolation to estimate output spectra for arbitrary values of stellar age and metallicity; this reduces the disk requirements and computational expense, allowing Sengi to serve the results in a client-based Javascript application.

[ascl:2012.002] NSCG: NOIRLab Source Catalog Generator

The NOIRLab Source Catalog Generator generates the NOIRLab Source Catalog (NSC), a catalog of all publicly available imagining data in the NOIRLab Astro Data Archive. The second data release (DR2) of the archive contains over 3.9 billion unique objects, 68 billion individual source measurements, covers 35,000 square degrees of the sky, has depths of 23rd magnitude in most broadband filters with 1-2% photometric precision, and astrometric accuracy of 7 mas. NSCG is written in Python and IDL. Three main steps generate the NSC: (1) Source Extractor (ascl:1010.064) is used to detect and measure sources in individual images; (2) astrometrics are calibrated with Gaia DR2 and photometric calibration using large public photometric catalogs such as Pan-STARRS1 and ATLAS-Refcat2; and, (3) measurements are clustered into unique objects, averaging photometric and morphological properties, and calculating proper motions and photometric variability indices.

[ascl:2012.001] getsf: Multi-scale, multi-wavelength sources and filaments extraction

getsf extracts sources and filaments in astronomical images by separating their structural components, and is designed to handle multi-wavelength sets of images and very complex filamentary backgrounds. The method spatially decomposes the original images and separates the structural components of sources and filaments from each other and from their backgrounds, flattening their resulting images. It spatially decomposes the flattened components, combines them over wavelengths, and detects the positions of sources and skeletons of filaments. Finally, getsf measures the detected sources and filaments and creates the output catalogs and images. This universal and fully automated method has a single user-definable free parameter, which reduces to a minimum dependence of its results on the human factor.

[submitted] ExoPix: Exoplanet Imaging with JWST

ExoPix is a collection of tutorials aimed at illustrating the imaging of exoplanets with the James Webb Space Telescope (JWST). ExoPix tutorials are meant to demonstrate the application of the PSF-subtraction algorithm pyKLIP (ascl:1506.001) to simulated JWST NIRCAM data. We provide simple walkthroughs of pyKLIP’s ability to reveal exoplanets, compute contrast curves, and measure exoplanet astrometry and photometry in imaged extrasolar systems.

[ascl:2011.030] DDCalc: Dark matter direct detection phenomenology package

DDCalc performs various dark matter direct detection calculations, including signal rate predictions, constraints on light DM, and likelihoods for several experiments. It offers eighteen non-relativistic effective operators to describe velocity and momentum transfer, and elastic scattering of DM particles off nucleons, and has an extended detector interface.

[ascl:2011.029] DarkBit: Dark matter constraints calculator

DarkBit computes dark matter constraints on extensions to the Standard Model of particle physics. Written in the GAMBIT (ascl:1708.030) framework, it seamlessly integrates with other tools in the statistical fitting framework; it is also available as a standalone tool. It offers a signal yield calculator for gamma-ray observations, provides likelihoods for arbitrary combinations of spin-independent and spin-dependent scattering processes, and provides a general solution for studying complex particle physics models that predict dark matter annihilation to a multitude of final states.

[ascl:2011.028] CWITools: Tools for Cosmic Web Imager data

CWITools analyzes integral field spectroscopy data from the Palomar and Keck Cosmic Web Imagers, and can be adapted for any three-dimensional integral field spectroscopy data. The package is modular, allowing users to construct data analysis pipelines to suit their own scientific needs, and includes tools for reducing data cubes, extracting a target signal, making emission maps, spectra, and other products. It also fits emission line and radial profiles and obtains final scalar quantities such as size and luminosity, among other tasks. It also contains helper functions that can, for example, obtain the wavelength axis from a 3D header, and create an auto-populated list of nebular emission lines or sky lines.

[ascl:2011.027] kiauhoku: Stellar model grid interpolation

Kiauhoku interacts with, manipulates, and interpolates between stellar evolutionary tracks in a model grid. It was built for interacting with YREC models, but other stellar evolution model grids, including MIST, Dartmouth, and GARSTEC, are also available.

[ascl:2011.026] DeepShadows: Finding low-surface-brightness galaxies in survey images

DeepShadows uses a convolutional neural networks (CNNs) to separate low-surface-brightness galaxies (LSBGs) from artifacts (such as Galactic cirrus and star-forming regions) in survey images. The model is trained and tested on labeled LSBGs and artifacts from the Dark Energy Survey and demonstrates that CNNs offer a promising path in the quest to study the low-surface-brightness universe.

[ascl:2011.025] PNICER: Extinction estimator

PNICER estimates extinction for individual sources and creates extinction maps using unsupervised machine learning algorithms. Extinction towards single sources is determined by fitting Gaussian Mixture Models along the extinction vector to (extinction-free) control field observations. PNICER also offers access to the well-established NICER technique in a simple unified interface and is capable of building extinction maps including the NICEST correction for cloud substructure.

[ascl:2011.024] ACStools: Python tools for Hubble Space Telescope Advanced Camera for Surveys data

The ACStools package contains Python tools to work with data from the Hubble Space Telescope (HST) Advanced Camera for Surveys (ACS). The package has several calibration utilities and a zeropoints calculator, can detect satellite trails, and offers destriping, polarization, and photometric tools.

[ascl:2011.023] reproject: Python-based astronomical image reprojection

reproject implements image reprojection (resampling) methods for astronomical images using various techniques via a uniform interface. Reprojection re-grids images from one world coordinate system to another (for example changing the pixel resolution, orientation, coordinate system). reproject works on celestial images by interpolation, as well as by finding the exact overlap between pixels on the celestial sphere. It can also reproject to/from HEALPIX projections by relying on the astropy-healpix package.

[ascl:2011.022] GPCAL: Instrumental polarization calibration in VLBI data

GPCAL performs instrumental polarization calibration in very long baseline interferometry (VLBI) data. It enhances the calibration accuracy by enabling users to fit the model to multiple calibrators data simultaneously and to take into account the calibrators linear polarization structures instead of using the conventional similarity assumption. GPCAL is based on AIPS (ascl:9911.003) and uses ParselTongue (ascl:1208.020) to run AIPS tasks.

[ascl:2011.021] HSTCosmicrays: Analyzing cosmic rays in HST calibration data

HSTCosmicrays finds and characterizes cosmic rays found in dark frames (exposures taken with the shutter closed) taken with instruments on the Hubble Space Telescope (HST). Dark exposures are obtained routinely by all the Hubble Space Telescope instruments for calibration. The main processing pipeline runs locally or in the cloud on AWS utilizing the HST Public Dataset.

[ascl:2011.020] REBOUNDx: Adding effects in REBOUND N-body integrations

REBOUNDx incorporates additional physics into REBOUND (ascl:1110.016) simulations. Users can add effects from a list of pre-implemented astrophysical forces or contribute new ones. The main code is written in C, and a Python wrapper is provided for interfacing with other libraries. The REBOUNDx source code is machine independent and a binary format to save REBOUNDx configurations interfaces with the SimulationArchive class in REBOUND, making it possible to share and reproduce results bit by bit.

[ascl:2011.019] Scintools: Pulsar scintillation data tools

SCINTOOLS (SCINtillation TOOLS) simulates and analyzes pulsar scintillation data. The code can be used for processing observed dynamic spectra, computing secondary spectra and ACFs, measuring scintillation arcs, simulating dynamic spectra, and modeling pulsar transverse velocities through scintillation arcs or diffractive timescales.

[ascl:2011.018] Clustering: Code for clustering single pulse events

Clustering is a modified version of the single-pulse sifting algorithm RRATrap (ascl:2011.017) combined with DBSCAN codes to cluster single pulse events.

[ascl:2011.017] RRATtrap: Rotating Radio Transient identifier

RRATtrap is a single-pulse sifting algorithm to identify Rotating Radio Transients (RRATs) and transients using output from the PRESTO (ascl:1107.017) routine single_pulse_search.py. It can be integrated into pulsar survey data analysis pipelines and, in addition to finding RRATs, it can also identify Fast Radio Bursts.

[ascl:2011.016] GoFish: Molecular line detections in protoplanetary disks

GoFish exploits the known rotation of a protoplanetary disk to shift all emission to a common line center in order to stack them, increasing the signal-to-noise of the spectrum, detecting weaker lines, or super-sampling the spectrum to better resolve the line profile.

[ascl:2011.015] EvapMass: Minimum mass of planets predictor

EvapMass predicts the minimum masses of planets in multi-planet systems using the photoevaporation-driven evolution model. The planetary system requires both a planet above and below the radius gap to be useful for this test. EvapMass includes an example Jupyter notebook for the Kepler-36 system. EvalMass can be used to identify TESS systems that warrant radial-velocity follow-up to further test the photoevaporation model.

[ascl:2011.014] SEDkit: Spectral energy distribution construction and analysis tools

SEDkit constructs and analyzes simple spectral energy distributions (SED). This collection of pure Python modules creates individual SEDs or SED catalogs from spectra and/or photometry and calculates fundamental parameters (fbol, Mbol, Lbol, Teff, mass, log(g)).

[ascl:2011.013] TLC: Tidally Locked Coordinates

Tidally Locked Coordinates converts global climate model (GCM) output from standard/Earth-like coordinates into a tidally locked coordinate system. The transformations in Tidally Locked Coordinates are useful for plotting and analyzing GCM simulations of slowly rotating tidally locked planets such as Earth-like planets inside the habitable zone of small stars. They can be used to leverage the fact that a slowly rotating planet's climate will start to look approximately symmetric about the axis of insolation.

[ascl:2011.012] wobble: Time-series spectra analyzer

wobble analyzes time-series spectra. It was designed with stabilized extreme precision radial velocity (EPRV) spectrographs in mind, but is highly flexible and extensible to a variety of applications. It takes a data-driven approach to deriving radial velocities and requires no a priori knowledge of the stellar spectrum or telluric features.

[ascl:2011.011] frbcat: Fast Radio Burst CATalog querying package

frbcat queries and downloads Fast Radio Burst (FRB) data from the FRBCAT Catalogue web page, the CHIME-REPEATERS web page and the Transient Name Server (TNS). It is written in Python and can be installed using pip.

[ascl:2011.010] ARES: Accelerated Reionization Era Simulations

The Accelerated Reionization Era Simulations (ARES) code rapidly generates models for the global 21-cm signal. It can also be used as a 1-D radiative transfer code, stand-alone non-equilibrium chemistry solver, or global radiation background calculator.

[ascl:2011.009] HaloGen: Modular halo model code

HaloGen computes all auto and cross spectra and halo model trispectrum in simple configurations. This modular halo model code computes 3d power spectra, and the corresponding projected 2d power spectra in the Limber and flat sky approximations. The observables include matter density, galaxy lensing, CMB lensing, thermal Sunyaev-Zel'dovich, cosmic infrared background, tracers with any dn/dz, b(z) and HOD.

[ascl:2011.008] GOTHIC: Double nuclei galaxy detector

GOTHIC (Graph-bOosTed iterated HIll Climbing) detects whether a given image of a galaxy has characteristic features of a double nuclei galaxy (DNG). Galaxy interactions and mergers play a crucial role in the hierarchical growth of structure in the universe; galaxy mergers can lead to the formation of elliptical galaxies and larger disk galaxies, as well as drive galaxy evolution through star formation and nuclear activity. During mergers, the nuclei of the individual galaxies come closer and finally form a double nuclei galaxy. Although mergers are common, the detection of double-nuclei galaxies (DNGs) is rare and fairly serendipitous. Their properties can help us understand the formation of supermassive black hole (SMBH) binaries, dual active galactic nuclei (DAGN) and the associated feedback effects. GOTHIC provides an automatic and systematic way to survey data for the discovery of double nuclei galaxies.

[ascl:2011.007] DYNAMITE: DYnamics, Age and Metallicity Indicators Tracing Evolution

DYNAMITE (DYnamics, Age and Metallicity Indicators Tracing Evolution) is a triaxial dynamical modeling code for stellar systems and is based on existing codes for Schwarzschild modeling in triaxial systems. DYNAMITE provides an easy-to-use object oriented Python wrapper that extends the scope of pre-existing triaxial Schwarzschild codes with a number of new features, including discrete kinematics, more flexible descriptions of line-of-sight velocity distributions, and modeling of stellar population information. It also offers more efficient steps through parameter space, and can use GPU acceleration.

[ascl:2011.006] tlpipe: Data processing pipeline for the Tianlai experiment

tlpipe processes the drift scan survey data from the Tianlai experiment; the Tainlai project is a 21cm intensity mapping experiment aimed at detecting dark energy by measuring the baryon acoustic oscillation (BAO) features in the large scale structure power spectrum. tlpipe performs offline data processing tasks such as radio frequency interference (RFI) flagging, array calibration, binning, and map-making, in addition to other tasks. It includes utility functions needed for the data analysis, such as data selection, transformation, visualization and others. tlpipe implements a number of new algorithms are implemented, including the eigenvector decomposition method for array calibration and the Tikhnov regularization for m-mode analysis.

[ascl:2011.005] DarkCapPy: Dark Matter Capture and Annihilation

DarkCapPy calculates rates associated with dark matter capture in the Earth, annihilation into light mediators, and observable decay of the light mediators near the surface of the Earth. This Python/Jupyter package can calculate the Sommerfeld enhancement at the center of the Earth and the timescale for capture-annihilation equilibrium, and can be modified for other compact astronomical objects and mediator spins.

[ascl:2011.004] MCMCDiagnostics: Markov Chain Monte Carlo convergence diagnostics

MCMCDiagnostics contains two diagnostics, written in Julia, for Markov Chain Monte Carlo. The first is potential_scale_reduction(chains...), which estimates the potential scale reduction factor, also known as Rhat, for multiple scalar chains
. The second, effective_sample_size(chain), calculates the effective sample size for scalar chains. These diagnostics are intended as building blocks for use by other libraries.

[ascl:2011.003] Kalkayotl: Inferring distances to stellar clusters from Gaia parallaxes

Kalkayotl obtains samples of the joint posterior distribution of cluster parameters and distances to the cluster stars from Gaia parallaxes using Bayesian inference. The code is designed to deal with the parallax spatial correlations of Gaia data, and can accommodate different values of parallax zero point and spatial correlation functions.

[ascl:2011.002] CAPTURE: Interferometric pipeline for image creation from GMRT data

CAPTURE (CAsa Pipeline-cum-Toolkit for Upgraded Giant Metrewave Radio Telescope data REduction) produces continuum images from radio interferometric data. Written in Python, it uses CASA (ascl:1107.013) tasks to analyze data obtained by the GMRT. It can produce self-calibrated images in a fully automatic mode or can run in steps to allow the data to be inspected throughout processing.

[ascl:2011.001] AdaMet: Adaptive Metropolis for Bayesian analysis

AdaMet (Adaptive Metropolis) performs efficient Bayesian analysis. The user-friendly Python package is an implementation of the Adaptive Metropolis algorithm. In many real-world applications, it is more efficient and robust than emcee (ascl:1303.002), which warm-up phase scales linearly with the number of walkers. For this reason, and because of its didactic value, the AdaMet code is provided as an alternative.

[ascl:2010.015] relxill: Reflection models of black hole accretion disks

relxill self-consistently connects an angle-dependent reflection model constructed with XILLVER (http://www.srl.caltech.edu/personnel/javier/xillver/index.html) with the relativistic blurring code RELLINE (ascl:1505.021). It calculates the proper emission angle of the radiation at each point on the accretion disk and then takes the corresponding reflection spectrum into account.

[ascl:2010.014] Pix2Prof: Deep learning for textraction of useful sequential information from galaxy imagery

Pix2Prof produces a surface brightness profile from an unprocessed galaxy image from the SDSS in either the g, r, or i bands. It is fast, and given suitable training data, Pix2Prof can be retrained to produce any galaxy profile from any galaxy image.

Would you like to view a random code?