ASCL.net

Astrophysics Source Code Library

Making codes discoverable since 1999

Browsing Codes

Results 3501-3750 of 3644 (3551 ASCL, 93 submitted)

Order
Title Date
 
Mode
Abstract Compact
Per Page
50100250All
[ascl:1802.006] VISIBLE: VISIbility Based Line Extraction

VISIBLE applies approximated matched filters to interferometric data, allowing line detection directly in visibility space. The filter can be created from a FITS image or RADMC3D output image, and the weak line data can be a CASA MS or uvfits file. The filter response spectrum can be output either to a .npy file or returned back to the user for scripting.

[ascl:1408.010] VisiOmatic: Celestial image viewer

VisiOmatic is a web client for IIPImage (ascl:1408.009) and is used to visualize and navigate through large science images from remote locations. It requires STIFF (ascl:1110.006), is based on the Leaflet Javascript library, and works on both touch-based and mouse-based devices.

[ascl:1103.007] VisIt: Interactive Parallel Visualization and Graphical Analysis Tool

VisIt is a free interactive parallel visualization and graphical analysis tool for viewing scientific data on Unix and PC platforms. Users can quickly generate visualizations from their data, animate them through time, manipulate them, and save the resulting images for presentations. VisIt contains a rich set of visualization features so that you can view your data in a variety of ways. It can be used to visualize scalar and vector fields defined on two- and three-dimensional (2D and 3D) structured and unstructured meshes. VisIt was designed to handle very large data set sizes in the terascale range and yet can also handle small data sets in the kilobyte range. See the table below for more details about the tool’s features.

VisIt was developed by the Department of Energy (DOE) Advanced Simulation and Computing Initiative (ASCI) to visualize and analyze the results of terascale simulations. It was developed as a framework for adding custom capabilities and rapidly deploying new visualization technologies. Although the primary driving force behind the development of VisIt was for visualizing terascale data, it is also well suited for visualizing data from typical simulations on desktop systems.

[ascl:1011.020] VisIVO: Integrated Tools and Services for Large-Scale Astrophysical Visualization

VisIVO is an integrated suite of tools and services specifically designed for the Virtual Observatory. This suite constitutes a software framework for effective visual discovery in currently available (and next-generation) very large-scale astrophysical datasets. VisIVO consists of VisiVO Desktop - a stand alone application for interactive visualization on standard PCs, VisIVO Server - a grid-enabled platform for high performance visualization and VisIVO Web - a custom designed web portal supporting services based on the VisIVO Server functionality. The main characteristic of VisIVO is support for high-performance, multidimensional visualization of very large-scale astrophysical datasets. Users can obtain meaningful visualizations rapidly while preserving full and intuitive control of the relevant visualization parameters. This paper focuses on newly developed integrated tools in VisIVO Server allowing intuitive visual discovery with 3D views being created from data tables. VisIVO Server can be installed easily on any web server with a database repository. We discuss briefly aspects of our implementation of VisiVO Server on a computational grid and also outline the functionality of the services offered by VisIVO Web. Finally we conclude with a summary of our work and pointers to future developments.

[ascl:1402.001] Vissage: ALMA VO Desktop Viewer

Vissage (VISualisation Software for Astronomical Gigantic data cubEs) is a FITS browser primarily targeting FITS data cubes obtained from ALMA. Vissage offers basic functionality for viewing three-dimensional data cubes, integrated intensity map, flipbook, channel map, and P-V diagram. It has several color sets and color scales available, offers panning and zooming, and can connect with the ALMA WebQL system and the JVO Subaru Image Cutout Service.

[ascl:1701.002] Vizic: Jupyter-based interactive visualization tool for astronomical catalogs

Vizic is a Python visualization library that builds the connection between images and catalogs through an interactive map of the sky region. The software visualizes catalog data over a custom background canvas using the shape, size and orientation of each object in the catalog and displays interactive and customizable objects in the map. Property values such as redshift and magnitude can be used to filter or apply colormaps, and objects can be selected for further analysis through standard Python functions from inside a Jupyter notebook.

Vizic allows custom overlays to be appended dynamically on top of the sky map; included are Voronoi, Delaunay, Minimum Spanning Tree and HEALPix layers, which are helpful for visualizing large-scale structure. Overlays can be generated, added or removed dynamically with one line of code. Catalog data is kept in a non-relational database. The Jupyter Notebook allows the user to create scripts to analyze and plot the data selected/displayed in the interactive map, making Vizic a powerful and flexible interactive analysis tool. Vizic be used for data inspection, clustering analysis, galaxy alignment studies, outlier identification or simply large-scale visualizations.

[ascl:2207.020] vKompth: Time-dependent Comptonization model for black-hole X-ray binaries

vKompth fits the energy-dependent rms-amplitude and phase-lag spectra of low-frequency quasi-periodic oscillations in low mass black-hole X-ray binaries using a variable Comptonization model. The accretion disc is modeled as a multi-temperature blackbody source emitting soft photons which are then Compton up-scattered in a spherical corona, including feedback of Comptonized photons that return to the disc.

[ascl:1908.014] Vlasiator: Hybrid-Vlasov simulation code

Vlasiator is a 6-dimensional Vlasov theory-based simulation. It simulates the entire near-Earth space at a global scale using the kinetic hybrid-Vlasov approach, to study fundamental plasma processes (reconnection, particle acceleration, shocks), and to gain a deeper understanding of space weather.

[ascl:2009.002] vlt-sphere: Automatic VLT/SPHERE data reduction and analysis

The high-contrast imager SPHERE at the Very Large Telescope combines extreme adaptive optics and coronagraphy to directly image exoplanets in the near-infrared. The vlt-sphere package enables easy reduction of the data coming from IRDIS and IFS, the two near-infrared subsystems of SPHERE. The package relies on the official ESO pipeline (ascl:1402.010), which must be installed separately.

[ascl:1304.005] VOBOZ/ZOBOV: Halo-finding and Void-finding algorithms

VOBOZ (VOronoi BOund Zones) is an algorithm to find haloes in an N-body dark matter simulation which has little dependence on free parameters.

ZOBOV (ZOnes Bordering On Voidness) is an algorithm that finds density depressions in a set of points without any free parameters or assumptions about shape. It uses the Voronoi tessellation to estimate densities to find both voids and subvoids. It also measures probabilities that each void or subvoid arises from Poisson fluctuations.

[ascl:1411.003] voevent-parse: Parse, manipulate, and generate VOEvent XML packets

voevent-parse, written in Python, parses, manipulates, and generates VOEvent XML packets; it is built atop lxml.objectify. Details of transients detected by many projects, including Fermi, Swift, and the Catalina Sky Survey, are currently made available as VOEvents, which is also the standard alert format by future facilities such as LSST and SKA. However, working with XML and adhering to the sometimes lengthy VOEvent schema can be a tricky process. voevent-parse provides convenience routines for common tasks, while allowing the user to utilise the full power of the lxml library when required. An earlier version of voevent-parse was part of the pysovo (ascl:1411.002) library.

[ascl:1811.016] VoigtFit: Absorption line fitting for Voigt profiles

VoigtFit fits Voigt profiles to absorption lines. It fits multiple components for various atomic lines simultaneously, allowing parameters to be tied and fixed, and can automatically fit a polynomial continuum model together with the line profiles. A physical model can be used to constrain thermal and turbulent broadening of absorption lines as well as implementing molecular excitation models. The code uses a χ2 minimization approach to find the best solution and offers interactive features such as manual continuum placement locally around each line, manual masking of undesired fitting regions, and interactive definition of velocity components for various elements, improving the ease of estimating initial guesses.

[ascl:2109.003] VOLKS2: VLBI Observation for transient Localization Keen Searcher

The VOLK2 (VLBI Observation for transient Localization Keen Searcher) pipeline conducts single pulse searches and localization in regular VLBI observations as well as single pulse detections from known sources in dedicated observations. In VOLKS2, the search and localization are two independent steps. The search step takes the idea of geodetic VLBI post processing, which fully utilizes the cross spectrum fringe phase information to maximize the signal power. Compared with auto spectrum based method, it is able to extract single pulses from highly RFI contaminated data. The localization uses the geodetic VLBI solving methods, which derives the single pulse location by solving a set of linear equations given the relation between the residual delay and the offset to a priori position.

[ascl:1309.007] VOMegaPlot: Plotting millions of points

VOMegaPlot, a Java based tool, has been developed for visualizing astronomical data that is available in VOTable format. It has been specifically optimized for handling large number of points (in the range of millions). It has the same look and feel as VOPlot (ascl:1309.006) and both these tools have certain common functionality.

[ascl:1309.006] VOPlot: Toolkit for Scientific Discovery using VOTables

VOPlot is a tool for visualizing astronomical data. It was developed in Java and acts on data available in VOTABLE, ASCII and FITS formats. VOPlot is available as a stand alone version, which is to be installed on the user's machine, or as a web-based version fully integrated with the VizieR database.

[ascl:1211.006] VorBin: Voronoi binning method

VorBin (Voronoi binning method) bins two-dimensional data to a constant signal-to-noise ratio per bin. It optimally solves the problem of preserving the maximum spatial resolution of general two-dimensional data, given a constraint on the minimum signal-to-noise ratio. The method is available in both IDL and Python.

[ascl:2410.003] vortex-p: Helmholtz-Hodge and Reynolds decomposition algorithm for particle-based simulations

vortex-p analyzes the velocity fields of astrophysical simulations of different natures (for example, SPH, moving-mesh, and meshless) usually spanning many orders of magnitude in scales involved. The code performs Helmholtz-Hodge decomposition (HHD); that is, it can decompose the velocity field into a solenoidal and an irrotational/compressive part Helmholtz-Hodge decomposition. vortex-p internally uses an AMR representation of the velocity field and can, in principle, capture the full dynamical range of the simulation. The package can also perform Reynolds decomposition (i.e., the decomposition of the velocity field into a bulk and a turbulent part). This is achieved by means of a multi-scale filtering of the velocity field, where the filtering scale around each point is determined by the local flow properties. vortex-p expands the vortex (ascl:2206.001) code, which had been coupled to the outputs of the MASCLET code, to a fully stand-alone tool capable of working with the outcomes of a broad range of simulation methods.

[ascl:2206.001] vortex: Helmholtz-Hodge decomposition for an AMR velocity field

vortex performs a Helmholtz-Hodge decomposition on vector fields defined on AMR grids, decomposing a vector field in its solenoidal (divergence-less) and compressive (curl-less) parts. It works natively on vector fields defined on Adaptive Mesh Refinement (AMR) grids, so that it can perform the decomposition over large dynamical ranges; it is also applicable to particle-based simulations. As vortex is devised primarily to investigate the properties of the turbulent velocity field in the Intracluster Medium (ICM), it also includes routines for multi-scale filtering the velocity field.

[ascl:1205.011] VOSpec: VO Spectral Analysis Tool

VOSpec is a multi-wavelength spectral analysis tool with access to spectra, theoretical models and atomic and molecular line databases registered in the VO. The standard tools of VOSpec include line and continuum fitting, redshift and reddening correction, spectral arithmetic and convolution between spectra, equivalent width and flux calculations, and a best fitting algorithm for fitting selected SEDs to a TSAP service. VOSpec offers several display modes (tree vs table) and organising functionalities according to the available metadata for each service, including distance from the observation position.

[ascl:1309.008] VOStat: Statistical analysis of astronomical data

VOStat allows astronomers to use both simple and sophisticated statistical routines on large datasets. This tool uses the large public-domain statistical computing package R. Datasets can be uploaded in either ASCII or VOTABLE (preferred) format. The statistical computations are performed by the VOStat and results are returned to the user.

[ascl:1408.015] VPFIT: Voigt profile fitting program

The VPFIT program fits multiple Voigt profiles (convolved with the instrument profiles) to spectroscopic data that is in FITS or an ASCII file. It requires CFITSIO (ascl:1010.001) and PGPLOT (ascl:1103.002); the tarball includes RDGEN (ascl:1408.017), which can be used with VPFIT to set up the fits, fit the profiles, and examine the result in interactive mode for setting up initial guesses; vpguess (ascl:1408.016) can also be used to set up an initial file.

[ascl:1408.016] vpguess: Fitting multiple Voigt profiles to spectroscopic data

vpguess facilitates the fitting of multiple Voigt profiles to spectroscopic data. It is a graphical interface to VPFIT (ascl:1408.015). Originally meant to simplify the process of setting up first guesses for a subsequent fit with VPFIT, it has developed into a full interface to VPFIT. It may also be used independently of VPFIT for displaying data, playing around with data and models, "chi-by-eye" fits, displaying the result of a proper fit, pretty plots, etc. vpguess is written in C, and the graphics are based on PGPLOT (ascl:1103.002).

[ascl:1811.017] VPLanet: Virtual planet simulator

VPLanet (Virtual Planetary Laboratory) simulates planetary system evolution with a focus on habitability. Physical models, typically consisting of ordinary differential equations for stellar, orbital, tidal, rotational, atmospheric, internal, magnetic, climate, and galactic evolution, are coupled together to simulate evolution for the age of a system.

[ascl:1407.013] VStar: Variable star data visualization and analysis tool

VStar is a multi-platform, easy-to-use variable star data visualization and analysis tool. Data for a star can be read from the AAVSO (American Association of Variable Star Observers) database or from CSV and TSV files. VStar displays light curves and phase plots, can produce a mean curve, and analyzes time-frequency with Weighted Wavelet Z-Transform. It offers tools for period analysis, filtering, and other functions.

[ascl:1704.011] VULCAN: Chemical Kinetics For Exoplanetary Atmospheres

VULCAN describes gaseous chemistry from 500 to 2500 K using a reduced C-H-O chemical network with about 300 reactions. It uses eddy diffusion to mimic atmospheric dynamics and excludes photochemistry, and can be used to examine the theoretical trends produced when the temperature-pressure profile and carbon-to-oxygen ratio are varied.

[ascl:1710.001] vysmaw: Fast visibility stream muncher

The vysmaw client library facilitates the development of code for processes to tap into the fast visibility stream on the National Radio Astronomy Observatory's Very Large Array correlator back-end InfiniBand network. This uses the vys protocol to allow loose coupling to clients that need to remotely access memory over an Infiniband network.

[ascl:2301.021] WALDO: Waveform AnomaLy DetectOr

WALDO (Waveform AnomaLy DetectOr) flags possible anomalous Gravitational Waves from Numerical Relativity catalogs using deep learning. It uses a U-Net architecture to learn the waveform features of a dataset. After computing the mismatch between those waveforms and the neural predictions, WALDO isolates high mismatch evaluations for anomaly search.

[ascl:2108.004] WaldoInSky: Anomaly detection algorithms for time-domain astronomy

WaldoInSky finds anomalous astronomical light curves and their analogs. The package contains four methods: an adaptation of the Unsupervised Random Forest for anomaly detection in light curves that operates on the light curve points and their power spectra; two manifold-learning methods (the t-SNE and UMAP) that operate on the DMDT maps (image representations of the light curves), and that can be used to find analog light curves in the low-dimensional representation; and an Isolation Forest method for evaluating approaches of light curve pre-processing, before they are passed to the anomaly detectors. WaldoInSky also contain code for random sparsification of light curves.

[ascl:2207.019] walter: Predictor for the number of resolved stars in a given observation from RST

walter calculates the number density of stars detected in a given observation aiming to resolve a stellar population. The code also calculates the exposure time needed to reach certain population features, such as the horizontal branch, and provides an estimate of the crowding limit. walter was written with the expectation that such calculations will be very useful for planning surveys with the Nancy Grace Roman Space Telescope (RST, formerly WFIRST).

[ascl:1807.002] Warpfield: Winds And Radiation Pressure: Feedback Induced Expansion, colLapse and Dissolution

Warpfield (Winds And Radiation Pressure: Feedback Induced Expansion, colLapse and Dissolution) calculates shell dynamics and shell structure simultaneously for isolated massive clouds (≥105 M). This semi-analytic 1D feedback model scans a large range of physical parameters (gas density, star formation efficiency, and metallicity) to estimate escape fractions of ionizing radiation fesc, I, the minimum star formation efficiency ∊min required to drive an outflow, and recollapse time-scales for clouds that are not destroyed by feedback.

[ascl:2307.038] WarpX: Time-based electromagnetic and electrostatic Particle-In-Cell code

WarpX is an advanced electromagnetic & electrostatic Particle-In-Cell code. It supports many features including Perfectly-Matched Layers (PML), mesh refinement, and the boosted-frame technique. A highly-parallel and highly-optimized code, WarpX can run on GPUs and multi-core CPUs, includes load balancing capabilities, and scales to the largest supercomputers.

[ascl:2206.024] Wavetrack: Arbitrary time-evolving solar object recognition and tracking

Wavetrack recognizes and tracks CME shock waves, filaments, and other solar objects. The code creates base images by averaging а series of images a few minutes prior to the start of the eruption and constructs base difference images by subtracting base images from the current raw image of the sequence. This enhances the change in intensity caused by coronal bright fronts, omits static details, and reduces noise. Wavetrack then chooses an appropriate intensity interval and decomposes the base difference or running difference image with an A-Trous wavelet transform, where each wavelet coefficient is obtained by convolving the image array with a corresponding iteration of the wavelet kernel. When the maximum value of the wavelet coefficients for a connected set of pixels satisfies certain conditions, this region is considered as a structure on the respective wavelet coefficient. Separate stand-alone object masks are obtained with a clustering algorithm and objects are renumbered according to the number of the quadrant they belong at each iteration.

[ascl:2311.001] wcpy: Wavelength Calibrator

The graphical user interface Wavelength Calibrator facilitates wavelength calibration. Although developed for astronomical data reduction, it can also be used in any place where wavelength calibration is needed.

[ascl:1108.003] WCSLIB and PGSBOX

WCSLIB is a C library, supplied with a full set of Fortran wrappers, that implements the "World Coordinate System" (WCS) standard in FITS (Flexible Image Transport System). It also includes a PGPLOT-based routine, PGSBOX, for drawing general curvilinear coordinate graticules and a number of utility programs.

[ascl:1109.015] WCSTools: Image Astrometry Toolkit

WCSTools is a package of programs and a library of utility subroutines for setting and using the world coordinate systems (WCS) in the headers of the most common astronomical image formats, FITS and IRAF .imh, to relate image pixels to sky coordinates. In addition to dealing with image WCS information, WCSTools has extensive catalog search, image header manipulation, and coordinate and time conversion tasks. This software is all written in very portable C, so it should compile and run on any computer with a C compiler.

[ascl:2004.004] WD: Wilson-Devinney binary star modeling

Wilson-Devinney binary star modeling code (WD) is a complete package for modeling binary stars and their eclipes and consists of two main modules. The LC module generates light and radial velocity curves, spectral line profiles, images, conjunction times, and timing residuals; the DC module handles differential corrections, performing parameter adjustment of light curves, velocity curves, and eclipse timings by the Least Squares criterion. WD handles eccentric orbits and asynchronous rotation, and can compute velocity curves (with proximity and eclipse effects). It offers options for detailed reflection and nonlinear (logarithmic law) limb darkening, adjustment of spot parameters, an optional provision for spots to drift over the surface, and can follow light curve development over large numbers of orbits. Absolute flux solution allow Direct Distance Estimation (DDE) and there are improved solutions for ellipsoidal variables and for eclipsing binaries (EBs) with very shallow eclipses. Absolute flux solutions also can estimate temperatures of both EB components under suitable circumstances.

[ascl:1806.012] WDEC: White Dwarf Evolution Code

WDEC (White Dwarf Evolution Code), written in Fortran, offers a fast and fairly easy way to produce models of white dwarfs. The code evolves hot (~100,000 K) input models down to a chosen effective temperature by relaxing the models to be solutions of the equations of stellar structure. The code can also be used to obtain g-mode oscillation modes for the models.

[ascl:1807.020] wdmerger: Simulate white dwarf mergers with CASTRO

wdmerger simulates binary white dwarf mergers (and related events) in CASTRO (ascl:1105.010) and provides useful information on the viability of mergers of white dwarfs as a progenitor for Type Ia supernovae.

[ascl:2307.037] WDMWaveletTransforms: Fast forward and inverse WDM wavelet transforms

WDMWaveletTransforms implements the fast forward and inverse WDM wavelet transforms in Python from both the time and frequency domains. The frequency domain transforms are inherently faster and more accurate. The wavelet domain->frequency domain and frequency domain->wavelet domain transforms are nearly exact numerical inverses of each other for a variety of inputs tested, including Gaussian random noise. WDMWaveletTransforms has both command line and Python interfaces.

[ascl:2206.012] WDPhotTools: White Dwarf Photometric SED fitter and luminosity function builder

WDPhotTools generates color-color diagrams and color-magnitude diagrams in various photometric systems, plots cooling profiles from different models, and computes theoretical white dwarf luminosity functions based on the built-in or supplied models of the (1) initial mass function, (2) total stellar evolution lifetime, (3) initial-final mass relation, and (4) white dwarf cooling time. The software has three main parts: the formatters that handle the output models from various works in the format as they are downloaded; the photometric fitter that solves for the WD parameters based on the photometry, with or without distance and reddening; and the generator of the white dwarf luminosity function in bolometric magnitudes or in any of the photometric systems available from the atmosphere model.

[ascl:2007.013] wdtools: Spectroscopic analysis of white dwarfs

wdtools characterizes the atmospheric parameters of white dwarfs using spectroscopic data. The flagship class is the generative fitting pipeline (GFP), which fits ab initio theoretical models to observed spectra in a Bayesian framework using high-speed neural networks to interpolate synthetic spectra.

[ascl:2206.016] wdwarfdate: White dwarfs age calculator

wdwarfdate derives the Bayesian total age of a white dwarf from an effective temperature and a surface gravity. It runs a chain of models assuming single star evolution and estimates the following parameters and their uncertainties: total age of the object, mass and cooling age of the white dwarf, and mass and lifetime of the progenitor star.

[ascl:2109.021] WeakLensingDeblending: Weak lensing fast simulations and analysis of blended objects

WeakLensingDeblending provides weak lensing fast simulations and analysis for the LSST Dark Energy Science Collaboration. It is used to study the effects of overlapping sources on shear estimation, photometric redshift algorithms, and deblending algorithms. Users can run their own simulations (of LSST and other surveys) or download the galaxy catalog and simulation outputs to use with their own code.

[ascl:2307.051] WeakLensingQML: Quadratic Maximum Likelihood estimator applied to Weak Lensing

WeakLensingQML implements the Quadratic Maximum Likelihood (QML) estimator and applies it to simulated cosmic shear data and compares the results to a Pseudo-Cl implementation. The package computes and saves relevant data files for later processes, such as the fiduciary cosmic shear power spectrum used in the analysis, the sky mask, and computing an analytic version of the QML's covariance matrix. The core of the package implements a conjugate-gradient approach for the quadratic estimator, and is parallelized for maximum performance. The code relies on the Eigen linear algebra package and the HealPix spherical harmonic transform library. A post-processing script analyzes the results and compares the QML's estimates with those from the Pseudo-Cl estimator; it then produces an array of plots highlighting the results.

[ascl:1504.007] WebbPSF: James Webb Space Telescope PSF Simulation Tool

WebbPSF provides a PSF simulation tool in a flexible and easy-to-use software package implemented in Python. Functionality includes support for spectroscopic modes of JWST NIRISS, MIRI, and NIRSpec, including modeling of slit losses and diffractive line spread functions.

[ascl:1609.007] Weighted EMPCA: Weighted Expectation Maximization Principal Component Analysis

Weighted EMPCA performs principal component analysis (PCA) on noisy datasets with missing values. Estimates of the measurement error are used to weight the input data such that the resulting eigenvectors, when compared to classic PCA, are more sensitive to the true underlying signal variations rather than being pulled by heteroskedastic measurement noise. Missing data are simply limiting cases of weight = 0. The underlying algorithm is a noise weighted expectation maximization (EM) PCA, which has additional benefits of implementation speed and flexibility for smoothing eigenvectors to reduce the noise contribution.

[ascl:1010.042] WeightMixer: Hybrid Cross-power Spectrum Estimation

This code, which requires HEALPix 2.x (ascl:1107.018), allows you to generate power spectrum estimators from WMAP 5-year maps and generate hybrid cross- and auto- power spectrum and covariance from general foreground-cleaned maps. In addition, it allows you to simulate combined maps or combinations of maps for individual detectors and do MPI spherical transforms of arrays of maps, calculate coupling matrices etc. The code includes all of LensPix (ascl:1102.025), the MPI framework used for doing spherical transforms (based on HealPix).

[ascl:1010.069] WeightWatcher: Code to Produce Control Maps

WeightWatcher is a program that combines weight-maps, flag-maps and polygon data in order to produce control maps which can directly be used in astronomical image-processing packages like Drizzle, SWarp or SExtractor.

[ascl:1705.015] WeirdestGalaxies: Outlier Detection Algorithm on Galaxy Spectra

WeirdestGalaxies finds the weirdest galaxies in the Sloan Digital Sky Survey (SDSS) by using a basic outlier detection algorithm. It uses an unsupervised Random Forest (RF) algorithm to assign a similarity measure (or distance) between every pair of galaxy spectra in the SDSS. It then uses the distance matrix to find the galaxies that have the largest distance, on average, from the rest of the galaxies in the sample, and defined them as outliers.

[ascl:2301.003] WF4Py: Gravitational waves waveform models in pure Python language

WF4Py implements frequency-domain gravitational wave waveform models in pure Python, thus enabling parallelization over multiple events at a time. Waveforms in WF4Py are built as classes; the functions take dictionaries containing the parameters of the events to analyze as input and provide Fourier domain waveform models. All the waveforms are accurately checked with their implementation in LALSuite (ascl:2012.021) and are a core element of GWFAST (ascl:2212.001).

[ascl:1404.013] WFC3UV_GC: WFC3 UVIS geometric-distortion correction

WFC3UV_GC is an improved geometric-distortion solution for the Hubble Space Telescope UVIS channel of Wide Field Camera 3 for ten broad-band filters. The solution is made up of three parts:

1.) a 3rd-order polynomial to deal with the general optical distortion;
2.) a table of residuals that accounts for both chip-related anomalies and fine-structure introduced by the filter; and,
3.) a linear transformation to put the two chips into a convenient master frame.

[ascl:2101.003] whereistheplanet: Predicting positions of directly imaged companions

whereistheplanet predicts the locations of directly imaged companions (mainly exoplanets and brown dwarfs) based on past orbital fits to the data. This tool helps coordinate follow-up observations to characterize their properties, as precise pointing of the instrument is often needed. It uses orbitize! (ascl:1910.009) as a backend. whereistheplanet is available as a Python API, a command line tool, and a web form at whereistheplanet.com.

[ascl:1911.018] WhereWolf: Galaxy/(sub)Halo ghosting tool for N-body simulations

WhereWolf tracks (sub)haloes even if they have been lost by a halo finder in cosmological simulations and supplements halo catalogs such as VELOCIraptor (ascl:1911.020) with these recovered (sub)haloes. The code can improve measurements of the subhalo/halo mass function and present estimates of the distribution of radii at which subhaloes merge.

[ascl:1010.084] WhiskyMHD: Numerical Code for General Relativistic Magnetohydrodynamics

Whisky is a code to evolve the equations of general relativistic hydrodynamics (GRHD) and magnetohydrodynamics (GRMHD) in 3D Cartesian coordinates on a curved dynamical background. It was originally developed by and for members of the EU Network on Sources of Gravitational Radiation and is based on the Cactus Computational Toolkit. Whisky can also implement adaptive mesh refinement (AMR) if compiled together with Carpet.

Whisky has grown from earlier codes such as GR3D and GRAstro_Hydro, but has been rewritten to take advantage of some of the latest research performed here in the EU. The motivation behind Whisky is to compute gravitational radiation waveforms for systems that involve matter. Examples would include the merger of a binary system containing a neutron star, which are expected to be reasonably common in the universe and expected to produce substantial amounts of radiation. Other possible sources are given in the projects list.

[ascl:2203.030] Wigglewave: Linearized governing equations solver

Wigglewave uses a finite difference method to solve the linearized governing equations for a torsion Alfvèn wave propagating in a plasma with negligible plasma beta and in a force-free axisymmetric magnetic field with no azimuthal component embedded in a high density divergent tube structure. Wigglewave is fourth order in time and space using a fourth-order central difference scheme for calculating spatial derivatives and a fourth-order Runge-Kutta (RK4) scheme for updating at each timestep. The solutions calculated are the perturbations to the velocity, v and to the magnetic field, b. All variables are calculated over a uniform grid in radius r and height z.

[ascl:2404.007] WignerFamilies: Compute families of wigner symbols with recurrence relations

WignerFamilies generates families of Wigner 3j and 6j symbols by recurrence relation. These exact methods are orders of magnitude more efficient than strategies such as prime factorization for problems which require every non-trivial symbol in a family, and are very useful for large quantum numbers. WignerFamilies is thread-safe and very fast, beating the standard Fortran routine DRC3JJ from SLATEC by a factor of 2-4.

[ascl:2112.010] WIMpy_NREFT: Dark Matter direct detection rates detector

WIMpy_NREFT (also known as WIMpy) calculates Dark Matter-Nucleus scattering rates in the framework of non-relativistic effective field theory (NREFT). It currently supports operators O1 to O11, as well as millicharged and magnetic dipole Dark Matter. It can be used to generate spectra for Xenon, Argon, Carbon, Germanium, Iodine and Fluorine targets. WIMpy_NREFT also includes functionality to calculate directional recoil spectra, as well as signals from coherent neutrino-nucleus scattering (including fluxes from the Sun, atmosphere and diffuse supernovae).

[ascl:2109.013] WimPyDD: WIMP direct–detection rates predictor

WimPyDD calculates accurate predictions for the expected rates in WIMP direct–detection experiments within the framework of Galilean–invariant non–relativistic effective theory. The object–oriented customizable Python code handles different scenarios including inelastic scattering, WIMP of arbitrary spin, and a generic velocity distribution of WIMP in the Galactic halo.

[ascl:9910.007] WINGSPAN: A WINdows Gamma-ray SPectral Analysis program

WINGSPAN is a program written to analyze spectral data from the Burst and Transient Source Experiment (BATSE) on NASA's Compton Gamma-Ray Observatory. Data files in the FITS (BFITS) format are suitable for input into the program. WINGSPAN can be used to view and manipulate event time histories or count spectra, and also has the capability to perform spectral deconvolution via a standard forward folding model fitting technique (Levenberg-Marquardt algorithm). Although WINGSPAN provides many functions for data manipulation, the program was designed to allow users to easily plug in their own external IDL routines. These external routines have access to all data read from the FITS files, as well as selection intervals created in the main part of WINGSPAN (background intervals and model, etc).

[ascl:2406.029] WinNet: Flexible, multi-purpose, single-zone nuclear reaction network

WinNet, a single zone nuclear reaction network, calculates many different nucleosynthesis processes, including r-process, nup-process, and explosive nucleosynthesis, and many more). It reads in a user-defined file with runtime parameters, then chooses the evolution mode, which is dependent on temperature. The temperature, density, and neutrino quantities are updated, after which the reaction network equations are solved numerically. If convergence is not achieved, the step size is halved and the iteration is repeated. Once convergence is reached, the output is generated and the time is evolved; the final output such as the final abundances and mass fractions are written.

[ascl:2409.017] WISE2MBH: Mass of supermassive black holes estimator

WISE2MBH uses infrared cataloged data from the Wide-field Infrared Survey Explorer (WISE) to estimate the mass of supermassive black holes (SMBH). It implements a Monte Carlo approach for error propagation, considering mean photometric errors from WISE magnitudes, errors in fits of scaling relations used and scatter of those relations, if available.

[ascl:1806.004] WiseView: Visualizing motion and variability of faint WISE sources

WiseView renders image blinks of Wide-field Infrared Survey Explorer (WISE) coadds spanning a multi-year time baseline in a browser. The software allows for easy visual identification of motion and variability for sources far beyond the single-frame detection limit, a key threshold not surmounted by many studies. WiseView transparently gathers small image cutouts drawn from many terabytes of unWISE coadds, facilitating access to this large and unique dataset. Users need only input the coordinates of interest and can interactively tune parameters including the image stretch, colormap and blink rate. WiseView was developed in the context of the Backyard Worlds: Planet 9 citizen science project, and has enabled hundreds of brown dwarf candidate discoveries by citizen scientists and professional astronomers.

[ascl:1812.001] WISP: Wenger Interferometry Software Package

WISP (Wenger Interferometry Software Package) is a radio interferometry calibration, reduction, imaging, and analysis package. WISP is a collection of Python code implemented through CASA (ascl:1107.013). Its generic and modular framework is designed to handle any continuum or spectral line radio interferometry data.

[ascl:1204.001] WM-basic: Modeling atmospheres of hot stars

WM-basic is an easy-to-use interface to a program package which models the atmospheres of Hot Stars (and also SN and GN). The release comprises all programs required to calculate model atmospheres which especially yield ionizing fluxes and synthetic spectra. WM-basic is a native 32-bit application, conforming to the Multiple Documents Interface (MDI) standards for Windows XP/2000/NT/9x. All components of the program package have been compiled with Digital Visual Fortran V6.6(Pro) and Microsoft Visual C++.

[ascl:1312.002] WND-CHARM: Multi-purpose image classifier

WND-CHARM quantitatively analyzes morphologies of galaxy mergers and associate galaxies by their morphology. It computes a large set (up to ~2700) of image features for each image based on the WND-CHARM algorithm. It can then split the images into training and test sets and classify them. The software extracts the image content descriptor from raw images, image transforms, and compound image transforms. The most informative features are then selected, and the feature vector of each image is used for classification and similarity measurement using Fisher discriminant scores and a variation of Weighted Nearest Neighbor analysis. WND-CHARM's results comparable favorably to the performance of task-specific algorithms developed for tested datasets. The simple user interface allows researchers who are not knowledgeable in computer vision methods and have no background in computer programming to apply image analysis to their data.

[ascl:2011.012] wobble: Time-series spectra analyzer

wobble analyzes time-series spectra. It was designed with stabilized extreme precision radial velocity (EPRV) spectrographs in mind, but is highly flexible and extensible to a variety of applications. It takes a data-driven approach to deriving radial velocities and requires no a priori knowledge of the stellar spectrum or telluric features.

[ascl:1212.007] WOLF: FITS file processor

WOLF processes FITS files and generates photometry files, annotated JPGs, opacity maps, background, transient detection and luminance changes detection. This software was used to process data for the Night Sky Live project.

[ascl:1204.014] WOMBAT: sWift Objects for Mhd BAsed on Tvd

WOMBAT (sWift Objects for Mhd BAsed on Tvd) is an astrophysical fluid code that is an implementation of a non-relativistic MHD TVD scheme; an extension for relativistic MHD has been added. The code operates on 1, 2, and 3D Eulerian meshes (cartesian and cylindrical coordinates) with magnetic field divergence restriction controlled by a constrained transport (CT) scheme. The user can tune code performance to a given processor based on chip cache sizes. Proper settings yield significant speed-ups due to efficient cache reuse.

[submitted] World Observatory

World Observatory visualizes S/N-versus-cost tradeoffs for large optical and near-infrared telescopes. Both mid-latitude and Arctic/Antarctic sites can be considered; the intent is a simple simulation to grow intuition for where major capital costs lie relative to key observatory design choices, and against expected scientific performance at various sites. User-defined unit costs for (a possibly "effective") roadway, enclosure, aperture, focal length, and adaptive optics can be scaled up for polar sites, and down for better seeing and lower sky brightness in K-band. Observatory models and results are immediately displayed side-by-side. Either point-source-detection S/N or recovery of bulge-to-total ratios in a simulated galaxy survey are divided by the total project cost, thus providing a universal metric.

[ascl:1907.030] Wōtan: Stellar detrending methods

Wōtan provides free and open source algorithms to remove trends from time-series data automatically as an aid to search efficiently for transits in stellar light curves from surveys. The toolkit helps determine empirically the best tool for a given job, serving as a one-stop solution for various smoothing tasks.

[ascl:2112.023] wpca: Weighted Principal Component Analysis in Python

wpca, written in Python, offers several implementations of Weighted Principal Component Analysis and uses an interface similar to scikit-learn's sklearn.decomposition.PCA. Implementations include a direct decomposition of a weighted covariance matrix to compute principal vectors, and then a weighted least squares optimization to compute principal components, and an iterative expectation-maximization approach to solve simultaneously for the principal vectors and principal components of weighted data. It also includes a standard non-weighted PCA implemented using the singular value decomposition, primarily to be useful for testing.

[ascl:1304.004] Wqed: Lightcurve Analysis Suite

Wqed (pronounced "Wicked") is a set of tools developed by the Delaware Asteroseismic Research Center (DARC) to simplify the process of reducing time-series CCD data on variable stars. It does not provide tools to measure the brightness of stars in individual frames, focusing instead on what comes next:

    - selecting and removing data lost to cloud,
    - removing the effects of light cloud and seeing variations,
    - keeping track of what star a given data set refers to, and when that data was taken, and
    - performing barycentric corrections to data.

[ascl:1408.023] WSClean: Widefield interferometric imager

WSClean (w-stacking clean) is a fast generic widefield imager. It uses the w-stacking algorithm and can make use of the w-snapshot algorithm. It supports full-sky imaging and proper beam correction for homogeneous dipole arrays such as the MWA. WSClean allows Hogbom and Cotton-Schwab cleaning, and can clean polarizations joinedly. All operations are performed on the CPU; it is not specialized for GPUs.

[ascl:1010.071] WSHAPE: Gravitational Softening and Adaptive Mass Resolution

Pairwise forces between particles in cosmological N-body simulations are generally softened to avoid hard collisions. Physically, this softening corresponds to treating the particles as diffuse clouds rather than point masses. For particles of unequal mass (and hence unequal softening length), computing the softened force involves a nontrivial double integral over the volumes of the two particles. We show that Plummer force softening is consistent with this interpretation of softening while spline softening is not. We provide closed-form expressions and numerical implementation for pairwise gravitational force laws for pairs of particles of general softening scales $epsilon_1$ and $epsilon_2$ assuming the commonly used cloud profiles: NGP, CIC, TSC, and PQS. Similarly, we generalize Plummer force law into pairs of particles of general softenings. We relate our expressions to the gaussian, Plummer and spline force softenings known from literature. Our expressions allow possible inclusions of pointlike particles such as stars or supermassive black holes.

[ascl:1402.029] wssa_utils: WSSA 12 micron dust map utilities

wssa_utils contains utilities for accessing the full-sky, high-resolution maps of the WSSA 12 micron data release. Implementations in both Python and IDL are included. The code allows users to sample values at (longitude, latitude) coordinates of interest with ease, transparently mapping coordinates to WSSA tiles and performing interpolation. The wssa_utils software also serves to define a unique WSSA 12 micron flux at every location on the sky.

[ascl:2209.013] wsynphot: Synthetic photometry package using quantities

wsynphot provides a broad set of filters, including observation facility, instrument, and wavelength range, and functions for imaging stars to produce a filter curve showing the transmission of light for each wavelength value. It can create a filter curve object, plot the curve, and allows the user to do calculations on the filter curve object.

[ascl:1207.014] wvrgcal: Correction of atmospheric phase fluctuations in ALMA observations

wvrgcal is a command line front end to LibAIR, the atmospheric inference library for phase correction of ALMA data using water vapour radiometers, and is the user-facing application for calculating atmospheric phase correction from WVR data. wvrgcal outputs a CASA gain calibration table which can then be applied to the observed data in the usual way.

Note: wvrgcal has been incorporated into the NRAO CASA suite.

[ascl:1211.003] WVT Binning: Spatially adaptive 2-D binning

WVT Binning is a spatially adaptive 2-dimensional binning algorithm designed to bin sparse X-ray data. It can handle background subtracted, exposure corrected data to produce intensity images, hardness ratio maps, or temperature maps. The algorithm is an extension of Cappellari & Copin's (2003) Voronoi binning code and uses Weighted Voronoi Tesselations (WVT) to produce a very compact binning structure with a constant S/N per bin. The bin size adjusts to the required resolution in single-pixel steps, which minimizes the scatter around the target S/N. The code is very versatile and can in principle be applied to any type of data. The user manual contains instructions on how to apply the WVT binning code to X-ray data and how to extend the algorithm to other problems.

[ascl:1909.011] WVTICs: SPH initial conditions using Weighted Voronoi Tesselations

WVTICs generates glass-like initial conditions for Smoothed Particle Hydrodynamics. Relaxation of the particle distribution is done using an algorithm based on Weighted Voronoi Tesselations; additional particle reshuffling can be enabled to improve over- and undersampled maxima/minima. The WBTICs package includes a full suite of analytical test problems.

[ascl:2310.003] wwz: Weighted wavelet z-transform code

wwz provides a python3 implementation of the Foster weighted wavelet z-transform, a wavelet-based method for periodicity analysis of unevenly sampled data.

[ascl:1601.019] WzBinned: Binned and uncorrelated estimates of dark energy EOS extractor

WzBinned extracts binned and uncorrelated estimates of dark energy equation of state w(z) using Type Ia supernovae Hubble diagram and other cosmological probes and priors. It can handle an arbitrary number of input distance modulus data (entered as an input file SNdata.dat) and various existing cosmological information.

[ascl:2102.005] X-PSI: X-ray Pulse Simulation and Inference

X-PSI simulates rotationally-modified (pulsed) surface X-ray emission from neutron stars, taking into account relativistic effects on the emitted radiation. This can then be used to perform Bayesian statistical inference on real or simulated astronomical data sets. Model parameters of interest may include neutron star mass and radius (useful to constrain the properties of ultradense nuclear matter) or the system geometry and properties of the hot emitting surface-regions. To achieve this, X-PSI couples code for likelihood functionality (simulation) with existing open-source software for posterior sampling (inference).

[ascl:1312.005] XAssist: Automatic analysis of X-ray astrophysics data

XAssist provides automation of X-ray astrophysics, specifically data reprocessing, source detection, and preliminary spatial, temporal and spectral analysis for each source with sufficient counts, with an emphasis on galaxies. It has been used for data from Chandra, ROSAT, XMM-Newton, and other various projects.

[ascl:1810.016] XCLASS: eXtended CASA Line Analysis Software Suite

XCLASS (eXtended CASA Line Analysis Software Suite) extends CASA (ascl:1107.013) with new functions for modeling interferometric and single dish data. It provides a tool for calculating synthetic spectra by solving the radiative transfer equation for an isothermal object in one dimension, taking into account the finite source size and dust attenuation. It also includes an interface for MAGIX (ascl:1303.009) to find the parameter set that most closely reproduces the data.

[ascl:1907.029] XDF-GAN: Mock astronomical survey generator

XDF-GAN generates mock galaxy surveys with a Spatial Generative Adversarial Network (SGAN)-like architecture. Mock galaxy surveys are generated from data that is preprocessed as little as possible (preprocessing is only a 99.99th percentile clipping). The outputs can also be tessellated together to create a very large survey, limited in size only by the RAM of the generation machine.

[ascl:1708.026] XDGMM: eXtreme Deconvolution Gaussian Mixture Modeling

XDGMM uses Gaussian mixtures to do density estimation of noisy, heterogenous, and incomplete data using extreme deconvolution (XD) algorithms which is compatible with the scikit-learn machine learning methods. It implements both the astroML and Bovy et al. (2011) algorithms, and extends the BaseEstimator class from scikit-learn so that cross-validation methods work. It allows the user to produce a conditioned model if values of some parameters are known.

[ascl:1302.016] XDQSO: Photometic quasar probabilities and redshifts

XDQSO, written in IDL, calculates photometric quasar probabilities to mimick SDSS-III’s BOSS quasar target selection or photometric redshifts for quasars, whether in three redshift ranges (z < 2.2; 2.2 leq z leq 3.5; z > 3.5) or arbitrary redshift ranges.

[ascl:1107.010] XDSPRES: CL-based package for Reducing OSIRIS Cross-dispersed Spectra

The CL-based package XDSPRES is a complete reducing facility for cross-dispersed spectra taken with the Ohio State Infrared Imager/Spectrometer, as installed at the SOAR telescope. This instrument provides spectra in the range between 1.2um and 2.35um in a single exposure, with resolving power of R ~ 1200. XDSPRES consists of two tasks, namely xdflat and doosiris. The former is a completely automated code for preparing normalized flat field images from raw flat field exposures. Doosiris provides a complete reduction pipeline that requires a minimum of user interaction. The user guide explains the general steps towards a fully reduced spectrum.

[ascl:1112.013] XEphem: Interactive Astronomical Ephemeris

XEphem is a scientific-grade interactive astronomical ephemeris package for UNIX-like systems. Written in C, X11 and Motif, it is easily ported to systems. XEphem computes heliocentric, geocentric and topocentric information for all objects and has built-in support for all planets, the moons of Mars, Jupiter, Saturn, Uranus and Earth, central meridian longitude of Mars and Jupiter, Saturn's rings, and Jupiter's Great Red Spot. It allows user-defined objects including stars, deepsky objects, asteroids, comets and Earth satellites, provides special efficient handling of large catalogs including Tycho, Hipparcos, GSC, displays data in configurable tabular formats in conjunction with several interactive graphical views, and displays a night-at-a-glance 24 hour graphic showing when any selected objects are up. It also displays 3-D stereo Solar System views that are particularly well suited for visualizing comet trajectories, quickly finds all close pairs of objects in the sky, and sorts and prints all catalogs with very flexible criteria for creating custom observing lists.

[ascl:1502.018] XFGLENSES: Gravitational lens visualizer

XFGL visualizes gravitational lenses. It has an XFORM GUI and is completely interactive with the mouse. It uses OpenGL for the simulations.

[ascl:2301.012] XGA: Efficient analysis of XMM observations

XGA (X-ray: Generate and Analyse) analyzes X-ray sources observed by the XMM-Newton Space telescope. It is based around declaring different types of source and sample objects which correspond to real X-ray sources, finding all available data, and then insulating the user from the tedious generation and basic analysis of X-ray data products. XGA generates photometric products and spectra for individual sources, or whole samples, with just a few lines of code. Though not a pipeline, pipelines for complex analysis can be built on top of it. XGA provides an easy to use (and parallelized) Python interface with XMM's Science Analysis System (ascl:1404.004), as well as with XSPEC (ascl:9910.005). All XMM products and fit results are read into an XGA source storage structure, thus avoiding the need to leave a Python environment at any point during the analysis. This module also supports more complex analyses for specific object types such as the easy generation of scaling relations, the measurement of gas masses for galaxy clusters, and the PSF correction of images.

[ascl:1807.031] xGDS: Exploration Ground Data Systems

xGDS (Exploration Ground Data Systems) synthesizes real world data (from sensors, robots, ROVs, mobile devices, etc) and human observations into rich, digital maps and displays for analysis, decision making, and collaboration. xGDS processes and maps data (including video) in real-time during operations and uses it to support live role-based geolocated note taking. Notes can be used to search for and display important data. The software enables real-time analysis of data, permitting one to make inferences and plan new data collection operations while still in the field.

[ascl:1511.004] Xgremlin: Interferograms and spectra from Fourier transform spectrometers analysis

Xgremlin is a hardware and operating system independent version of the data analysis program Gremlin used for Fourier transform spectrometry. Xgremlin runs on PCs and workstations that use the X11 window system, including cygwin in Windows. It is used to Fourier transform interferograms, plot spectra, perform phase corrections, perform intensity and wavenumber calibration, and find and fit spectral lines. It can also be used to construct synthetic spectra, subtract continua, compare several different spectra, and eliminate ringing around lines.

[ascl:1704.012] XID+: Next generation XID development

XID+ is a prior-based source extraction tool which carries out photometry in the Herschel SPIRE (Spectral and Photometric Imaging Receiver) maps at the positions of known sources. It uses a probabilistic Bayesian framework that provides a natural framework in which to include prior information, and uses the Bayesian inference tool Stan to obtain the full posterior probability distribution on flux estimates.

[ascl:1303.021] Xmatch: GPU Enhanced Astronomic Catalog Cross-Matching

Xmatch is a cross-platform, multi-GPU tool which allows for extremely fast cross-matching between two Astronomic catalogs. It is capable of asyncronously managing multiple GPUs, ideal for workstation and cluster environments.

[ascl:1402.020] XNS: Axisymmetric equilibrium configuration of neutron stars

XNS solves for the axisymmetric equilibrium configuration of neutron stars in general relativity. It can model differentially rotating and magnetic fields that are either purely toroidal, purely poloidal or in the mixed twisted torus configuration. Einsten's equations are solved using the XCFC approximation for the metric in spherical coordinates.

[ascl:2110.022] XookSuut: Model circular and noncircular flows on 2D velocity maps

XookSuut models circular and noncircular flows on resolved velocity maps. The code performs nonparametric fits to derive kinematic models without assuming analytical functions on the different velocity components of the models. It recovers the circular and radial motions in galaxies in dynamical equilibrium and can derive the noncircular motions induced by oval distortions, such as that produced by stellar bars. XookSuut explores the full space of parameters on a N-dimensional space to derive their mean values; this combined method efficiently recovers the constant parameters and the different kinematic components.

[ascl:1502.019] XPCell: Convective plasma cells simulator

XPCell simulates convective plasma cells. The program is implemented in two versions, one using GNUPLOT and the second OpenGL. XPCell offers a GUI to introduce the parameter required by the program.

[ascl:1212.002] XPHOT: Estimation of properties of weak X-ray sources

XPHOT is an IDL implementation of a non-parametric method for estimating the apparent and intrinsic broad-band fluxes and absorbing X-ray column densities of weak X-ray sources. XPHOT is intended for faint sources with greater than ∼5-7 counts but fewer than 100-300 counts where parametric spectral fitting methods will be superior. This method is similar to the long-standing use of color-magnitude diagrams in optical and infrared astronomy, with X-ray median energy replacing color index and X-ray source counts replacing magnitude. Though XPHOT was calibrated for thermal spectra characteristic of stars in young stellar clusters, recalibration should be possible for some other classes of faint X-ray sources such as extragalactic active galactic nuclei.

[ascl:2301.009] Xpol: Pseudo-Cl power spectrum estimator

Xpol computes angular power spectra based on cross-correlation between maps and covariance matrices. The code is written in C and is fully MPI parallelized in CPU and memory using spherical transform by s2hat (ascl:1110.013). It has been used to derive CMB and dust power spectra for Archeops and CMB, dust, CIB, SZ, SZ-CIB for Planck, among others.

[ascl:1509.001] XSHPipelineManager: Wrapper for the VLT/X-shooter Data Reduction Pipeline

XSHPipelineManager provides a framework for reducing spectroscopic observations taken by the X-shooter spectrograph at the Very Large Telescope. This Python code wraps recipes developed by the European Southern Observatory and runs the full X-shooter data reduction pipeline. The code offers full flexibility in terms of what data reduction recipes to include and which calibration files to use. During the data reduction chain restart-files are saved, making it possible to restart at any step in the chain.

[submitted] Xsmurf - Measuring multifractal properties with the continuous wavelet transform modulus maxima (WTMM) method

Xsmurf is a software package written in C/Tcl/Tk that implements the continuous wavelet transform modulus maxima method, an image processing tool for measuring fractal and multifractal properties in experimental and simulation data.
Multifractal analysis is described in the following page: http://www.scholarpedia.org/article/Wavelet-based_multifractal_analysis

Xsmurf has been used in multiple applications in astrophysics, e.g. :
- analysis of solar magnetograms for characterizing complexity of evolving regions
- fractal/multifractal nature and anisotropic structure of Galactic atomic hydrogen (H I)
- analysis of simulation data (velocity field, ...) of turbulent flow

[ascl:1207.008] xSonify: Sonification software

xSonify maps scientific data to acoustic sequences. Listening to data can help discover patterns in huge amounts of data. Written in Java, xSonify allows visually impaired people to examine numerical data for patterns. The data can be imported from local files or from remote databases via the Internet. Single results of measurements from spacecraft instruments can be selected by their corresponding variables in a specific time frame. The results are transformed into MIDI sequences which can be played with a selection of different instruments from a soundbank. Another software module enables xSonify to convert the sonified data into other sound formats to make it easier to archive and exchange the Sonification results with other scientists.

[ascl:1805.016] xspec_emcee: XSPEC-friendly interface for the emcee package

XSPEC_EMCEE is an XSPEC-friendly interface for emcee (ascl:1303.002). It carries out MCMC analyses of X-ray spectra in the X-ray spectral fitting program XSPEC (ascl:9910.005). It can run multiple xspec processes simultaneously, speeding up the analysis, and can switch to parameterizing norm
parameters in log space.

[ascl:9910.005] XSPEC: An X-ray spectral fitting package

It has been over a decade since the first paper was published containing results determined using the general X-ray spectral-fitting program XSPEC. Since then XSPEC has become the most widely used program for this purpose, being the de facto standard for the ROSAT and the de jure standard for the ASCA and XTE satellites. Probably the most important features of XSPEC are the large number of theoretical models available and the facilities for adding new models.

[ascl:9910.008] XSTAR: A program for calculating conditions and spectra of photoionized gases

XSTAR is a command-driven, interactive, computer program for calculating the physical conditions and emission spectra of photoionized gases. It may be applied in a wide variety of astrophysical contexts. Stripped to essentials, its job may be described simply: A spherical gas shell surrounding a central source of ionizing radiation absorbs some of this radiation and reradiates it in other portions of the spectrum; XSTAR computes the effects on the gas of absorbing this energy, and the spectrum of reradiated light. The user supplies the shape and strength of the incident continuum, the elemental abundances in the gas, its density or pressure, and its thickness; the code can be directed to return any of a large number of derived quantities, including (but not limited to) the ionization balance and temperature, opacity tables, and emitted line and continuum fluxes.

[ascl:2212.011] xwavecal: Wavelength calibrating echelle spectrographs

The xwavecal library automatically wavelength calibrates echelle spectrographs for high precision radial velocity work. The routines are designed to operate on data with extracted 1D spectra. The library provides a convienience function which returns a list of wavelengths from just a list of spectral feature coordinates (pixel and order) and a reference line list. The returned wavelengths are the wavelengths of the measured spectral features under the best fit wavelength model. xwavecal also provides line identification and spectral reduction utilities. The library is modular; each step of the wavelength calibration is a stage which can be disabled by removing the associated line in the config.ini file. Wavelength calibrating data which already have spectra means only using the wavelength calibration stages. Using the full experimental pipeline means enabling the other data reduction stages, such as overscan subtraction.

[ascl:1306.016] Yaxx: Yet another X-ray extractor

Yaxx is a Perl script that facilitates batch data processing using Perl open source software and commonly available software such as CIAO/Sherpa, S-lang, SAS, and FTOOLS. For Chandra and XMM analysis it includes automated spectral extraction, fitting, and report generation. Yaxx can be run without climbing an extensive learning curve; even so, yaxx is highly configurable and can be customized to support complex analysis. yaxx uses template files and takes full advantage of the unique Sherpa / S-lang environment to make much of the processing user configurable. Although originally developed with an emphasis on X-ray data analysis, yaxx evolved to be a general-purpose pipeline scripting package.

[ascl:1908.022] YMW16: Electron-density model

YMW16 models the distribution of free electrons in the Galaxy, the Magellanic Clouds and the inter-galactic medium and can be used to estimate distances for real or simulated pulsars and fast radio bursts (FRBs) based on their position and dispersion measure. The Galactic model is based on 189 pulsars that have independently determined distances as well as dispersion measures, whereas simpler models are used for the electron density in the MC and the IGM.

[ascl:1305.008] YNOGK: Calculating null geodesics in the Kerr spacetime

YNOGK, written in Fortran, calculates the null geodesics in the Kerr spacetime. It uses Weierstrass' and Jacobi's elliptic functions to express all coordinates and affine parameters as analytical and numerical functions of a parameter $p$, which is an integral value along the geodesic. The information about the turning points do not need to be specified in advance by the user, allowing applications such as imaging, the calculation of line profiles or the observer-emitter problem to become root finding problems. Elliptic integrations are computed by Carlson's elliptic integral method, which allows fast computation.

[ascl:1403.012] YNOGKM: Time-like geodesics in the Kerr-Newmann Spacetime calculations

YNOGKM (Yun-Nan observatories geodesic in a Kerr-Newman spacetime for massive particles) performs fast calculation of time-like geodesics in the Kerr-Newman (K-N) spacetime; it is a direct extension of YNOGK (Yun-Nan observatories geodesic Kerr) calculating null geodesics in a Kerr spacetime. The four Boyer-Lindquis coordinates and proper time are expressed as functions of a parameter p semi-analytically by using the Weierstrass' and Jacobi's elliptic functions and integrals. The elliptic integrals are computed by Carlson's elliptic integral method, which guarantees the fast speed of the code. The source Fortran file ynogkm.f90 contains three modules: constants, rootfind, ellfunction, and blcoordinates.

[ascl:1312.009] YODA: Yet another Object Detection Application

YODA, implemented in C++, performs object detection, photometry and star-galaxy classification on astronomical images. Developed specifically to cope with the multi-band imaging data common in modern extragalactic imaging surveys, it is modular and therefore easily adaptable to specific needs. YODA works under conditions of inhomogeneous background noise across the detection frame, and performs accurate aperture photometry in image sets not sharing a common coordinate system or pixel scale as is often the case in present-day extragalactic survey work.

[ascl:2208.025] Yonder: Data denoising and reconstruction

YONDER uses singular value decomposition to perform low-rank data denoising and reconstruction. It takes a tabular data matrix and an error matrix as input and returns a denoised version of the original dataset as output. The approach enables a more accurate data analysis in the presence of uncertainties. Consequently, this package can be used as a simple toolbox to perform astronomical data cleaning.

[ascl:1203.010] Youpi: YOUr processing PIpeline

Youpi is a portable, easy to use web application providing high level functionalities to perform data reduction on scientific FITS images. Built on top of various open source reduction tools released to the community by TERAPIX (http://terapix.iap.fr), Youpi can help organize data, manage processing jobs on a computer cluster in real time (using Condor) and facilitate teamwork by allowing fine-grain sharing of results and data. Youpi is modular and comes with plugins which perform, from within a browser, various processing tasks such as evaluating the quality of incoming images (using the QualityFITS software package), computing astrometric and photometric solutions (using SCAMP), resampling and co-adding FITS images (using SWarp) and extracting sources and building source catalogues from astronomical images (using SExtractor). Youpi is useful for small to medium-sized data reduction projects; it is free and is published under the GNU General Public License.

[submitted] ysoisochrone: A Python package to estimate masses and ages for YSOs

ysoisochrone is a Python3 package that handles the isochrones for young stellar objects (YSOs), and utilize isochrones to derive the stellar mass and ages. Our primary method is a Bayesian inference approach, and the Python code builds on the IDL version developed in Pascucci et al. (2016). The code estimates the stellar masses, ages, and associated uncertainties by comparing their stellar effective temperature, bolometric luminosity, and their uncertainties with different stellar evolutionary models, including those specifically developed for YSOs. User-developed evolutionary tracks can also be utilized when provided in the specific format described in the code documentation.

[ascl:1011.022] yt: A Multi-Code Analysis Toolkit for Astrophysical Simulation Data

yt is an open source, community-developed volumetric analysis and visualization toolkit. Originally designed for handling Enzo's (ascl:1010.072) structure adaptive mesh refinement (AMR) data, yt has been extended to work with numerous simulation methods and simulation codes including Orion, RAMSES (ascl:1011.007), and FLASH (ascl:1010.082). Analysis and visualization with yt are oriented around physically relevant quantities rather than quantities native to data representation on-disk or in-memory. yt can be used for projections, multivariate volume rendering, multi-dimensional histograms, halo finding, light cone generation and topologically-connected isocontour identification.

yt benefits from the contributions of a broad range of community members, and a full list of credits for the code can be found on the yt website or in the source repository.

[ascl:2406.017] ytree: yt-based merger-tree code

ytree reads and works with merger tree data from multiple formats. An extension of yt (ascl:1011.022), which can analyze snapshots from cosmological simulations, ytree can be thought of as the yt of merger trees. ytree's online documentation lists supported formats; support for additional formats can be added, as in principle, any type of tree-like data where an object has one or more ancestors and a single descendant can be supported.

[ascl:1602.003] ZAP: Zurich Atmosphere Purge

ZAP (Zurich Atmosphere Purge) provides sky subtraction for integral field spectroscopy; its approach is based on principal component analysis (PCA) developed for the Multi Unit Spectrographic Explorer (MUSE) integral field spectrograph. ZAP employs filtering and data segmentation to enhance the inherent capabilities of PCA for sky subtraction. ZAP reduces sky emission residuals while robustly preserving the flux and line shapes of astronomical sources; this method works in a variety of observational situations from sparse fields with a low density of sources to filled fields in which the target source fills the field of view. With the inclusion of both of these situations the method is generally applicable to many different science cases and should also be useful for other instrumentation.

[ascl:1607.012] ZASPE: Zonal Atmospheric Stellar Parameters Estimator

ZASPE (Zonal Atmospheric Stellar Parameters Estimator) computes the atmospheric stellar parameters (Teff, log(g), [Fe/H] and vsin(i)) from echelle spectra via least squares minimization with a pre-computed library of synthetic spectra. The minimization is performed only in the most sensitive spectral zones to changes in the atmospheric parameters. The uncertainities and covariances computed by ZASPE assume that the principal source of error is the systematic missmatch between the observed spectrum and the sythetic one that produces the best fit. ZASPE requires a grid of synthetic spectra and can use any pre-computed library minor modifications.

[ascl:1807.017] ZBARYCORR: Barycentric redshift calculator

ZBARYCORR determines the barycentric redshift (zB) for a given star. It calculates the positions and velocities of solar system objects, applies the rotation, precession, nutation, and polar motion of the Earth, applies the stellar motion using the Markwardt library (ascl:1807.016), Shapiro delay, and light-travel term, and finally calculates the quantity zB—the barycentric correction independent of the measured redshift. A Python wrapper, BARYCORR (ascl:1807.018), is available.

[ascl:1907.017] ZChecker: Zwicky Transient Facility moving target checker for short object lists

ZChecker finds, measures, and visualizes known comets in the Zwicky Transient Facility time-domain survey. Images of targets are identified using on-line ephemeris generation and survey metadata. The photometry of the targets are measured and the images are processed with temporal filtering to highlight morphological variations in time.

[ascl:2310.007] zCluster: Measure photometric redshifts for galaxy clusters

zCluster measures galaxy cluster photometric redshifts using data from broadband photometry in large public surveys, given a priori knowledge of the cluster position. The code retrieves and uses redshift probability distributions in order to create a projected two-dimensional density map of a targeted galaxy cluster, which is later convolved with a Gaussian kernel to smooth the map. zCluster also produces photometric redshift estimates and galaxy density maps for any point in the sky using the included zField tool.

[ascl:1404.002] ZDCF: Z-Transformed Discrete Correlation Function

The cross-correlation function (CCF) is commonly employed in the study of AGN, where it is used to probe the structure of the broad line region by line reverberation, to study the continuum emission mechanism by correlating multi-waveband light curves and to seek correlations between the variability and other AGN properties. The z -transformed discrete correlation function (ZDCF) is a method for estimating the CCF of sparse, unevenly sampled light curves. Unlike the commonly used interpolation method, it does not assume that the light curves are smooth and it does provide errors on its estimates.

[ascl:1110.005] ZEBRA: Zurich Extragalactic Bayesian Redshift Analyzer

The current version of the Zurich Extragalactic Bayesian Redshift Analyzer (ZEBRA) combines and extends several of the classical approaches to produce accurate photometric redshifts down to faint magnitudes. In particular, ZEBRA uses the template-fitting approach to produce Maximum Likelihood and Bayesian redshift estimates based on: (1.) An automatic iterative technique to correct the original set of galaxy templates to best represent the SEDs of real galaxies at different redshifts; (2.) A training set of spectroscopic redshifts for a small fraction of the photometric sample; and (3.) An iterative technique for Bayesian redshift estimates, which extracts the full two-dimensional redshift and template probability function for each galaxy.

[ascl:2205.012] Zelda: Generate correlation functions and power spectra from a galaxy catalog

The Zelda command-line tool extracts correlation functions in velocity space from a galaxy catalog. Zelda is modular, extendable, and can be generalized to produce power spectra and to work in position space. Written in C, it was heavily inspired by the cosmological Boltzmann code CLASS (ascl:1106.020). Zelda is a parallel code using the OpenMP standard.

[ascl:1605.016] zeldovich-PLT: Zel'dovich approximation initial conditions generator

zeldovich-PLT generates Zel'dovich approximation (ZA) initial conditions (i.e. first-order Lagrangian perturbation theory) for cosmological N-body simulations, optionally applying particle linear theory (PLT) corrections.

[ascl:1512.016] ZeldovichRecon: Halo correlation function using the Zeldovich approximation

ZeldovichRecon computes the halo correlation function using the Zeldovich approximation. It includes 3 variants: 1.) zelrecon.cpp, which computes the various contributions to the correlation function; 2.) zelrecon_ctypes.cpp, which is designed to be called from Python using the ctypes library; and 3.) a version which implements the "ZEFT" formalism of "A Lagrangian effective field theory" [arxiv:1506.05264] including the alpha term described in that paper.

[ascl:1911.012] Zeltron: Explicit 3D relativistic electromagnetic Particle-In-Cell code

Zeltron is an explicit 3D relativistic electromagnetic Particle-In-Cell code suited for modeling particle acceleration in astrophysical plasmas. The code is efficiently parallelized with the Message Passing Interface, and can be run on a laptop computer or on multiple cores on current supercomputers. Zeltron takes into account the effect of the radiation reaction force on the motion of the particles; it assigns variable weights to the macro-particles to model particle density gradients, and does not strictly conserve the total energy. The code uses linear interpolation to deposit the charges and currents generated by each particle at the nodes of the computational grid, and computes the charge and current densities for Maxwell's equations. Zeltron contains a large set of analysis tools, including plasma density, particle spectrum, optically thin synchrotron and inverse Compton spectra, angular distributions, and stress-energy tensor.

[ascl:1102.027] ZENO: N-body and SPH Simulation Codes

The ZENO software package integrates N-body and SPH simulation codes with a large array of programs to generate initial conditions and analyze numerical simulations. Written in C, the ZENO system is portable between Mac, Linux, and Unix platforms. It is in active use at the Institute for Astronomy (IfA), at NRAO, and possibly elsewhere.

Zeno programs can perform a wide range of simulation and analysis tasks. While many of these programs were first created for specific projects, they embody algorithms of general applicability and embrace a modular design strategy, so existing code is easily applied to new tasks. Major elements of the system include structured data file utilities facilitate basic operations on binary data, including import/export of ZENO data to other systems; snapshot generation routines to create particle distributions with various properties; systems with user-specified density profiles can be realized in collisionless or gaseous form; multiple spherical and disk components may be set up in mutual equilibrium; and snapshot manipulation routines permit the user to sift, sort, and combine particle arrays, translate and rotate particle configurations, and assign new values to data fields associated with each particle.

Simulation codes include both pure N-body and combined N-body/SPH programs. Pure N-body codes are available in both uniprocessor and parallel versions. SPH codes offer a wide range of options for gas physics, including isothermal, adiabatic, and radiating models. Snapshot analysis programs calculate temporal averages, evaluate particle statistics, measure shapes and density profiles, compute kinematic properties, and identify and track objects in particle distributions. Visualization programs generate interactive displays and produce still images and videos of particle distributions; the user may specify arbitrary color schemes and viewing transformations.

[ascl:1306.014] ZEUS-2D: Simulation of fluid dynamical flows

ZEUS-2D is a hydrodynamics code based on ZEUS which adds a covariant differencing formalism and algorithms for compressible hydrodynamics, MHD, and radiation hydrodynamics (using flux-limited diffusion) in Cartesian, cylindrical, or spherical polar coordinates.

[ascl:1102.028] ZEUS-MP/2: Computational Fluid Dynamics Code

ZEUS-MP is a multiphysics, massively parallel, message-passing implementation of the ZEUS code. ZEUS-MP offers an MHD algorithm that is better suited for multidimensional flows than the ZEUS-2D module by virtue of modifications to the method of characteristics scheme first suggested by Hawley & Stone. This MHD module is shown to compare quite favorably to the TVD scheme described by Ryu et al. ZEUS-MP is the first publicly available ZEUS code to allow the advection of multiple chemical (or nuclear) species. Radiation hydrodynamic simulations are enabled via an implicit flux-limited radiation diffusion (FLD) module. The hydrodynamic, MHD, and FLD modules can be used, singly or in concert, in one, two, or three space dimensions. In addition, so-called 1.5D and 2.5D grids, in which the "half-D'' denotes a symmetry axis along which a constant but nonzero value of velocity or magnetic field is evolved, are supported. Self-gravity can be included either through the assumption of a GM/r potential or through a solution of Poisson's equation using one of three linear solver packages (conjugate gradient, multigrid, and FFT) provided for that purpose. Point-mass potentials are also supported.

Because ZEUS-MP is designed for large simulations on parallel computing platforms, considerable attention is paid to the parallel performance characteristics of each module in the code. Strong-scaling tests involving pure hydrodynamics (with and without self-gravity), MHD, and RHD are performed in which large problems (2563 zones) are distributed among as many as 1024 processors of an IBM SP3. Parallel efficiency is a strong function of the amount of communication required between processors in a given algorithm, but all modules are shown to scale well on up to 1024 processors for the chosen fixed problem size.

[ascl:2008.010] zeus: Lightning Fast MCMC

Zeus is a pure-Python implementation of the Ensemble Slice Sampling method. Ensemble Slice Sampling improves upon Slice Sampling by bypassing some of that method's difficulties; it also exploits an ensemble of parallel walkers, thus making it immune to linear correlations. Zeus offers fast and robust Bayesian inference and efficient Markov Chain Monte Carlo without hand-tuning. The code provides excellent performance in terms of autocorrelation time and convergence rate, can scale to multiple CPUs without any extra effort, and includes convergence diagnostics.

[ascl:2306.017] Zeus21: Simulations of 21-cm at cosmic dawn

Zeus21 (Zippy Early-Universe Solver for 21-cm) captures the nonlocal and nonlinear physics of cosmic dawn to create an effective model for the 21-cm power spectrum and global signal. The code takes advantage of the approximate log-normality of the star-formation rate density (SFRD) during cosmic dawn to compute the 21-cm power spectrum analytically. It agrees with more expensive semi-numerical simulations to roughly 10% precision, but has comparably negligible computational cost (~ s) and memory requirements. Zeus21 pairs well with data from HERA, but can be used for any 21-cm inference or prediction. Its capabilities include finding the 21-cm power spectrum (at a broad range of k and z), the global signal, IGM temperatures (Tk, Ts, Tcolor), neutral fraction xHI, Lyman-alpha fluxes, and the evolution of the SFRD; all across cosmic dawn z=5-35. It can also predict UVLFs for HST and JWST. Zeus21 can use three different astrophysical models, one of which emulates 21cmFAST (ascl:1102.023), and can vary the cosmology through CLASS (ascl:1106.020).

[ascl:1511.022] ZInCo: Zoomed Initial Conditions

ZInCo manipulates existing initial conditions (ICs) compatible with GADGET-2/3 (ascl:0003.001) ICs, allowing different flavors of zoom-in simulations rather then producing new ICs from scratch. The code can manipulate initial conditions with multiple types of particles, unlike the vast majority of zoom-in ICs codes available, preserving their properties and random field. This allows ZInCo to take advantage of other codes that produce ICs featuring a broad range of different cosmologies; it can be used also on existing ICs even in the unlikely case nothing is known about their properties. The code is written in C++ and parallelized using MPI.

[ascl:1202.002] ZODIPIC: Zodiacal Cloud Image Synthesis

ZODIPIC synthesizes images of exozodiacal clouds. As a default, ZODIPIC creates an image of the solar zodiacal cloud as seen from 10 pc, but it contains many parameters that are tweakable from the command line, making it a handy general-purpose model for optically-thin debris disks that yields both accurate images and photometric information simultaneously. Written in IDL, ZODIPIC includes dust with real optical constants, user-specified dust maps and can compute images as seen through a linear polarizer.

[ascl:2306.012] ZodiPy: Zodiacal emission simulations in timestreams or HEALPix for solar system observers

ZodiPy simulates the zodiacal emission in intensity that an arbitrary solar system observer is predicted to see given an interplanetary dust model, either in the form of timestreams or full-sky HEALPix maps. Written in Python, the code makes zodiacal emission simulations more accessible by providing a simple interface to existing models.

[ascl:2105.010] ZOGY: Python implementation of proper image subtraction

ZOGY performs optimal image subtraction; the code is designed specifically for the MeerLICHT and BlackGEM pipelines, but should also be useful to apply to images of other telescopes. The module accepts a new and a reference FITS image, runs SExtractor (ascl:1010.064) on them, and finds their WCS solution using Astrometry.net (ascl:1208.001). ZOGY then uses PSFex (ascl:1301.001) to infer the position-dependent PSFs of the images and SWarp (ascl:1010.068) to map the reference image to the new image and performs optimal image subtraction. This produces the subtracted image, the significance image, the corrected significance image, and the PSF photometry image and associated error image. The inferred PSFs are also used to extract optimal photometry of all sources detected by SExtractor.

[ascl:2203.027] Zoobot: Deep learning galaxy morphology classifier

Zoobot classifies galaxy morphology with Bayesian CNN. Deep learning models were trained on volunteer classifications; these models were able to both learn from uncertain volunteer responses and predict full posteriors (rather than point estimates) for what volunteers would have said. The code reproduces and improves Galaxy Zoo DECaLS automated classifications, and can be finetuned for new tasks.

[ascl:1011.003] ZPEG: An Extension of the Galaxy Evolution Model PEGASE.2

Photometric redshifts are estimated on the basis of template scenarios with the help of the code ZPEG, an extension of the galaxy evolution model PEGASE.2 and available on the PEGASE web site. The spectral energy distribution (SED) templates are computed for nine spectral types including starburst, irregular, spiral and elliptical. Dust, extinction and metal effects are coherently taken into account, depending on evolution scenarios. The sensitivity of results to adding near-infrared colors and IGM absorption is analyzed. A comparison with results of other models without evolution measures the evolution factor which systematically increases the estimated photometric redshift values by $Delta z$ > 0.2 for z > 1.5. Moreover we systematically check that the evolution scenarios match observational standard templates of nearby galaxies, implying an age constraint of the stellar population at z=0 for each type. The respect of this constraint makes it possible to significantly improve the accuracy of photometric redshifts by decreasing the well-known degeneracy problem. The method is applied to the HDF-N sample. From fits on SED templates by a $chi^2$-minimization procedure, not only is the photometric redshift derived but also the corresponding spectral type and the formation redshift $z_for$ when stars first formed. Early epochs of galaxy formation z > 5 are found from this new method and results are compared to faint galaxy count interpretations.

[ascl:2106.034] ztf-viewer: SNAD ZTF data releases object viewer

The SNAD ZTF DR4 object viewer enables quick expert investigation of objects within the public Zwicky Transient Facility (ZTF) data releases. The viewer allows visualization of raw and folded light curves and metadata, as well as cross-match information with the General Catalog of Variable Stars, the International Variable Stars Index, the ATLAS Catalog of Variable Stars, the ZTF Catalog of Periodic Variable Stars, the Transient Name Server, the Open Astronomy Catalogs, the OGLE III Catalog of Variable Stars, the Simbad Astronomical Data Base, Gaia DR2 distances (Bailer-Jones+, 2018), and Vizier. The viewer is also available for ZTF DR2 and ZTF DR3.

[ascl:2106.033] ZWAD: Anomaly detection pipeline

ZWAD (ZTF anomaly detection pipeline) examines data and performs tailored feature extraction. The code then uses machine learning methods to searches for outliers, and identifies anomalies to be examined for validation by experts. Used with the SNAD ZTF data releases object viewer (ascl:2106.034), the infrastructure helps experts to form global views of specific scientifically interesting candidates.

[ascl:2202.003] Zwindstroom: Cosmological growth factors from fluid calculations

Zwindstroom computes background quantities and scale-dependent growth factors for cosmological models with free-streaming species, such as massive neutrinos. Following the earlier REPS code (ascl:1612.022), the code uses a Newtonian fluid approximation with external neutrino sound speed to close the Boltzmann hierarchy. Zwindstroom supports multi-fluid models with distinct transfer functions and sound speeds. A flexible python interface facilitates interaction with CLASS (ascl:1106.020) through classy. There is also a Zwindstroom plugin for the cosmological initial conditions generator monofonIC (ascl:2008.024) that allows for higher-order LPT ICs for massive neutrino simulations in a single step.

[ascl:2306.006] β-SGP: Scaled Gradient Projection algorithm using β-divergence

β-SGP deconvolves an astronomical image with a known Point Spread Function, providing a means for restoration of telescopic images due to issues ranging from atmospheric turbulence to instrumental aberrations. The code supports improved astrometry, deblending of overlapping sources, faint source detection, and identification of point sources near bright extended objects, and other tasks. β-SGP generalizes the Scaled Gradient Projection (SGP) image deconvolution algorithm using β-divergence as a loss function to restore distorted stellar shapes.

[ascl:2307.050] νHawkHunter: Forecasting of PBH neutrinos

νHawkHunter explores the prospects of detecting neutrinos produced by the evaporation of primordial black holes in ground-based experiments. It makes use of neutrino fluxes from Hawking radiation computed with BlackHawk (ascl:2012.020). νHawkHunter is also be used for Diffuse Supernova Neutrino Background or similar studies by replacing the signal fluxes by the proper ones.

Would you like to view a random code?