[ascl:1602.013]
TailZ: Redshift distributions estimator of photometric samples of galaxies

TailZ estimates redshift distributions of photometric samples of galaxies selected photometrically given a subsample with measured spectroscopic redshifts. The approach uses a non-parametric Voronoi tessellation density estimator to interpolate the galaxy distribution in the redshift and photometric color space. The Voronoi tessellation estimator performs well at reconstructing the tails of the redshift distribution of individual galaxies and gives unbiased estimates of the first and second moments.

[ascl:1202.004]
TALYS: Nuclear Reaction Simulator

TALYS is software that simulates nuclear reactions which involve neutrons, gamma-rays, protons, deuterons, tritons, helions and alpha-particles, in the 1 keV-200MeV energy range. A suite of nuclear reaction models has been implemented into a single code system, enabling one to evaluate basically all nuclear reactions beyond the resonance range. In particular, TALYS estimates the Maxwellian-averaged reaction rates that are of astrophysical relevance. This enables reaction rates to be calculated with increased accuracy and reliability and the approximations of previous codes to be investigated. The TALYS predictions for the thermonuclear rates of relevance to astrophysics are detailed and compared with those derived by widely-used codes for the same nuclear ingredients. TALYS predictions may differ significantly from those of previous codes, in particular for nuclei for which no or little nuclear data is available. The pre-equilibrium process is shown to influence the astrophysics rates of exotic neutron-rich nuclei significantly. The TALYS code provides a tool to estimate all nuclear reaction rates of relevance to astrophysics with improved accuracy and reliability.

[ascl:1503.003]
TAME: Tool for Automatic Measurement of Equivalent-width

TAME measures the equivalent width (EWs) in high-resolution spectra. Written by IDL, TAME provides the EWs of spectral lines by profile fitting in an automatic or interactive mode and is reliable for measuring EWs in a spectrum with a spectral resolution of R ≳ 20000. It offers an interactive mode for more flexible measurement of the EW and a fully automatic mode that can simultaneously measure the EWs for a large set of lines.

[ascl:1306.007]
Tapir: A web interface for transit/eclipse observability

Tapir is a set of tools, written in Perl, that provides a web interface for showing the observability of periodic astronomical events, such as exoplanet transits or eclipsing binaries. The package provides tools for creating finding charts for each target and airmass plots for each event. The code can access target lists that are stored on-line in a Google spreadsheet or in a local text file.

[ascl:1402.018]
TARDIS: Temperature And Radiative Diffusion In Supernovae

TARDIS creates synthetic spectra for supernova ejecta and is sufficiently fast to allow exploration of the complex parameter spaces of models for SN ejecta. TARDIS uses Monte Carlo methods to obtain a self-consistent description of the plasma state and to compute a synthetic spectrum. It is written in Python with a modular design that facilitates the implementation of a range of physical approximations that can be compared to assess both accuracy and computational expediency; this allows users to choose a level of sophistication appropriate for their application.

[ascl:1305.014]
TAU: 1D radiative transfer code for transmission spectroscopy of extrasolar planet atmospheres

TAU is a 1D line-by-line radiative transfer code for modeling transmission spectra of close-in extrasolar planets. The code calculates the optical path through the planetary atmosphere of the radiation from the host star and quantifies the absorption due to the modeled composition in a transmission spectrum of transit depth as a function of wavelength. The code is written in C++ and is parallelized using OpenMP.

[ascl:1505.031]
TEA: Thermal Equilibrium Abundances

TEA (Thermal Equilibrium Abundances) calculates gaseous molecular abundances under thermochemical equilibrium conditions. Given a single T,P point or a list of T,P pairs (the thermal profile of an atmosphere) and elemental abundances, TEA calculates mole fractions of the desired molecular species. TEA uses 84 elemental species and thermodynamical data for more then 600 gaseous molecular species, and can adopt any initial elemental abundances.

[ascl:1405.002]
TelFit: Fitting the telluric absorption spectrum

TelFit calculates the best-fit telluric absorption spectrum in high-resolution optical and near-IR spectra. The best-fit model can then be divided out to remove the telluric contamination. Written in Python, TelFit is essentially a wrapper to LBLRTM, the Line-By-Line Radiative Transfer Model, and simplifies the process of generating a telluric model.

[ascl:1509.002]
Tempo: Pulsar timing data analysis

Manchester, R.; Taylor, J.; Peters, W.; Weisberg, J.; Irwin, A.; Wex, N.; Stairs, I.; Demorest, P.; Nice, D.

Tempo analyzes pulsar timing data. Pulse times of arrival (TOAs), pulsar model parameters, and coded instructions are read from one or more input files. The TOAs are fitted by a pulse timing model incorporating transformation to the solar-system barycenter, pulsar rotation and spin-down and, where necessary, one of several binary models. Program output includes parameter values and uncertainties, residual pulse arrival times, chi-squared statistics, and the covariance matrix of the model. In prediction mode, ephemerides of pulse phase behavior (in the form of polynomial expansions) are calculated from input timing models. Tempo is the basis for the Tempo2 (ascl:1210.015) code.

[ascl:1210.015]
Tempo2: Pulsar Timing Package

Tempo2 is a pulsar timing package developed to be used both for general pulsar timing applications and also for pulsar timing array research in which data-sets from multiple pulsars need to be processed simultaneously. It was initially developed by George Hobbs and Russell Edwards as part of the Parkes Pulsar Timing Array project. Tempo2 is based on the original Tempo (ascl:1509.002) code and can be used (from the command-line) in a similar fashion. It is very versatile and can be extended by plugins.

[ascl:1611.002]
tf_unet: Generic convolutional neural network U-Net implementation in Tensorflow

tf_unet mitigates radio frequency interference (RFI) signals in radio data using a special type of Convolutional Neural Network, the U-Net, that enables the classification of clean signal and RFI signatures in 2D time-ordered data acquired from a radio telescope. The code is not tied to a specific segmentation and can be used, for example, to detect radio frequency interference (RFI) in radio astronomy or galaxies and stars in widefield imaging data. This U-Net implementation can outperform classical RFI mitigation algorithms.

[ascl:1505.019]
TFIT: Mixed-resolution data set photometry package

Laidler, Victoria G.; Papovich, Casey; Grogin, Norman A.; Idzi, Rafal; Dickinson, Mark; Ferguson, Henry C.; Hilbert, Bryan; Clubb, Kelsey; Ravindranath, Swara

TFIT measures galaxy photometry using prior knowledge of sources in a deep, high‐resolution image (HRI) to improve photometric measurements of objects in a corresponding low‐resolution image (LRI) of the same field, usually at a different wavelength. For background‐limited data, this technique produces optimally weighted photometry that maximizes signal‐to‐noise ratio (S/N). For objects not significantly detected in the low‐resolution image, it provides useful and quantitative information for setting upper limits.

This code is no longer updated and has been superseded by T-PHOT (ascl:1609.001).

[ascl:1303.012]
TGCat: Chandra Transmission Grating Catalog and Archive

TGCat is an archive of Chandra transmission grating spectra and a suite of software for processing such data. Users can browse and categorize Chandra gratings observations quickly and easily, generate custom plots of resulting response corrected spectra on-line without the need for special software and download analysis ready products from multiple observations in one convenient operation. Data processing for the catalog is done with a suite of ISIS/S-Lang scripts; the software is available for download. These ISIS scripts wrap and call CIAO tools for reprocessing from "Level 1" (acis_process_events or hrc_process_events) through "Level 2" (binned spectra, via tg_resolve_events and tgextract), compute responses (grating "RMFs" and "ARFs", via mkgrmf and mkgarf), and make summary plots.

[ascl:1602.010]
The Cannon: Data-driven method for determining stellar parameters and abundances from stellar spectra

The Cannon is a data-driven method for determining stellar labels (physical parameters and chemical abundances) from stellar spectra in the context of vast spectroscopic surveys. It fits for the spectral model given training spectra and labels, with the polynomial order for the spectral model decided by the user, infers labels for the test spectra, and provides diagnostic output for monitoring and evaluating the process. It offers SNR-independent continuum normalization, performs well at lower signal-to-noise, and is very accurate.

[ascl:1105.003]
The DTFE public software: The Delaunay Tessellation Field Estimator code

We present the DTFE public software, a code for reconstructing fields from a discrete set of samples/measurements using the maximum of information contained in the point distribution. The code is written in C++ using the CGAL library and is parallelized using OpenMP. The software was designed for the analysis of cosmological data but can be used in other fields where one must interpolate quantities given at a discrete point set. The software comes with a wide suite of options to facilitate the analysis of 2- and 3-dimensional data and of both numerical simulations and galaxy redshift surveys. For comparison purposes, the code also implements the TSC and SPH grid interpolation methods. The code comes with an extensive user guide detailing the program options, examples and the inner workings of the code.

[ascl:1405.003]
The Hammer: An IDL Spectral Typing Suite

The Hammer can classify spectra in a variety of formats with targets spanning the MK spectral sequence. It processes a list of input spectra by automatically estimating each object's spectral type and measuring activity and metallicity tracers in late type stars. Once automatic processing is complete, an interactive interface allows the user to manually tweak the final assigned spectral type through visual comparison with a set of templates.

[ascl:1701.001]
The Joker: A custom Monte Carlo sampler for binary-star and exoplanet radial velocity data

Given sparse or low-quality radial-velocity measurements of a star, there are often many qualitatively different stellar or exoplanet companion orbit models that are consistent with the data. The consequent multimodality of the likelihood function leads to extremely challenging search, optimization, and MCMC posterior sampling over the orbital parameters. The Joker is a custom-built Monte Carlo sampler that can produce a posterior sampling for orbital parameters given sparse or noisy radial-velocity measurements, even when the likelihood function is poorly behaved. The method produces correct samplings in orbital parameters for data that include as few as three epochs. The Joker can therefore be used to produce proper samplings of multimodal pdfs, which are still highly informative and can be used in hierarchical (population) modeling.

[ascl:1407.001]
The Starfish Diagram: Statistical visualization tool

The Starfish Diagram is a statistical visualization tool that simultaneously displays the properties of an individual and its parent sample through a series of histograms. The code is useful for large datasets for which one needs to understand the standing or significance of a single entry.

[ascl:1604.008]
The Tractor: Probabilistic astronomical source detection and measurement

The Tractor optimizes or samples from models of astronomical objects. The approach is generative: given astronomical sources and a description of the image properties, the code produces pixel-space estimates or predictions of what will be observed in the images. This estimate can be used to produce a likelihood for the observed data given the model: assuming the model space actually includes the truth (it doesn’t, in detail), then if we had the optimal model parameters, the predicted image would differ from the actually observed image only by noise. Given a noise model of the instrument and assuming pixelwise independent noise, the log-likelihood is the negative chi-squared difference: (image - model) / noise.

[ascl:1706.008]
the-wizz: Clustering redshift estimation code

the-wizz clusters redshift estimates for any photometric unknown sample in a survey. The software is composed of two main parts: a pair finder and a pdf maker. The pair finder finds spatial pairs and stores the indices of all closer pairs around target reference objects in an output HDF5 data file. Users then query this data file using the indices of their unknown sample to produce an output clustering-z.

[ascl:1308.013]
THELI GUI: Optical, near- & mid-infrared imaging data reduction

THELI is an easy-to-use, end-to-end pipeline for the reduction of any optical, near-IR and mid-IR imaging data. It combines a variety of processing algorithms and third party software into a single, homogeneous tool. Over 90 optical and infrared instruments at observatories world-wide are pre-configured; more can be added by the user. The code's online appendix contains three walk-through examples using public data (optical, near-IR and mid-IR) and additional online documentation is available for training and troubleshooting.

[ascl:1112.003]
THERMINATOR 2: THERMal heavy IoN generATOR 2

THERMINATOR is a Monte Carlo event generator dedicated to studies of the statistical production of particles in relativistic heavy-ion collisions. The increased functionality of the code contains the following features: The input of any shape of the freeze-out hypersurface and the expansion velocity field, including the 3+1 dimensional profiles, in particular those generated externally with various hydrodynamic codes. The hypersufraces may have variable thermal parameters, which allows for studies departing significantly from the mid-rapidity region, where the baryon chemical potential becomes large. We include a library of standard sets of hypersurfaces and velocity profiles describing the RHIC Au+Au data at sqrt(s_(NN)) = 200 GeV for various centralities, as well as those anticipated for the LHC Pb+Pb collisions at sqrt(s_(NN)) = 5.5 TeV. A separate code, FEMTO-THERMINATOR, is provided to carry out the analysis of femtoscopic correlations which are an important source of information concerning the size and expansion of the system. We also include several useful scripts that carry out auxiliary tasks, such as obtaining an estimate of the number of elastic collisions after the freeze-out, counting of particles flowing back into the fireball and violating causality (typically very few), or visualizing various results: the particle p_T-spectra, the elliptic flow coefficients, and the HBT correlation radii. We also investigate the problem of the back-flow of particles into the hydrodynamic region, as well as estimate the elastic rescattering in terms of trajectory crossings. The package is written in C++ and uses the CERN ROOT environment.

[ascl:1212.014]
Thrust: Productivity-Oriented Library for CUDA

Thrust is a parallel algorithms library which resembles the C++ Standard Template Library (STL). Thrust's high-level interface greatly enhances programmer productivity while enabling performance portability between GPUs and multicore CPUs. Interoperability with established technologies (such as CUDA, TBB, and OpenMP) facilitates integration with existing software.

[ascl:1609.021]
TIDEV: Tidal Evolution package

TIDEV (Tidal Evolution package) calculates the evolution of rotation for tidally interacting bodies using Efroimsky-Makarov-Williams (EMW) formalism. The package integrates tidal evolution equations and computes the rotational and dynamical evolution of a planet under tidal and triaxial torques. TIDEV accounts for the perturbative effects due to the presence of the other planets in the system, especially the secular variations of the eccentricity. Bulk parameters include the mass and radius of the planet (and those of the other planets involved in the integration), the size and mass of the host star, the Maxwell time and Andrade's parameter. TIDEV also calculates the time scale that a planet takes to be tidally locked as well as the periods of rotation reached at the end of the spin-orbit evolution.

[ascl:1206.012]
Time Utilities

Time Utilities are software tools that, in principal, allow one to calculate BJD to a precision of 1 μs for any target from anywhere on Earth or from any spacecraft. As the quality and quantity of astrophysical data continue to improve, the precision with which certain astrophysical events can be timed becomes limited not by the data themselves, but by the manner, standard, and uniformity with which time itself is referenced. While some areas of astronomy (most notably pulsar studies) have required absolute time stamps with precisions of considerably better than 1 minute for many decades, recently new areas have crossed into this regime. In particular, in the exoplanet community, we have found that the (typically unspecified) time standards adopted by various groups can differ by as much as a minute. Left uncorrected, this ambiguity may be mistaken for transit timing variations and bias eccentricity measurements. We recommend using BJD_TDB, the Barycentric Julian Date in the Barycentric Dynamical Time standard for any astrophysical event. The BJD_TDB is the most practical absolute time stamp for extraterrestrial phenomena, and is ultimately limited by the properties of the target system. We compile a general summary of factors that must be considered in order to achieve timing precisions ranging from 15 minutes to 1 μs, and provide software for download and online webapps for use.

[ascl:1010.057]
Tiny Tim: Simulated Hubble Space Telescope PSFs

Tiny Tim generates simulated Hubble Space Telescope point spread functions (PSFs). It is written in C and distributed as source code and runs on a wide variety of UNIX and VMS systems. Tiny Tim includes mirror zonal errors, time dependent aberrations (for the pre-repair instruments), field dependent obscuration patterns (for WF/PC-1 and WFPC2), and filter passband effects. It can produce a normally sampled or subsampled PSF. Output is a FITS image file.

[ascl:1111.015]
TIPSY: Code for Display and Analysis of N-body Simulations

The development of TIPSY was motivated by the need to quickly display and analyze the results of N-body simulations. Most data visualization packages are designed for the display of gridded data, and hence are unsuitable for use with particle data. Therefore, a special package was built that could easily perform the following functions:

- Display particle positions (as points), and velocities (as line segments) from an arbitrary viewpoint.
- Zoom in to a chosen position. Due to their extremely clustered nature, structure of interest in an N-body simulation is often so small that it can not be seen when looking at the simulation as a whole.
- Color particles to display scalar fields. Examples of such fields are potential energy, or for SPH particles, density and temperature.
- Selection of a subset of the particles for display and analysis. Regions of interest are generally small subsets of the simulation.
- Following selected particles from one timestep to another.
- Finding cumulative properties of a collection of particles. This usually involves just a sum over the particles.

[ascl:1208.008]
TiRiFiC: Tilted Ring Fitting Code

Tilted Ring Fitting Code (TiRiFiC) is a prototype computer program to construct simulated (high-resolution) astronomical spectroscopic 3d-observations (data cubes) of simple kinematical and morphological models of rotating (galactic) disks. It is possible to automatically optimize the parameterizations of constructed model disks to fit spectroscopic (3d-) observations via a χ2 minimization. TiRiFiC is currently implemented as an add-on to the Groningen Image Processing System (GIPSY) software package and attempts to provide a method to automatically fit an extended tilted-ring model directly to a data cube.

[ascl:1108.012]
TITAN: General-purpose Radiation Hydrodynamics Code

TITAN is a general-purpose radiation hydrodynamics code developed at the Laboratory for Computational Astrophysics (NCSA, University of Illinois at Urbana-Champaign). TITAN solves the coupled sets of radiation transfer and fluid dynamics equations on an adaptive mesh in one spatial dimension.

[ascl:1109.021]
TLUSTY: Stellar Atmospheres, Accretion Disks, and Spectroscopic Diagnostics

TLUSTY is a user-oriented package written in FORTRAN77 for modeling stellar atmospheres and accretion disks and wide range of spectroscopic diagnostics. In the program's maximum configuration, the user may start from scratch and calculate a model atmosphere of a chosen degree of complexity, and end with a synthetic spectrum in a wavelength region of interest for an arbitrary stellar rotation and an arbitrary instrumental profile. The user may also model the vertical structure of annuli of an accretion disk.

[ascl:1512.014]
TM: Torus Mapper

TM (Torus Mapper) produces models for orbits in action-angle coordinates in axisymmetric potentials using torus mapping, a non-perturbative technique for creating orbital tori for specified values of the action integrals. It can compute a star's position at any time given an orbital torus and a star’s position at a reference time, and also provides a way to choose initial conditions for N-body simulations of realistic disc

galaxies that start in perfect equilibrium. TM provides some advantages over use of a standard time-stepper to

create orbits.

[ascl:1212.015]
TMAP: Tübingen NLTE Model-Atmosphere Package

The Tübingen NLTE Model-Atmosphere Package (TMAP) is a tool to calculate stellar atmospheres in spherical or plane-parallel geometry in hydrostatic and radiative equilibrium allowing departures from local thermodynamic equilibrium (LTE) for the population of atomic levels. It is based on the Accelerated Lambda Iteration (ALI) method and is able to account for line blanketing by metals. All elements from hydrogen to nickel may be included in the calculation with model atoms which are tailored for the aims of the user.

[ascl:1605.005]
TMBIDL: Single dish radio astronomy data reduction package

The IDL package reduces and analyzes radio astronomy data. It translates SDFITS files into TMBIDL format, and can average and display spectra, remove baselines, and fit Gaussian models.

[ascl:1209.007]
TMCalc: Fast estimation of stellar metallicity [Fe/H]

TMCalc is a C code developed as an extension to ARES. Using the line list given, the code can be used as a precise and fast indicator of the spectroscopic temperature and metallicity for dwarf FKG stars with effective temperatures ranging from 4500 K to 6500 K and with [Fe/H] ranging from -0.8 dex to 0.4 dex.

[ascl:1104.001]
TomograPy: A Fast, Instrument-Independent, Solar Tomography Software

TomograPy is an open-source software freely available on the Python Package Index that can perform fast tomographic inversions that scale linearly with the number of measurements, linearly with the length of the reconstruction cube (and not the number of voxels) and linearly with the number of cores and can use data from different sources and with a variety of physical models. For performance, TomograPy uses a parallelized-projection algorithm. It relies on the World Coordinate System standard to manage various data sources. A variety of inversion algorithms are provided to perform the tomographic-map estimation. A test suite is provided along with the code to ensure software quality. Since it makes use of the Siddon algorithm it is restricted to rectangular parallelepiped voxels but the spherical geometry of the corona can be handled through proper use of priors.

[ascl:1101.010]
TOPCAT: Tool for OPerations on Catalogues And Tables

TOPCAT is an interactive graphical viewer and editor for tabular data. Its aim is to provide most of the facilities that astronomers need for analysis and manipulation of source catalogues and other tables, though it can be used for non-astronomical data as well. It understands a number of different astronomically important formats (including FITS and VOTable) and more formats can be added.

It offers a variety of ways to view and analyse tables, including a browser for the cell data themselves, viewers for information about table and column metadata, and facilities for 1-, 2-, 3- and higher-dimensional visualisation, calculating statistics and joining tables using flexible matching algorithms. Using a powerful and extensible Java-based expression language new columns can be defined and row subsets selected for separate analysis. Table data and metadata can be edited and the resulting modified table can be written out in a wide range of output formats.

It is a stand-alone application which works quite happily with no network connection. However, because it uses Virtual Observatory (VO) standards, it can cooperate smoothly with other tools in the VO world and beyond, such as VODesktop, Aladin and ds9. Between 2006 and 2009 TOPCAT was developed within the AstroGrid project, and is offered as part of a standard suite of applications on the AstroGrid web site, where you can find information on several other VO tools.

The program is written in pure Java and available under the GNU General Public Licence. It has been developed in the UK within the Starlink and AstroGrid projects, and under PPARC and STFC grants. Its underlying table processing facilities are provided by STIL.

[ascl:1404.006]
TORUS: Radiation transport and hydrodynamics code

TORUS is a flexible radiation transfer and radiation-hydrodynamics code. The code has a basic infrastructure that includes the AMR mesh scheme that is used by several physics modules including atomic line transfer in a moving medium, molecular line transfer, photoionization, radiation hydrodynamics and radiative equilibrium. TORUS is useful for a variety of problems, including magnetospheric accretion onto T Tauri stars, spiral nebulae around Wolf-Rayet stars, discs around Herbig AeBe stars, structured winds of O supergiants and Raman-scattered line formation in symbiotic binaries, and dust emission and molecular line formation in star forming clusters. The code is written in Fortran 2003 and is compiled using a standard Gnu makefile. The code is parallelized using both MPI and OMP, and can use these parallel sections either separately or in a hybrid mode.

[ascl:1507.006]
Toyz: Large datasets and astronomical images analysis framework

Toyz is a python web framework that allows scientists to interact with large images and data sets stored on a remote server. A web application is run on the server containing the data and clients are run from web browsers on the user's computer. Toyz displays large FITS images also also renders any image format supported by Pillow (a fork of the Python Imaging Library), contains a GUI to interact with linked plots, and offers a customizable framework that allows students and researchers to create their own work spaces inside a Toyz environment. Astro-Toyz extends the features of the Toyz image viewer, allowing users to view world coordinates and align images based on their WCS.

[ascl:1603.012]
tpipe: Searching radio interferometry data for fast, dispersed transients

Visibilities from radio interferometers have not traditionally been used to study the fast transient sky. Millisecond transients (e.g., fast radio bursts) and periodic sources (e.g., pulsars) have been studied with single-dish radio telescopes and a software stack developed over the past few decades. tpipe is an initial attempt to develop the fast transient algorithms for visibility data. Functions exist for analysis of visibilties, such as reading data, flagging data, applying interferometric gain calibration, and imaging. These functions are given equal footing as time-domain techniques like filters and dedispersion.

[ascl:1305.003]
TPM: Tree-Particle-Mesh code

TPM carries out collisionless (dark matter) cosmological N-body simulations, evolving a system of N particles as they move under their mutual gravitational interaction. It combines aspects of both Tree and Particle-Mesh algorithms. After the global PM forces are calculated, spatially distinct regions above a given density contrast are located; the tree code calculates the gravitational interactions inside these denser objects at higher spatial and temporal resolution. The code is parallel and uses MPI for message passing.

[ascl:1304.011]
TPZ: Trees for Photo-Z

TPZ, a parallel code written in python, produces robust and accurate photometric redshift PDFs by using prediction tree and random forests. The code also produces ancillary information about the sample used, such as prior unbiased errors estimations (giving an estimation of performance) and a ranking of importance of variables as well as a map of performance indicating where extra training data is needed to improve overall performance. It is designed to be easy to use and a tutorial is available.

[ascl:1601.001]
TRADES: TRAnsits and Dynamics of Exoplanetary Systems

TRADES (TRAnsits and Dynamics of Exoplanetary Systems) simultaneously fits observed radial velocities and transit times data to determine the orbital parameters of exoplanetary systems from observational data. It uses a dynamical simulator for N-body systems that also fits the available data during the orbital integration and determines the best combination of the orbital parameters using grid search, χ2 minimization, genetic algorithms, particle swarm optimization, and bootstrap analysis.

[ascl:1501.011]
transfer: The Sloan Digital Sky Survey Data Transfer Infrastructure

The Sloan Digital Sky Survey (SDSS) produces large amounts of data daily. transfer, written in Python, provides the effective automation needed for daily data transfer operations and management and operates essentially free of human intervention. This package has been tested and used successfully for several years.

[ascl:1106.014]
Transit Analysis Package (TAP and autoKep): IDL Graphical User Interfaces for Extrasolar Planet Transit Photometry

We present an IDL graphical user interface-driven software package designed for the analysis of extrasolar planet transit light curves. The Transit Analysis Package (TAP) software uses Markov Chain Monte Carlo (MCMC) techniques to fit light curves using the analytic model of Mandel and Agol (2002). The package incorporates a wavelet based likelihood function developed by Carter and Winn (2009) which allows the MCMC to assess parameter uncertainties more robustly than classic chi-squared methods by parameterizing uncorrelated "white" and correlated "red" noise. The software is able to simultaneously analyze multiple transits observed in different conditions (instrument, filter, weather, etc). The graphical interface allows for the simple execution and interpretation of Bayesian MCMC analysis tailored to a user's specific data set and has been thoroughly tested on ground-based and Kepler photometry. AutoKep provides a similar GUI for the preparation of Kepler MAST archive data for analysis by TAP or any other analysis software. This paper describes the software release and provides instructions for its use.

[ascl:1611.008]
Transit Clairvoyance: Predicting multiple-planet systems for TESS

Transit Clairvoyance uses Artificial Neural Networks (ANNs) to predict the most likely short period transiters to have additional transiters, which may double the discovery yield of the TESS (Transiting Exoplanet Survey Satellite). Clairvoyance is a simple 2-D interpolant that takes in the number of planets in a system with period less than 13.7 days, as well as the maximum radius amongst them (in Earth radii) and orbital period of the planet with maximum radius (in Earth days) in order to predict the probability of additional transiters in this system with period greater than 13.7 days.

[ascl:1011.018]
Transit of a Spherical Planet of a Stellar Chromosphere which is Geometrically Thin

Transit light curves for stellar continua have only one minimum and a "U" shape. By contrast, transit curves for optically thin chromospheric emission lines can have a "W" shape because of stellar limb-brightening. We calculate light curves for an optically thin shell of emission and fit these models to time-resolved observations of Si IV absorption by the planet HD209458b. We find that the best fit Si IV absorption model has R_p,SIV/R_*= 0.34+0.07-0.12, similar to the Roche lobe of the planet. While the large radius is only at the limit of statistical significance, we develop formulae applicable to transits of all optically thin chromospheric emission lines.

[ascl:1704.008]
Transit: Radiative-transfer code for planetary atmospheres

Cubillos, Patricio; Blecic, Jasmina; Harrington, Joe; Rojo, Patricio; Foster, Austin J.; Stemm, Madison; Challener,Ryan; Foster, Andrew S. D.

Transit calculates the transmission or emission spectrum of a planetary atmosphere with application to extrasolar-planet transit and eclipse observations, respectively. It computes the spectra by solving the one-dimensional line-by-line radiative-transfer equation for an atmospheric model.

[ascl:1703.010]
TransitSOM: Self-Organizing Map for Kepler and K2 transits

A self-organizing map (SOM) can be used to identify planetary candidates from Kepler and K2 datasets with accuracies near 90% in distinguishing known Kepler planets from false positives. TransitSOM classifies a Kepler or K2 lightcurve using a self-organizing map (SOM) created and pre-trained using PyMVPA (ascl:1703.009). It includes functions for users to create their own SOMs.

[ascl:1412.011]
TraP: Transients discovery pipeline for image-plane surveys

The TraP is a pipeline for detecting and responding to transient and variable sources in a stream of astronomical images. Images are initially processed using a pure-Python source-extraction package, PySE, which is bundled with the TraP. Source positions and fluxes are then loaded into a SQL database for association and variability detection. The database structure allows for estimation of past upper limits on newly detected sources, and for forced fitting of previously detected sources which have since dropped below the blind-extraction threshold. Developed with LOFAR data in mind, the TraP has been used with data from other radio observatories.

[ascl:1508.007]
TreeCorr: Two-point correlation functions

TreeCorr efficiently computes two-point correlation functions. It can compute correlations of regular number counts, weak lensing shears, or scalar quantities such as convergence or CMB temperature fluctuations. Two-point correlations may be auto-correlations or cross-correlations, including any combination of shear, kappa, and counts. Two-point functions can be done with correct curved-sky calculation using RA, Dec coordinates, on a Euclidean tangent plane, or in 3D using RA, Dec and a distance. The front end is written in Python, which can be used as a Python module or as a standalone executable using configuration files; the actual computation of the correlation functions is done in C++ using ball trees (similar to kd trees), making the calculation extremely efficient, and when available, OpenMP is used to run in parallel on multi-core machines.

[ascl:1612.019]
Trident: Synthetic spectrum generator

Trident creates synthetic absorption-line spectra from astrophysical hydrodynamics simulations. It uses the yt package (ascl:1011.022) to read in simulation datasets and extends it to provide realistic synthetic observations appropriate for studies of the interstellar, circumgalactic, and intergalactic media.

[ascl:1508.009]
Trilogy: FITS image conversion software

Trilogy automatically scales and combines FITS images to produce color or grayscale images using Python scripts. The user assigns images to each color channel (RGB) or a single image to grayscale luminosity. Trilogy determines the intensity scaling automatically and independently in each channel to display faint features without saturating bright features. Each channel's scaling is determined based on a sample of the image (or summed images) and two input parameters. One parameter sets the output luminosity of "the noise," currently determined as 1-sigma above the sigma-clipped mean. The other parameter sets what fraction of the data (if any) in the sample region should be allowed to saturate. Default values for these parameters (0.15% and 0.001%, respectively) work well, but the user is able to adjust them. The scaling is accomplished using the logarithmic function y = a log(kx + 1) clipped between 0 and 1, where a and k are constants determined based on the data and desired scaling parameters as described above.

[ascl:1210.014]
TRIP: General computer algebra system for celestial mechanics

TRIP is an interactive computer algebra system that is devoted to perturbation series computations, and specially adapted to celestial mechanics. Its development started in 1988, as an upgrade of the special purpose FORTRAN routines elaborated by J. Laskar for the demonstration of the chaotic behavior of the Solar System. TRIP is a mature and efficient tool for handling multivariate generalized power series, and embeds two kernels, a symbolic and a numerical kernel. This numerical kernel communicates with Gnuplot or Grace to plot the graphics and allows one to plot the numerical evaluation of symbolic objects.

[ascl:1405.008]
TRIPP: Time Resolved Imaging Photometry Package

Geckeler, Ralf D.; Schuh, Sonja; Dreizler, Stefan; Deetjen, Jochen; Gleissner, Thomas; Risse, Patrick; Rauch, Thomas; Göhler, Eckart; Hügelmeyer, Simon; Husser, Tim-Oliver; Israel, Holger; Benlloch-Garcia, Sara; Pottschmidt, Katja; Wilms, Jörn

Written in IDL, TRIPP performs CCD time series reduction and analysis. It provides an on-line check of the incoming frames, performs relative aperture photometry and provides a set of time series tools, such as calculation of periodograms including false alarm probability determination, epoc folding, sinus fitting, and light curve simulations.

[ascl:1605.010]
TRIPPy: Python-based Trailed Source Photometry

Fraser, Wesley C.; Alexandersen, Mike; Schwamb, Megan E.; Marsset, Michael E.; Pike, Rosemary E.; Kavelaars, JJ; Bannister, Michele T.; Benecchi, Susan; Delsanti, Audrey

TRIPPy (TRailed Image Photometry in Python) uses a pill-shaped aperture, a rectangle described by three parameters (trail length, angle, and radius) to improve photometry of moving sources over that done with circular apertures. It can generate accurate model and trailed point-spread functions from stationary background sources in sidereally tracked images. Appropriate aperture correction provides accurate, unbiased flux measurement. TRIPPy requires numpy, scipy, matplotlib, Astropy (ascl:1304.002), and stsci.numdisplay; emcee (ascl:1303.002) and SExtractor (ascl:1010.064) are optional.

[ascl:1509.005]
TRUVOT: True Background Technique for the Swift UVOT Grisms

TRUVOT decontaminates Swift UVOT grism spectra for transient objects. The technique makes use of template images in a process similar to image subtraction.

[ascl:1406.011]
TSP: Time-Series/Polarimetry Package

TSP is an astronomical data reduction package that handles time series data and polarimetric data from a variety of different instruments, and is distributed as part of the Starlink software collection (ascl:1110.012).

[ascl:1409.002]
Tsyganenko Geomagnetic Field Models

The Tsyganenko models are semi-empirical best-fit representations for the magnetic field, based on a large number of satellite observations (IMP, HEOS, ISEE, POLAR, Geotail, GOES, etc). The models include the contributions from major external magnetospheric sources: ring current, magnetotail current system, magnetopause currents, and large-scale system of field-aligned currents.

[ascl:1404.015]
TTVFast: Transit timing inversion

TTVFast efficiently calculates transit times for n-planet systems and the corresponding radial velocities. The code uses a symplectic integrator with a Keplerian interpolator for the calculation of transit times (Nesvorny et al. 2013); it is available in both C and Fortran.

[ascl:1604.012]
TTVFaster: First order eccentricity transit timing variations (TTVs)

TTVFaster implements analytic formulae for transit time variations (TTVs) that are accurate to first order in the planet–star mass ratios and in the orbital eccentricities; the implementations are available in several languages, including IDL, Julia, Python and C. These formulae compare well with more computationally expensive N-body integrations in the low-eccentricity, low mass-ratio regime when applied to simulated and to actual multi-transiting Kepler planet systems.

[ascl:1011.011]
turboGL: Accurate Modeling of Weak Lensing

turboGL is a fast Mathematica code based on a stochastic approach to cumulative weak lensing. It can easily compute the lensing PDF relative to arbitrary halo mass distributions, selection biases, number of observations, halo profiles and evolutions, making it a useful tool to study how lensing depends on cosmological parameters and impact on observations.

[ascl:1205.004]
Turbospectrum: Code for spectral synthesis

Turbospectrum is a 1D LTE spectrum synthesis code which covers 600 molecules, is fast with many lines, and uses the treatment of line broadening described by Barklem & O’Mara (1998).

[ascl:1304.015]
TVD: Total Variation Diminishing code

TVD solves the magnetohydrodynamic (MHD) equations by updating the fluid variables along each direction using the flux-conservative, second-order, total variation diminishing (TVD), upwind scheme of Jin & Xin. The magnetic field is updated separately in two-dimensional advection-constraint steps. The electromotive force (EMF) is computed in the advection step using the TVD scheme, and this same EMF is used immediately in the constraint step in order to preserve ∇˙B=0 without the need to store intermediate fluxes. The code is extended to three dimensions using operator splitting, and Runge-Kutta is used to get second-order accuracy in time. TVD offers high-resolution per grid cell, second-order accuracy in space and time, and enforcement of the ∇˙B=0 constraint to machine precision. Written in Fortran, It has no memory overhead and is fast. It is also available in a fully scalable message-passing parallel MPI implementation.

[ascl:1708.015]
TWO-POP-PY: Two-population dust evolution model

TWO-POP-PY runs a two-population dust evolution model that follows the upper end of the dust size distribution and the evolution of the dust surface density profile and treats dust surface density, maximum particle size, small and large grain velocity, and fragmentation. It derives profiles that describe the dust-to-gas ratios and the dust surface density profiles well in protoplanetary disks, in addition to the radial flux by solid material rain out.

[ascl:1407.002]
TWODSPEC: Long-slit and optical fiber array spectra extensions for FIGARO

TWODSPEC offers programs for the reduction and analysis of long-slit and optical fiber array spectra, implemented as extensions to the FIGARO package (ascl:1203.013). The software are currently distributed as part of the Starlink software collection (ascl:1110.012). These programs are designed to do as much as possible for the user, to assist quick reduction and analysis of data; for example, LONGSLIT can fit multiple Gaussians to line profiles in batch and decides how many components to fit.

[ascl:1210.025]
TwoDSSM: Self-gravitating 2D shearing sheet

TwoDSSM solves the equations of self-gravitating hydrodynamics in the shearing sheet, with cooling. TwoDSSM is configured to use a simple, exponential cooling model, although it contains code for a more complicated (and perhaps more realistic) cooling model based on a one-zone vertical model. The complicated cooling model can be switched on using a flag.

[ascl:1303.008]
TYCHO: Stellar evolution code

TYCHO is a general, one dimensional (spherically symmetric) stellar evolution code written in structured Fortran 77; it is designed for hydrostatic and hydrodynamic stages including mass loss, accretion, pulsations and explosions. Mixing and convection algorithms are based on 3D time-dependent simulations. It offers extensive on-line graphics using Tim Pearson's PGPLOT with X-windows and runs effectively on Linux and Mac OS X laptop and desktop computers.

NOTE: This code is no longer being supported.

[ascl:1303.006]
UCL_CHEM: time and depth dependent gas-grain chemical model

UCL_CHEM is a time and depth dependent gas-grain chemical model that can be used to estimate the fractional abundances (with respect to hydrogen) of gas and surface species in every environment where molecules are present. The model includes both gas and surface reactions. The code starts from the most diffuse state where all the gas is in atomic form and evolve sthe gas to its final density. Depending on the temperature, atoms and molecules from the gas freeze on to the grains and they hydrogenate where possible. The advantage of this approach is that the ice composition is not assumed but it is derived by a time-dependent computation of the chemical evolution of the gas-dust interaction process. The code is very modular, has been used to model a variety of regions and can be coupled with the UCL_PDR and SMMOL codes.

[ascl:1303.004]
UCL_PDR: Time dependent photon-dissociation regions model

UCL_PDR is a time dependent photon-dissociation regions model that calculates self consistently the thermal balance. It can be used with gas phase only species as well as with surface species. It is very modular, has the possibility of accounting for density and pressure gradients and can be coupled with UCL_CHEM as well as with SMMOL. It has been used to model small scale (e.g. knots in proto-planetary nebulae) to large scale regions (high redshift galaxies).

[ascl:1704.002]
UDAT: A multi-purpose data analysis tool

UDAT is a pattern recognition tool for mass analysis of various types of data, including image and audio. Based on its WND-CHARM (ascl:1312.002) prototype, UDAT computed a large set of numerical content descriptors from each file it analyzes, and selects the most informative features using statistical analysis. The tool can perform automatic classification of galaxy images by training with annotated galaxy images. It also has unsupervised learning capabilities, such as query-by-example of galaxies based on morphology. That is, given an input galaxy image of interest, the tool can search through a large database of images to retrieve the galaxies that are the most similar to the query image. The downside of the tool is its computational complexity, which in most cases will require a small or medium cluster.

[ascl:1611.001]
UltraNest: Pythonic Nested Sampling Development Framework and UltraNest

This three-component package provides a Pythonic implementation of the Nested Sampling integration algorithm for Bayesian model comparison and parameter estimation. It offers multiple implementations for constrained drawing functions and a test suite to evaluate the correctness, accuracy and efficiency of various implementations. The three components are:

- a modular framework for nested sampling algorithms (nested_sampling) and their development;
- a test framework to evaluate the performance and accuracy of algorithms (testsuite); and
- UltraNest, a fast C implementation of a mixed RadFriends/MCMC nested sampling algorithm.

[ascl:1104.007]
ULySS: A Full Spectrum Fitting Package

ULySS (University of Lyon Spectroscopic Analysis Software) is an open-source software package written in the GDL/IDL language to analyze astronomical data. ULySS fits a spectrum with a linear combination of non-linear components convolved with a line-of-sight velocity distribution (LOSVD) and multiplied by a polynomial continuum. ULySS is used to study stellar populations of galaxies and star clusters and atmospheric parameters of stars.

[ascl:1403.007]
Unified EOS for neutron stars

The equation of state (EOS) of dense matter is a crucial input for the neutron-star structure calculations. This Fortran code can obtain a "unified EOS" in the many-body calculations based on a single effective nuclear Hamiltonian, and is valid in all regions of the neutron star interior. For unified EOSs, the transitions between the outer crust and the inner crust and between the inner crust and the core are obtained as a result of many-body calculations.

[ascl:1503.007]
UniPOPS: Unified data reduction suite

Maddalena, Ronald J.; Garwood, Robert W.; Salter, Christopher J.; Stobie, Elizabeth B; Cram, Thomas R.; Morgan, Lorrie; Vance, Bob; Hudson, Jerome

UniPOPS, a suite of programs and utilities developed at the National Radio Astronomy Observatory (NRAO), reduced data from the observatory's single-dish telescopes: the Tucson 12-m, the Green Bank 140-ft, and archived data from the Green Bank 300-ft. The primary reduction programs, 'line' (for spectral-line reduction) and 'condar' (for continuum reduction), used the People-Oriented Parsing Service (POPS) as the command line interpreter. UniPOPS unified previous analysis packages and provided new capabilities; development of UniPOPS continued within the NRAO until 2004 when the 12-m was turned over to the Arizona Radio Observatory (ARO). The submitted code is version 3.5 from 2004, the last supported by the NRAO.

[ascl:1110.021]
Univiewer: Visualisation Program for HEALPix Maps

Univiewer is a visualisation program for HEALPix maps. It is written in C++ and uses OpenGL and the wxWidgets library for cross-platform portability. Using it you can:

- Rotate and zoom maps on the sphere in 3D;
- Create high-resolution views of square patches of the map;
- Change maximum and minimum values of the colourmap interactively;
- Calculate the power spectrum of the full-sky map or a patch;
- Display any column of a HEALPix map FITS file on the sphere.

In the 3D view, a HEALPix map is projected onto a ECP pixelation to create a texture which is wrapped around the sphere. In calculating the power spectrum, the spherical harmonic transforms are computed using the same ECP pixelation. This inevitably leads to some discrepancies at small scales due to repixelation effects, but they are reasonably small.

[ascl:1504.001]
UPMASK: Unsupervised Photometric Membership Assignment in Stellar Clusters

UPMASK, written in R, performs membership assignment in stellar clusters. It uses photometry and spatial positions, but can take into account other types of data. UPMASK takes into account arbitrary error models; the code is unsupervised, data-driven, physical-model-free and relies on as few assumptions as possible. The approach followed for membership assessment is based on an iterative process, principal component analysis, a clustering algorithm and a kernel density estimation.

[ascl:1512.019]
UPSILoN: AUtomated Classification of Periodic Variable Stars using MachIne LearNing

UPSILoN (AUtomated Classification of Periodic Variable Stars using MachIne LearNing) classifies periodic variable stars such as Delta Scuti stars, RR Lyraes, Cepheids, Type II Cepheids, eclipsing binaries, and long-period variables (i.e. superclasses), and their subclasses (e.g. RR Lyrae ab, c, d, and e types) using well-sampled light curves from any astronomical time-series surveys in optical bands regardless of their survey-specific characteristics such as color, magnitude, and sampling rate. UPSILoN consists of two parts, one which extracts variability features from a light curve, and another which classifies a light curve, and returns extracted features, a predicted class, and a class probability. In principle, UPSILoN can classify any light curves having arbitrary number of data points, but using light curves with more than ~80 data points provides the best classification quality.

[ascl:1412.009]
URCHIN: Reverse ray tracer

URCHIN is a Smoothed Particle Hydrodynamics (SPH) reverse ray tracer (i.e. from particles to sources). It calculates the amount of shielding from a uniform background that each particle experiences. Preservation of the adaptive density field resolution present in many gas dynamics codes and uniform sampling of gas resolution elements with rays are two of the benefits of URCHIN; it also offers preservation of Galilean invariance, high spectral resolution, and preservation of the standard uniform UV background in optically thin gas.

[ascl:1411.012]
util_2comp: Planck-based two-component dust model utilities

The util_2comp software utilities generate predictions of far-infrared Galactic dust emission and reddening based on a two-component dust emission model fit to Planck HFI, DIRBE and IRAS data from 100 GHz to 3000 GHz. These predictions and the associated dust temperature map have angular resolution of 6.1 arcminutes and are available over the entire sky. Implementations in IDL and Python are included.

[ascl:1412.003]
UTM: Universal Transit Modeller

The Universal Transit Modeller (UTM) is a light-curve simulator for all kinds of transiting or eclipsing configurations between arbitrary numbers of several types of objects, which may be stars, planets, planetary moons, and planetary rings. A separate fitting program, UFIT (Universal Fitter) is part of the UTM distribution and may be used to derive best fits to light-curves for any set of continuously variable parameters. UTM/UFIT is written in IDL code and its source is released in the public domain under the GNU General Public License.

[ascl:1606.006]
uvmcmcfit: Parametric models to interferometric data fitter

Uvmcmcfit fits parametric models to interferometric data. It is ideally suited to extract the maximum amount of information from marginally resolved observations with interferometers like the Atacama Large Millimeter Array (ALMA), Submillimeter Array (SMA), and Plateau de Bure Interferometer (PdBI). uvmcmcfit uses emcee (ascl:1303.002) to do Markov Chain Monte Carlo (MCMC) and can measure the goodness of fit from visibilities rather than deconvolved images, an advantage when there is strong gravitational lensing and in other situations. uvmcmcfit includes a pure-Python adaptation of Miriad’s (ascl:1106.007) uvmodel task to generate simulated visibilities given observed visibilities and a model image and a simple ray-tracing routine that allows it to account for both strongly lensed systems (where multiple images of the lensed galaxy are detected) and weakly lensed systems (where only a single image of the lensed galaxy is detected).

[ascl:1402.017]
UVMULTIFIT: Fitting astronomical radio interferometric data

UVMULTIFIT, written in Python, is a versatile library for fitting models directly to visibility data. These models can depend on frequency and fitting parameters in an arbitrary algebraic way. The results from the fit to the visibilities of sources with sizes smaller than the diffraction limit of the interferometer are superior to the output obtained from a mere analysis of the deconvolved images. Though UVMULTIFIT is based on the CASA package, it can be easily adapted to other analysis packages that have a Python API.

[ascl:1410.004]
UVOTPY: Swift UVOT grism data reduction

The two Swift UVOT grisms provide uv (170.0-500.0 nm) and visible (285.0-660.0 nm) spectra with a resolution of R~100 and 75. To reduce the grism data, UVOTPY extracts a spectrum given source sky position, and outputs a flux calibrated spectrum. UVOTPY is a replacement for the UVOTIMGRISM FTOOL (ascl:9912.002) in the HEADAS Swift package. Its extraction uses a curved aperture for the uv spectra, accounts the coincidence losses in the detector, provides more accurate anchor positions for the wavelength scale, and is valid for the whole detector.

[ascl:1207.003]
VAC: Versatile Advection Code

The Versatile Advection Code (VAC) is a freely available general hydrodynamic and magnetohydrodynamic simulation software that works in 1, 2 or 3 dimensions on Cartesian and logically Cartesian grids. VAC runs on any Unix/Linux system with a Fortran 90 (or 77) compiler and Perl interpreter. VAC can run on parallel machines using either the Message Passing Interface (MPI) library or a High Performance Fortran (HPF) compiler.

[ascl:1406.009]
VADER: Viscous Accretion Disk Evolution Resource

VADER is a flexible, general code that simulates the time evolution of thin axisymmetric accretion disks in time-steady potentials. VADER handles arbitrary viscosities, equations of state, boundary conditions, and source and sink terms for both mass and energy.

[ascl:1702.004]
Validation: Codes to compare simulation data to various observations

Validation provides codes to compare several observations to simulated data with stellar mass and star formation rate, simulated data stellar mass function with observed stellar mass function from PRIMUS or SDSS-GALEX in several redshift bins from 0.01-1.0, and simulated data B band luminosity function with observed stellar mass function, and to create plots for various attributes, including stellar mass functions, and stellar mass to halo mass. These codes can model predictions (in some cases alongside observational data) to test other mock catalogs.

[ascl:1309.002]
VAPHOT: Precision differential aperture photometry package

VAPHOT is an aperture photometry package for precise time−series photometry of uncrowded fields, geared towards the extraction of target lightcurves of eclipsing or transiting systems. Its photometric main routine works within the IRAF (ascl:9911.002) environment and is built upon the standard aperture photometry task 'phot' from IRAF, using optimized aperture sizes. The associated analysis program 'VANALIZ' works in the IDL environment. It performs differential photometry with graphical and numerical output. VANALIZ produces plots indicative of photometric stability and permits the interactive evaluation and weighting of comparison stars. Also possible is the automatic or manual suppression of data-points and the output of statistical analyses. Several methods for the calculation of the reference brightness are offered. Specific routines for the analysis of transit 'on'-'off' photometry, comparing the target brightness inside against outside a transit are also available.

[ascl:1506.010]
VAPID: Voigt Absorption-Profile [Interstellar] Dabbler

VAPID (Voigt Absorption Profile [Interstellar] Dabbler) models interstellar absorption lines. It predicts profiles and optimizes model parameters by least-squares fitting to observed spectra. VAPID allows cloud parameters to be optimized with respect to several different data set simultaneously; those data sets may include observations of different transitions of a given species, and may have different S/N ratios and resolutions.

[ascl:1111.012]
VAPOR: Visualization and Analysis Platform for Ocean, Atmosphere, and Solar Researchers

VAPOR is the Visualization and Analysis Platform for Ocean, Atmosphere, and Solar Researchers. VAPOR provides an interactive 3D visualization environment that runs on most UNIX and Windows systems equipped with modern 3D graphics cards. VAPOR provides:

- A visual data discovery environment tailored towards the specialized needs of the astro and geosciences CFD community
- A desktop solution capable of handling terascale size data sets
- Advanced interactive 3D visualization tightly coupled with quantitative data analysis
- Support for multi-variate, time-varying data
- Close coupling with RSI's powerful interpretive data language, IDL
- Support for 3D visualization of WRF-ARW datasets

[ascl:1208.016]
VARTOOLS: Light Curve Analysis Program

The VARTOOLS program is a command line utility that provides tools for analyzing time series astronomical data. It implements a number of routines for calculating variability/periodicity statistics of light curves, as well as tools for modifying and filtering light curves.

[ascl:1704.005]
VaST: Variability Search Toolkit

VaST (Variability Search Toolkit) finds variable objects on a series of astronomical images in FITS format. The software performs object detection and aperture photometry using SExtractor (ascl:1010.064) on each image, cross-matches lists of detected stars, performs magnitude calibration with respect to the first (reference) image and constructs a lightcurve for each object. The sigma-magnitude, Stetson's L variability index, Robust Median Statistic (RoMS) and other plots may be used to visually identify variable star candidates. The two distinguishing features of VaST are its ability to perform accurate aperture photometry of images obtained with non-linear detectors and to handle complex image distortions. VaST can be used in cases of unstable PSF (e.g., bad guiding or with digitized wide-field photographic images), and has been successfully applied to images obtained with telescopes ranging from 0.08 to 2.5m in diameter equipped with a variety of detectors including CCD, CMOS, MIC and photographic plates.

[ascl:1610.009]
velbin: radial velocity corrected for binary orbital motions

Velbin convolves the radial velocity offsets due to binary orbital motions with a Gaussian to model an observed velocity distribution. This can be used to measure the mean velocity and velocity dispersion from an observed radial velocity distribution, corrected for binary orbital motions. Velbin fits single- or multi-epoch data with any arbitrary binary orbital parameter distribution (as long as it can be sampled properly), however it always assumes that the intrinsic velocity distribution (i.e. corrected for binary orbital motions) is a Gaussian. Velbin samples (and edits) a binary orbital parameter distribution, fits an observed radial velocity distribution, and creates a mock radial velocity distribution that can be used to provide the fitted radial velocities in the single_epoch or multi_epoch methods.

[ascl:1010.021]
velfit: A Code for Modeling Non-Circular Flows in Disk Galaxies

High-quality velocity maps of galaxies frequently exhibit signatures of non-circular streaming motions. velfit yields results that are more easily interpreted than the commonly used procedure. It can estimate the magnitudes of forced non-circular motions over a broad range of bar strengths from a strongly barred galaxy, through cases of mild bar-like distortions to placing bounds on the shapes of halos in galaxies having extended rotation curves.

[submitted]
venice

venice is a mask utility program that reads a mask file (DS9 or fits type) and a catalogue of objects (ascii or fits type) to:

1. create a pixelized mask,

2. find objects inside/outside a mask,

3. or generate a random catalogue of objects inside/outside a mask.

The program reads the mask file and checks if a point, giving its coordinates, is inside or outside the mask, i.e. inside or outside at least one polygon of the mask.

[ascl:1503.011]
VESPA: False positive probabilities calculator

Validation of Exoplanet Signals using a Probabilistic Algorithm (VESPA) calculates false positive probabilities and statistically validates transiting exoplanets. Written in Python, it uses isochrones [ascl:1503.010] and the package simpledist.

[ascl:1204.007]
VH-1: Multidimensional ideal compressible hydrodynamics code

VH-1 is a multidimensional ideal compressible hydrodynamics code written in FORTRAN for use on any computing platform, from desktop workstations to supercomputers. It uses a Lagrangian remap version of the Piecewise Parabolic Method developed by Paul Woodward and Phil Colella in their 1984 paper. VH-1 comes in a variety of versions, from a simple one-dimensional serial variant to a multi-dimensional version scalable to thousands of processors.

[ascl:1306.015]
VHD: Viscous pseudo-Newtonian accretion

VHD is a numerical study of viscous fluid accretion onto a black hole. The flow is axisymmetric and uses a pseudo-Newtonian potential to model relativistic effects near the event horizon. VHD is based on ZEUS-2D (Stone & Norman 1992) with the addition of an explicit scheme for the viscosity.

[ascl:1404.010]
VictoriaReginaModels: Stellar evolutionary tracks

The Victoria–Regina stellar models are comprised of seventy-two grids of stellar evolutionary tracks accompanied by complementary zero-age horizontal branches and are presented in “equivalent evolutionary phase” (.eep) files. This Fortran 77 software interpolates isochrones, isochrone population functions, luminosity functions, and color functions of stellar evolutionary tracks.

[ascl:1407.014]
VIDE: The Void IDentification and Examination toolkit

Sutter, P. M.; Lavaux, Guilhem; Hamaus, Nico; Pisani, Alice; Wandelt, Benjamin D.; Warren, Michael S.; Villaescusa-Navarro, Francisco; Zivick, Paul; Mao, Qingqing; Thompson, Benjamin B.

The Void IDentification and Examination toolkit (VIDE) identifies voids using a modified version of the parameter-free void finder ZOBOV (ascl:1304.005); a Voronoi tessellation of the tracer particles is used to estimate the density field followed by a watershed algorithm to group Voronoi cells into zones and subsequently voids. Output is a summary of void properties in plain ASCII; a Python API is provided for analysis tasks, including loading and manipulating void catalogs and particle members, filtering, plotting, computing clustering statistics, stacking, comparing catalogs, and fitting density profiles.

[ascl:1403.016]
Viewpoints: Fast interactive linked plotting of large multivariate data sets

Viewpoints is an interactive tool for exploratory visual analysis of large high-dimensional (multivariate) data. It uses linked scatterplots to find relations in a few seconds that can take much longer with other plotting tools. Its features include linked scatter plots with brushing, dynamic histograms, normalization, and outlier detection/removal.

Would you like to view a random code?