Results 3101-3200 of 3550 (3461 ASCL, 89 submitted)

[ascl:2106.022]
STaRS: Sejong Radiative Transfer through Raman and Rayleigh Scattering with atomic hydrogen

The 3D grid-based Monte Carlo code STaRS (Sejong Radiative Transfer through Raman and Rayleigh Scattering with atomic hydrogen) traces radiative transfer through Raman and Rayleigh scattering. This can be used to investigate line formation of Raman-scattered features in a thick neutral region illuminated by a strong far-UV emission source. Favorable conditions for Raman scattering with atomic hydrogen are easily met in symbiotic stars, young planetary nebulae, and active galactic nuclei.

[ascl:1703.005]
starsense_algorithms: Performance evaluation of various star sensors

Sarpotdar, Mayuresh; Mathew, Joice; Sreejith, A. G.; Nirmal, K.; Ambily, S.; Prakash, Ajin; Safonova, Margarita; Murthy, Jayant

The Matlab starsense_algorithms package evaluates the performance of various star sensors through the implementation of centroiding, geometric voting and QUEST algorithms. The physical parameters of a star sensor are parametrized and by changing these parameters, performance estimators such as sky coverage, memory requirement, and timing requirements can be estimated for the selected star sensor.

[ascl:1805.010]
StarSmasher: Smoothed Particle Hydrodynamics code for smashing stars and planets

Smoothed Particle Hydrodynamics (SPH) is a Lagrangian particle method that approximates a continuous fluid as discrete nodes, each carrying various parameters such as mass, position, velocity, pressure, and temperature. In an SPH simulation the resolution scales with the particle density; StarSmasher is able to handle both equal-mass and equal number-density particle models. StarSmasher solves for hydro forces by calculating the pressure for each particle as a function of the particle's properties - density, internal energy, and internal properties (e.g. temperature and mean molecular weight). The code implements variational equations of motion and libraries to calculate the gravitational forces between particles using direct summation on NVIDIA graphics cards. Using a direct summation instead of a tree-based algorithm for gravity increases the accuracy of the gravity calculations at the cost of speed. The code uses a cubic spline for the smoothing kernel and an artificial viscosity prescription coupled with a Balsara Switch to prevent unphysical interparticle penetration. The code also implements an artificial relaxation force to the equations of motion to add a drag term to the calculated accelerations during relaxation integrations. Initially called StarCrash, StarSmasher was developed originally by Rasio.

[ascl:1704.004]
STATCONT: Statistical continuum level determination method for line-rich sources

STATCONT determines the continuum emission level in line-rich spectral data by inspecting the intensity distribution of a given spectrum by using different statistical approaches. The sigma-clipping algorithm provides the most accurate continuum level determination, together with information on the uncertainty in its determination; this uncertainty is used to correct the final continuum emission level. In general, STATCONT obtains accuracies of < 10 % in the continuum determination, and < 5 % in most cases. The main products of the software are the continuum emission level, together with its uncertainty, and data cubes containing only spectral line emission, i.e. continuum-subtracted data cubes. STATCONT also includes the option to estimate the spectral index or variation of the continuum emission with frequency.

[ascl:2201.010]
statmorph: Non-parametric morphological diagnostics of galaxy images

statmorph calculates non-parametric morphological diagnostics of galaxy images (e.g., Gini-M_{20} and CAS statistics), and fits 2D Sérsic profiles. Given a background-subtracted image and a corresponding segmentation map indicating the source(s) of interest, statmorph calculates the following morphological statistics for each source:

- Gini-M20 statistics;

- Concentration, Asymmetry and Smoothness (CAS) statistics;

- Multimode, Intensity and Deviation (MID) statistics;

- outer asymmetry and shape asymmetry;

- Sérsic index; and,

- several shape and size measurements associated to the above statistics, such as ellipticity, Petrosian radius, and half-light radius, among others.

[ascl:1206.006]
statpl: Goodness-of-fit for power-law distributed data

statpl estimates the parameter of power-law distributed data and calculates goodness-of-fit tests for them. Many objects studied in astronomy follow a power-law distribution function (DF), for example the masses of stars or star clusters. Such data is often analyzed by generating a histogram and fitting a straight line to it. The parameters obtained in this way can be severely biased, and the properties of the underlying DF, such as its shape or a possible upper limit, are difficult to extract. statpl is an (effectively) bias-free estimator for the exponent and the upper limit.

[ascl:2112.006]
STDPipe: Simple Transient Detection Pipeline

STDPipe is a set of Python routines for astrometry, photometry and transient detection related tasks, intended for quick and easy implementation of custom pipelines, as well as for interactive data analysis. It is implemented as a library of routines covering most common tasks and operates on standard Python objects, including NumPy arrays for images and Astropy (ascl:1304.002) tables for catalogs and object lists. The pipeline does not re-implement code already implemented in other Python packages; instead, it transparently wraps external codes, such as SExtractor (ascl:1010.064), SCAMP (ascl:1010.063), PSFEx (ascl:1301.001), HOTPANTS (ascl:1504.004), and Astrometry.Net (ascl:1208.001), that do not have their own Python interfaces. STDPipe operates on temporary files, keeping nothing after the run unless something is explicitly requested.

[ascl:1108.018]
STECKMAP: STEllar Content and Kinematics via Maximum A Posteriori likelihood

STECKMAP stands for STEllar Content and Kinematics via Maximum A Posteriori likelihood. It is a tool for interpreting galaxy spectra in terms of their stellar populations through the derivation of their star formation history, age-metallicity relation, kinematics and extinction. The observed spectrum is projected onto a temporal sequence of models of single stellar populations, so as to determine a linear combination of these models that best fits the observed spectrum. The weights of the various components of this linear combination indicate the stellar content of the population. This procedure is regularized using various penalizing functions. The principles of the method are detailed in Ocvirk et al. 2006.

[ascl:1108.013]
STELLA: Multi-group Radiation Hydrodynamics Code

STELLA is a one-dimensional multi-group radiation hydrodynamics code. STELLA incorporates implicit hydrodynamics coupled to a multi-group non-equilibrium radiative transfer for modeling SN II-L light curves. The non-equilibrium description of radiation is crucial for this problem since the presupernova envelope may be of low mass and very dilute. STELLA implicitly treats time dependent equations of the angular moments of intensity averaged over a frequency bin. Local thermodynamic equilibrium is assumed to determine the ionization levels of materials.

[ascl:2010.007]
stella: Stellar flares identifier

stella creates and trains a neural network to identify stellar flares. Within stella, users can simulate flares as a training set, run a neural network, and feed in their own data to the neural network model. The software returns a probability at each data point as to whether that data point is part of a flare; the code can also characterize the flares identified.

[ascl:1505.009]
StellaR: Stellar evolution tracks and isochrones tools

stellaR accesses and manipulates publicly available stellar evolutionary tracks and isochrones from the Pisa low-mass database. It retrieves and plots the required calculations from CDS, constructs by interpolation tracks or isochrones of compositions different to the ones available in the database, constructs isochrones for age not included in the database, and extracts relevant evolutionary points from tracks or isochrones.

[ascl:1303.028]
Stellarics: Inverse Compton scattering from stellar heliospheres

Cosmic ray electrons scatter on the photon fields around stars, including the sun, to create gamma rays by the inverse Compton effect. Stellarics computes the spectrum and angular distribution of this emission. The software also includes general-purpose routines for inverse Compton scattering on a given electron spectrum, for example for interstellar or astrophysical source modelling.

[ascl:1901.012]
stellarWakes: Dark matter subhalo searches using stellar kinematic data

stellarWakes uses stellar kinematic data to search for dark matter (DM) subhalos through their gravitational perturbations to the stellar phase-space distribution.

[ascl:2108.014]
StelNet: Stellar mass and age predictor

StelNet predicts mass and age from absolute luminosity and effective temperature for stars with close to solar metallicity. It uses a Deep Neural Network trained on stellar evolutionary tracks. The underlying model makes no assumption on the evolutionary stage and includes the pre-main sequence phase. A mix of models are trained and bootstrapped to quantify the uncertainty of the model, and data is through all trained models to provide a predictive distribution from which an expectation value and uncertainty level can be estimated.

[ascl:1907.018]
StePar: Inferring stellar atmospheric parameters using the EW method

StePar computes the stellar atmospheric parameters T_{eff}, log g, [Fe/H], and ξ of FGK-type stars using the Equivalent Width (*EW*) method. The code implements a grid of MARCS model atmospheres and uses the MOOG radiative transfer code (ascl:1202.009) and TAME (ascl:1503.003). StePar uses a Downhill Simplex minimization algorithm, running it twice for any given star, to compute the stellar atmospheric parameters.

[ascl:2111.016]
SteParSyn: Stellar atmospheric parameters using the spectral synthesis method

SteParSyn infers stellar atmospheric parameters (Teff, log g, [Fe/H], and Vbroad) of FGKM-type stars using the spectral synthesis method. The code uses the MCMC sampler emcee (ascl:1303.002) in conjunction with an spectral emulator that can interpolate spectra down to a precision < 1%. A grid of synthetic spectra that allow the user to characterize the spectra of FGKM-type stars with parameters in the range of 3500 to 7000 K in Teff, 0.0 to 5.5 dex in log g, and −2.0 to 1.0 dex in [Fe/H] is also provided.

[ascl:1809.014]
stepped_luneburg: Stacked-based ray tracing code to model a stepped Luneburg lens

stepped_luneburg investigates the scattered light properties of a Luneburg lens approximated as a series of concentric shells with discrete refractive indices. The optical Luneburg lens has promising applications for low-cost, continuous all-sky monitoring to obtain transit light curves of bright, nearby stars. This code implements a stack-based algorithm that tracks all reflected and refracted rays generated at each optical interface of the lens as described by Snell's law. The Luneburg lens model parameters, such as number of lens layers, the power-law that describes the refractive indices, the number of incident rays, and the initial direction of the incident wavefront can be altered to optimize lens performance. The stepped_luneburg module can be imported within the Python environment or used with scripting, and it is accompanied by two other modules, enc_int and int_map, that help the user to determine the resolving power of the lens and the strength of scattered light haloes for the purpose of quality assessment.

[ascl:1805.006]
StePS: Stereographically Projected Cosmological Simulations

StePS (Stereographically Projected Cosmological Simulations) compactifies the infinite spatial extent of the Universe into a finite sphere with isotropic boundary conditions to simulate the evolution of the large-scale structure. This eliminates the need for periodic boundary conditions, which are a numerical convenience unsupported by observation and which modifies the law of force on large scales in an unrealistic fashion. StePS uses stereographic projection for space compactification and naive O(N2) force calculation; this arrives at a correlation function of the same quality more quickly than standard (tree or P3M) algorithms with similar spatial and mass resolution. The N2 force calculation is easy to adapt to modern graphics cards, hence StePS can function as a high-speed prediction tool for modern large-scale surveys.

[ascl:2305.019]
sterile-dm: Sterile neutrino production

The sterile neutrino production code sterile-dm incorporates new elements to the calculations of the neutrino opacity at temperatures 10 MeV ≤ T ≤ 10 GeV and folds the asymmetry redistribution and opacity calculations into the sterile neutrino production computation, providing updated PSDs for the range of parameters relevant to the X-ray excess. The code requires several data files, which are included. With each run, sterile-dm creates a new output sub-directory that contains a parameter file listing the mass, mixing angle, initial lepton asymmetry and other information, a state file, which includes, among other states, the temperature and FRW coordinate time, and a set of snapshot files, one for each line in the state file.

[ascl:1306.009]
STF: Structure Finder

STF is a general structure finder designed to find halos, subhaloes, and tidal debris in N-body simulations. The current version is designed to read in "particle data" (that is SPH N-body data), but a simple modification of the I/O can have it read grid data from Grid based codes.

This code has been updated and renamed to VELOCIraptor-STF (ascl:1911.020).

[submitted]
stginga: Ginga for STScI

stginga customizes Ginga to aid data analysis for the data supported by STScI (e.g., HST or JWST). For instance, it provides plugins and configuration files that understand HST and JWST data products.

[ascl:1810.014]
STiC: Stockholm inversion code

STiC is a MPI-parallel non-LTE inversion code for observed full-Stokes observations. The code processes lines from multiple atoms in non-LTE, including partial redistribution effects of scattered photons in angle and frequency of scattered photons (PRD), and can be used with model atmospheres that have a complex depth stratification without introducing artifacts.

[ascl:1110.006]
STIFF: Converting Scientific FITS Images to TIFF

STIFF converts scientific FITS images to the more popular TIFF format for illustration purposes. Most FITS readers and converters do not do a proper job at converting FITS image data to 8 bits. 8-bit images stored in JPEG, PNG or TIFF files have the intensities implicitly stored in a non-linear way. Most current FITS image viewers and converters provide the user an incorrect translation of the FITS image content by simply rescaling linearly input pixel values. A first consequence is that the people working on astronomical images usually have to apply narrow intensity cuts or square-root or logarithmic intensity transformations to actually see something on their deep-sky images. A less obvious consequence is that colors obtained by combining images processed this way are not consistent across such a large range of surface brightnesses. Though with other software the user is generally afforded a choice of nonlinear transformations to apply in order to make the faint stuff stand out more clearly in the images, with the limited selection of choices provides, colors will not be accurately rendered, and some manual tweaking will be necessary. The purpose of STIFF is to produce beautiful pictures in an automatic and consistent way.

[ascl:1105.001]
STILTS: Starlink Tables Infrastructure Library Tool Set

The STIL Tool Set is a set of command-line tools based on STIL, the Starlink Tables Infrastructure Library. It deals with the processing of tabular data; the package has been designed for, but is not restricted to, astronomical tables such as object catalogues. Some of the tools are generic and can work with multiple formats (including FITS, VOTable, CSV, SQL and ASCII), and others are specific to the VOTable format. In some ways, STILTS forms the command-line counterpart of the GUI table analysis tool TOPCAT. The package is robust, fully documented, and designed for efficiency, especially with very large datasets.

Facilities offered include:

- format conversion

- crossmatching

- plotting

- column calculation and rearrangement

- row selections

- data and metadata manipulation and display

- sorting

- statistical calculations

- histogram calculation

- data validation

- VO service access

A powerful and extensible expression language is used for specifying data calculations. These facilities can be put together in very flexible and efficient ways. For tasks in which the data can be streamed, the size of table STILTS can process is effectively unlimited. For other tasks, million-row tables usually do not present a problem. STILTS is written in pure Java (J2SE1.5 or later), and can be run from the command line or from Jython, or embedded into java applications. It is released under the GPL.

[ascl:2305.007]
Stimela: Containerized radio interferometry scripting framework

stimela provides a system-agnostic scripting framework for simulating, processing, and imaging radio interferometric data. The framework executes radio interferometry related tasks such as imaging, calibration, and data synthesis in Docker containers using Python modules. stimela offers a simple interface to packages that perform these tasks rather than doing any data processing, synthesis or analysis itself. stimela only requires Docker and Python. Moreover, because of Docker, a stimela script runs the same way (in the same isolated environment) regardless of the host machine’s settings, thus providing a user-friendly and modular scripting environment that gives general users easy access to novel radio interferometry calibration, imaging, and synthesis packages.

[ascl:1608.001]
Stingray: Spectral-timing software

Stingray is a spectral-timing software package for astrophysical X-ray (and more) data. The package merges existing efforts for a (spectral-)timing package in Python and is composed of a library of time series methods (including power spectra, cross spectra, covariance spectra, and lags); scripts to load FITS data files from different missions; a simulator of light curves and event lists that includes different kinds of variability and more complicated phenomena based on the impulse response of given physical events (e.g. reverberation); and a GUI to ease the learning curve for new users.

[ascl:1204.009]
STOKES: Modeling Radiative Transfer and Polarization

STOKES was designed to perform three-dimensional radiative transfer simulations for astronomical applications. The code also considers the polarization properties of the radiation. The program is based on the Monte-Carlo method and treats optical and ultraviolet polarization induced by scattering off free electrons or dust grains. Emission and scattering regions can be arranged in various geometries within the model space, the computed continuum and line spectra can be evaluated at different inclinations and azimuthal viewing angles.

[ascl:1708.005]
STools: IDL Tools for Spectroscopic Analysis

STools contains a variety of simple tools for spectroscopy, such as reading an IRAF-formatted (multispec) echelle spectrum in FITS, measuring the wavelength of the center of a line, Gaussian convolution, deriving synthetic photometry from an input spectrum, and extracting and interpolating a MARCS model atmosphere (standard composition).

[ascl:2101.018]
stratsi: Stratified streaming instability

Stratsi calculates stratified and vertically-shearing streaming instabilities. It solves one- and two-fluid linearized equations, and, for two-fluid models, also provides the parameters and analytic vertical structure and solves for equilibrium horizontal velocity profiles. It offers utilities and various plotting options, including plots to compare one- and two-fluid results, viscous results to inviscid results, and results from two different stokes numbers or two different metallicities. stratsi requires Dedalus (ascl:1603.015) and Eigentools (ascl:2101.017).

[ascl:1702.009]
stream-stream: Stellar and dark-matter streams interactions

Stream-stream analyzes the interaction between a stellar stream and a disrupting dark-matter halo. It requires galpy (ascl:1411.008), NEMO (ascl:1010.051), and the usual common scientific Python packages.

[ascl:1702.010]
streamgap-pepper: Effects of peppering streams with many small impacts

streamgap-pepper computes the effect of subhalo fly-bys on cold tidal streams based on the action-angle representation of streams. A line-of-parallel-angle approach is used to calculate the perturbed distribution function of a given stream segment by undoing the effect of all impacts. This approach allows one to compute the perturbed stream density and track in any coordinate system in minutes for realizations of the subhalo distribution down to 10^5 Msun, accounting for the stream's internal dispersion and overlapping impacts. This code uses galpy (ascl:1411.008) and the streampepperdf.py galpy extension, which implements the fast calculation of the perturbed stream structure.

[ascl:1106.021]
StringFast: Fast Code to Compute CMB Power Spectra induced by Cosmic Strings

StringFast implements a method for efficient computation of the C_l spectra induced by a network of strings, which is fast enough to be used in Markov Chain Monte Carlo analyses of future data. This code allows the user to calculate TT, EE, and BB power spectra (scalar [for TT and EE], vector, and tensor modes) for "wiggly" cosmic strings. StringFast uses the output of the public code CMBACT (ascl:1106.023). The properties of the strings are described by four parameters: Gμ—dimensionless string tension; v—rms transverse velocity (as fraction of c); α—"wiggliness"; ξ—comoving correlation length of the string network. It is written as a Fortran 90 module.

[ascl:2404.025]
stringgen: Scattering based cosmic string emulation

Price, Matthew A.; Mars, Matthijs; Docherty, Matthew M.; Spurio Mancini, Alessio; Marignier, Augustin; McEwen, Jason D.

stringgen creates emulations of cosmic string maps with statistics similar to those of a single (or small ensemble) of reference simulations. It uses wavelet phase harmonics to calculate a compressed representation of these reference simulations, which may then be used to synthesize new realizations with accurate statistical properties, *e.g.*, 2 and 3 point correlations, skewness, kurtosis, and Minkowski functionals.

[ascl:2401.019]
StructureFunction: Bayesian estimation of the AGN structure function for Poisson data

Georgakakis, A.; Buchner, J.; Ruiz, A.; Boller, T.; Akylas, A.; Paolillo, M.; Salvato, M.; Merloni, A.; Nandra, K.; Dwelly, T.

StructureFunction determines the X-ray Structure Function of a population of Active Galactic Nuclei (AGN) for which two epoch X-ray observations are available and are separated by rest frame time interval. The calculation of the X-ray structure function is Bayesian. The sampling of the likelihood uses Stan (ascl:1801.003) for statistical modeling and high-performance statistical computation.

[ascl:1206.003]
STSDAS: IRAF Tools for Hubble Space Telescope data reduction

The Space Telescope Science Data Analysis System (STSDAS) is a software package for reducing and analyzing astronomical data. It is layered on top of IRAF and provides general-purpose tools for astronomical data analysis as well as routines specifically designed for HST data. In particular, STSDAS contains all the programs used for the calibration and reduction of HST data in the STScI post-observation processing pipelines.

[ascl:2010.003]
stsynphot: synphot for HST and JWST

An extension to synphot (ascl:1811.001), stsynphot implements synthetic photometry package for HST and JWST support. The software constructs spectra from various grids of model atmosphere spectra, parameterized spectrum models, and atlases of stellar spectrophotometry. It also simulates observations specific to HST and JWST, computes photometric calibration parameters for any supported instrument mode, and plots instrument-specific sensitivity curves and calibration target spectra.

[ascl:1010.067]
Stuff: Simulating “Perfect” Astronomical Catalogues

Stuff is a program that simulates “perfect” astronomical catalogues. It generate object lists in ASCII which can read by the SkyMaker program to produce realistic astronomical fields. Stuff is part of the EFIGI development project.

[ascl:2312.035]
SubGen: Fast subhalo sampler

SubGen generates Monte-Carlo samples of dark matter subhaloes. It fully describes the joint distribution of subhaloes in final mass, infall mass, and radius; it can be used to predict derived distributions involving combinations of these quantities, including the universal subhalo mass function, the subhalo spatial distribution, the gravitational lensing profile, the dark matter annihilation radiation profile and boost factor. SubGen works only for CDM subhaloes; for an extension of the code to also work with WDM subhaloes, see SubGen2 (ascl:2312.036).

[ascl:2312.036]
SubGen2: Subhalo population generator

The SubGen2 subhalo population generator works for both CDM and WDM of arbitrary DM particle mass. It can be used to generate a population of subhaloes according to the joint distribution of subhalo bound mass, infall mass and halo-centric distance in a halo of a given mass. SubGen2 is an extension to SubGen (ascl:2312.035), which works only for CDM subhaloes.

[ascl:2306.050]
SubgridClumping: Clumping factor for large low-resolution N-body simulations

Bianco, Michele; Iliev, Ilian T.; Ahn, Kyungjin; Giri, Sambit K.; Mao, Yi; Park, Hyunbae; Shapiro, Paul R.

SubgridClumping derives the parameters for the global, in-homogeneous and stochastic clumping model and then computes the clumping factor for large low-resolution N-body simulations smoothed on a regular grid. Written for the CUBEP3M simulation, the package contains two main modules. The first derives the three clumping model parameters for a given small high-resolution simulation; the second computes a clumping factor cube (same mesh-size as input) for the three models for the given density field of a large low-resolution simulation.

[ascl:2405.015]
sunbather: Escaping exoplanet atmospheres and transit spectra simulator

sunbather simulates the upper atmospheres of exoplanets and their observational signatures. The code constructs 1D Parker wind profiles using p-winds (ascl:2111.011) to simulate these with Cloudy (ascl:9910.001), and postprocesses the output with a custom radiative transfer module to predict the transmission spectra of exoplanets.

[ascl:2312.015]
SUNBIRD: Neural-network-based models for galaxy clustering

Cuesta-Lazaro, Carolina; Paillas, Enrique; Yuan, Sihan; Cai, Yan-Chuan; Nadathur, Seshadri; Percival, Will J.; Beutler, Florian; de Mattia, Arnaud; Eisenstein, Daniel; Forero-Sanchez, Daniel; Padilla, Nelson; Pinon, Mathilde; Ruhlmann-Kleider, Vanina; Sánchez, Ariel G.; Valogiannis, Georgios; Zarrouk, Pauline

SUNBIRD trains neural-network-based models for galaxy clustering. It also incorporates pre-trained emulators for different summary statistics, including galaxy two-point correlation function, density-split clustering statistics, and old-galaxy cross-correlation function. These models have been trained on mock galaxy catalogs, and were calibrated to work for specific samples of galaxies. SUNBIRD implements routines with PyTorch to train new neural-network emulators.

[ascl:2202.024]
SunnyNet: Neural network framework for solving 3D NLTE radiative transfer in stellar atmospheres

SunnyNet learns the mapping the between LTE and NLTE populations of a model atom and predicts the NLTE populations based on LTE populations for an arbitrary 3D atmosphere. To use SunnyNet, one must already have a set of LTE and NLTE populations computed in 3D, to train the network. These must come from another code, as SunnyNet is unable to solve the formal problem. Once SunnyNet is trained, one can feed it LTE populations from a different 3D atmosphere, and obtain predicted NLTE populations. The NLTE populations can then be used to synthesize any spectral line that is included in the model atom. SunnyNet's output is a file with predicted NLTE populations. SunnyNet itself does not calculate synthetic spectra, but a sample script written in the Julia language that quickly computes Hα spectra is included.

[ascl:1401.010]
sunpy: Python for Solar Physicists

sunpy is a community-developed free and open-source software package for solar physics and is an alternative to the SolarSoft (ascl:1208.013) data analysis environment. SunPy provides data structures for representing the most common solar data types (images, lightcurves, and spectra) and integration with the Virtual Solar Observatory (VSO) and the Heliophysics Event Knowledgebase (HEK) for data acquisition.

[ascl:1303.030]
Sunrise: Radiation transfer through interstellar dust

Sunrise is a Monte Carlo radiation transfer code for calculating absorption and scattering of light to study the effects of dust in hydrodynamic simulations of interacting galaxies. It uses an adaptive mesh refinement grid to describe arbitrary geometries of emitting and absorbing/scattering media, with spatial dynamical range exceeding 10^{4}; it can efficiently generate images of the emerging radiation at arbitrary points in space and spectral energy distributions of simulated galaxies run with the Gadget (ascl:0003.001), Gasoline (ascl:1710.019), Arepo (ascl:1909.010), Enzo (ascl:1010.072) or ART codes. In addition to the monochromatic radiative transfer typically used by Monte Carlo codes, Sunrise can propagate a range of wavelengths simultaneously. This "polychromatic" algorithm gives significant improvements in efficiency and accuracy when spectral features are calculated.

[ascl:1105.007]
Sunspot Models

These IDL codes create a thick magneto-static structure with parameters of a typical sunspot in a solar like photosphere - chromosphere. The variable parameters are field strength on the axis, radius, and Wilson depression (displacement of the atmosphere on the axis with respect to the field-free atmosphere). Output are magnetic field vector, pressure and density distributions with radius and height. The structure has azimuthal symmetry. The codes are relatively self explanatory and the download packages contain README files.

[ascl:2404.009]
superABC: Cosmological constraints from SN light curves using Approximate Bayesian Computation

The superABC sampling method obtains cosmological constraints from supernova light curves using Approximate Bayesian Computation (ABC) without any likelihood assumptions. It provides an interface to two forward model simulations, SNCosmo (ascl:1611.017) and SNANA (ascl:1010.027), for supernova cosmology.

[ascl:1109.007]
SuperBayeS: Supersymmetry Parameters Extraction Routines for Bayesian Statistics

SuperBayeS is a package for fast and efficient sampling of supersymmetric theories. It uses Bayesian techniques to explore multidimensional SUSY parameter spaces and to compare SUSY predictions with observable quantities, including sparticle masses, collider observables, dark matter abundance, direct detection cross sections, indirect detection quantities etc. Scanning can be performed using Markov Chain Monte Carlo (MCMC) technology or even more efficiently by employing a new scanning technique called MultiNest (ascl:1109.006). which implements the nested sampling algorithm. Using MultiNest, a full 8-dimensional scan of the CMSSM takes about 12 hours on 10 2.4GHz CPUs. There is also an option for old-style fixed-grid scanning. A discussion forum for SuperBayeS is available.

The package combines SoftSusy, DarkSusy, FeynHiggs, Bdecay, MultiNest and MicrOMEGAs. Some of the routines and the plotting tools are based on CosmoMC (ascl:1106.025).

SuperBayeS comes with SuperEGO, a MATLAB graphical user interface tool for interactive plotting of the results. SuperEGO has been developed by Rachid Lemrani and is based on CosmoloGUI by Sarah Bridle.

[ascl:1609.019]
SuperBoL: Module for calculating the bolometric luminosities of supernovae

SuperBoL calculates the bolometric lightcurves of Type II supernovae using observed photometry; it includes three different methods for calculating the bolometric luminosity: quasi-bolometric, direct, and bolometric correction. SuperBoL propagates uncertainties in the input data through the calculations made by the code, allowing for error bars to be included in plots of the lightcurve.

[ascl:1507.002]
SUPERBOX: Particle-multi-mesh code to simulate galaxies

SUPERBOX is a particle-mesh code that uses moving sub-grids to track and resolve high-density peaks in the particle distribution and a nearest grid point force-calculation scheme based on the second derivatives of the potential. The code implements a fast low-storage FFT-algorithm and allows a highly resolved treatment of interactions in clusters of galaxies, such as high-velocity encounters between elliptical galaxies and the tidal disruption of dwarf galaxies, as sub-grids follow the trajectories of individual galaxies. SUPERBOX is efficient in that the computational overhead is kept as slim as possible and is also memory efficient since it uses only one set of grids to treat galaxies in succession.

[ascl:1511.001]
SuperFreq: Numerical determination of fundamental frequencies of an orbit

SuperFreq numerically estimates the fundamental frequencies and orbital actions of pre-computed orbital time series. It is an implementation of a version of the Numerical Analysis of Fundamental Frequencies close to that by Monica Valluri, which itself is an implementation of an algorithm first used by Jacques Laskar.

[ascl:2406.018]
SuperLite: Spectral synthesis code for interacting transients

SuperLite produces synthetic spectra for astrophysical transient phenomena affected by circumstellar interaction. It uses Monte Carlo methods and multigroup structured opacity calculations for semi-implicit, semirelativistic radiation transport in high-velocity shocked outflows, and can reproduce spectra of typical Type Ia, Type IIP, and Type IIn supernovae. SuperLite also generates high-quality spectra that can be compared with observations of transient events, including superluminous supernovae, pulsational pair-instability supernovae, and other peculiar transients.

[ascl:2008.009]
SuperNNova: Photometric classification

SuperNNova performs photometric classification by leveraging recent advances in deep neural networks. It can train either a recurrent neural network or random forest to classify light-curves using only photometric information. It also allows additional information, such as host-galaxy redshift, to be incorporated to improve performance.

[ascl:1109.014]
Supernova Flux-averaging Likelihood Code

Flux-averaging justifies the use of the distance-redshift relation for a smooth universe in the analysis of type Ia supernova (SN Ia) data. Flux-averaging of SN Ia data is required to yield cosmological parameter constraints that are free of the bias induced by weak gravitational lensing. SN Ia data are converted into flux. For a given cosmological model, the distance dependence of the data is removed, then the data are binned in redshift, and placed at the average redshift in each redshift bin. The likelihood of the given cosmological model is then computed using "flux statistics''. These Fortran codes compute the likelihood of an arbitrary cosmological model [with given H(z)/H_0] using flux-averaged Type Ia supernova data.

[ascl:1705.017]
supernovae: Photometric classification of supernovae

Supernovae classifies supernovae using their light curves directly as inputs to a deep recurrent neural network, which learns information from the sequence of observations. Observational time and filter fluxes are used as inputs; since the inputs are agnostic, additional data such as host galaxy information can also be included.

[ascl:2103.019]
SUPERNU: Radiative transfer code for explosive outflows using Monte Carlo methods

SuperNu simulates time-dependent radiation transport in local thermodynamic equilibrium with matter. It applies the methods of Implicit Monte Carlo (IMC) and Discrete Diffusion Monte Carlo (DDMC) for static or homologously expanding spatial grids. The radiation field affects material temperature but does not affect the motion of the fluid. SuperNu may be applied to simulate radiation transport for supernovae with ejecta velocities that are not affected by radiation momentum. The physical opacity calculation includes elements from Hydrogen up to Cobalt. SuperNu is motivated by the ongoing research into the effect of variation in the structure of progenitor star explosions on observables: the brightness and shape of light curves and the temporal evolution of the spectra. Consequently, the code may be used to post-process data from hydrodynamic simulations. SuperNu does not include any capabilities or methods that allow for non-trivial hydrodynamics.

[ascl:1612.015]
Superplot: Graphical interface for plotting and analyzing data

Superplot calculates and plots statistical quantities relevant to parameter inference from a "chain" of samples drawn from a parameter space produced by codes such as MultiNest (ascl:1109.006), BAYES-X (ascl:1505.027), and PolyChord (ascl:1502.011). It offers a graphical interface for browsing a chain of many variables quickly and can produce numerous kinds of publication quality plots, including one- and two-dimensional profile likelihood, three-dimensional scatter plots, and confidence intervals and credible regions. Superplot can also save plots in PDF format, create a summary text file, and export a plot as a pickled object for importing and manipulating in a Python interpreter.

[ascl:2306.016]
SuperRad: Black hole superradiance gravitational waveform modeler

SuperRad models ultralight boson clouds that arise through black hole superradiance. It uses numerical results in the relativistic regime combined with analytic estimates to describe the dynamics and gravitational wave signals of ultralight scalar or vector clouds. Written in Python, SuperRad includes a set of testing routines that check the internal consistency of the package; these tests mainly serve the purpose of ensuring functionality of the waveform model but can also be utilized to check that SuperRad works as intended.

[ascl:2008.014]
SuperRAENN: Supernova photometric classification pipeline

SuperRAENN performs photometric classification of supernovae in the following categories: Type I superluminos supernovae, Type II, Type IIn, Type Ia and Type Ib/c. Though the code is optimized for use with complete (rather than realtime) light curves from the Pan-STARRS Medium Deep Survey, the classifier can be trained on other data. SuperRAENN can be used on a dataset containing both spectroscopically labelled and unlabelled SNe; all events will be used to train the RAENN, while labelled events will be used to train the random forest.

[submitted]
Supervised star, galaxy and QSO classification with sharpened dimensionality reduction

Aims. We explore the use of broadband colors to classify stars, galaxies and QSOs. Specifically, we apply sharpened dimensionality reduction (SDR)-aided classification to this problem, with the aim of enhancing cluster separation in the projections of high-dimensional data clusters to allow for better classification performance and more informative projections.

Methods. The main objective of this work is to apply SDR to large sets of broadband colors derived from the CPz catalog first introduced by Fotopoulou & Paltani (2018) to obtain projections with clusters of star, galaxy and QSO data that exhibit a high degree of separation. The SDR method achieves this by combining density-based clustering with conventional dimensionality-reduction techniques. To make SDR scalable and have out-of-sample ability, we use a deep neural network trained to reproduce the SDR projections. Subsequently classification is done by applying a k-nearest neighbors (k-NN) classifier to the sharpened projections.

Results. Based on a qualitative and quantitative analysis of the embeddings produced by SDR, we find that SDR consistently produces accurate projections with a high degree of cluster separation. A number of projection performance metrics are used to evaluate this separation, including the trustworthiness, continuity, Shepard goodness, and distribution consistency metrics. Using the k-NN classifier and consolidating the results of various data sets we obtain precisions of 99.7%, 98.9%, and 98.5% for classifying stars, galaxies, and QSOs, respectively. Furthermore, we achieve completenesses of 97.8%, 99.3%, and 86.8%, respectively. In addition to classification we explore the structure of the embeddings produced by SDR by cross-matching with data from Gaia DR3, Galaxy Zoo 1 and a catalog of specific star formation rates, stellar masses and dust luminosities. We discover that the embeddings reveal astrophysical information, which allows one to understand the structure of the high-dimensional broadband color data in greater detail.

Conclusions. We find that SDR-aided star, galaxy, and QSO classification performs comparably to another unsupervised learning method using hierarchical density-based spatial clustering of applications with noise (HDBSCAN) but offers advantages in terms of scalability and interpretability. Furthermore, it outperforms traditional color selection methods in terms of QSO classification performance. Overall, we demonstrate the potential of SDR-aided classification to provide an accurate and physically insightful classification of astronomical objects based on their broadband colors.

[ascl:2202.004]
SUPPNet: Spectrum normalization neural network

SUPPNet performs fully automated precise continuum normalization of merged echelle spectra and offers flexible manual fine-tuning, if necessary. The code uses a fully convolutional deep neural network (SUPP Network) trained to predict a pseudo-continuum. The post-processing step uses smoothing splines that give access to regressed knots, which are useful for optional manual corrections. The active learning technique controls possible biases that may arise from training with synthetic spectra and extends the applicability of the method to features absent in this kind of spectra.

[ascl:1403.008]
SURF: Submm User Reduction Facility

SURF reduces data from the SCUBA instrument from the James Clerk Maxwell Telescope. Facilities are provided for reducing all the SCUBA observing modes including jiggle, scan and photometry modes. SURF uses the Starlink environment (ascl:1110.012).

[ascl:1809.007]
surfinBH: Surrogate final black hole properties for mergers of binary black holes

surfinBH predicts the final mass, spin and recoil velocity of the remnant of a binary black hole merger. Trained directly against numerical relativity simulations, these models are extremely accurate, reproducing the results of the simulations at the same level of accuracy as the simulations themselves. Fits such as these play a crucial role in waveform modeling and tests of general relativity with gravitational waves, performed by LIGO.

[ascl:1605.017]
Surprise Calculator: Estimating relative entropy and Surprise between samples

The Surprise is a measure for consistency between posterior distributions and operates in parameter space. It can be used to analyze either the compatibility of separately analyzed posteriors from two datasets, or the posteriors from a Bayesian update. The Surprise Calculator estimates relative entropy and Surprise between two samples, assuming they are Gaussian. The software requires the R package CompQuadForm to estimate the significance of the Surprise, and rpy2 to interface R with Python.

[ascl:1804.016]
surrkick: Black-hole kicks from numerical-relativity surrogate models

surrkick quickly and reliably extract recoils imparted to generic, precessing, black hole binaries. It uses a numerical-relativity surrogate model to obtain the gravitational waveform given a set of binary parameters, and from this waveform directly integrates the gravitational-wave linear momentum flux. This entirely bypasses the need of fitting formulae which are typically used to model black-hole recoils in astrophysical contexts.

[ascl:1208.012]
Swarm-NG: Parallel n-body Integrations

Dindar, Saleh; Ford, Eric B.; Juric, Mario; Young, In Yeo; Gao, Jianwei; Boley, Aaron C.; Nelson, Benjamin; Peters, Jorg

Swarm-NG is a C++ library for the efficient direct integration of many n-body systems using highly-parallel Graphics Processing Units (GPU). Swarm-NG focuses on many few-body systems, e.g., thousands of systems with 3...15 bodies each, as is typical for the study of planetary systems; the code parallelizes the simulation, including both the numerical integration of the equations of motion and the evaluation of forces using NVIDIA's "Compute Unified Device Architecture" (CUDA) on the GPU. Swarm-NG includes optimized implementations of 4th order time-symmetrized Hermite integration and mixed variable symplectic integration as well as several sample codes for other algorithms to illustrate how non-CUDA-savvy users may themselves introduce customized integrators into the Swarm-NG framework. Applications of Swarm-NG include studying the late stages of planet formation, testing the stability of planetary systems and evaluating the goodness-of-fit between many planetary system models and observations of extrasolar planet host stars (e.g., radial velocity, astrometry, transit timing). While Swarm-NG focuses on the parallel integration of many planetary systems,the underlying integrators could be applied to a wide variety of problems that require repeatedly integrating a set of ordinary differential equations many times using different initial conditions and/or parameter values.

[ascl:1010.068]
SWarp: Resampling and Co-adding FITS Images Together

SWarp resamples and co-adds together FITS images using any arbitrary astrometric projection defined in the WCS standard. It operates on pre-reduced images and their weight-maps. Based on the astrometric and photometric calibrations derived at an earlier phase of the pipeline, SWarp re-maps ("warps") the pixels to a perfect projection system, and co-adds them in an optimum way, according to their relative weights. SWarp's astrometric engine is based on a customized version of Calabretta's WCSLib 2.6 and supports all of the projections defined in the 2000 version of the WCS proposal.

[ascl:1303.001]
SWIFT: A solar system integration software package

SWIFT follows the long-term dynamical evolution of a swarm of test particles in the solar system. The code efficiently and accurately handles close approaches between test particles and planets while retaining the powerful features of recently developed mixed variable symplectic integrators. Four integration techniques are included: Wisdom-Holman Mapping; Regularized Mixed Variable Symplectic (RMVS) method; fourth order T+U Symplectic (TU4) method; and Bulirsch-Stoer method. The package is designed so that the calls to each of these look identical so that it is trivial to replace one with another. Complex data manipulations and results can be analyzed with the graphics packace SwiftVis.

[ascl:1805.020]
SWIFT: SPH With Inter-dependent Fine-grained Tasking

Schaller, Matthieu; Gonnet, Pedro; Draper, Peter W.; Chalk, Aidan B. G.; Bower, Richard G.; Willis, James; Hausammann, Loïc

SWIFT runs cosmological simulations on peta-scale machines for solving gravity and SPH. It uses the Fast Multipole Method (FMM) to calculate gravitational forces between nearby particles, combining these with long-range forces provided by a mesh that captures both the periodic nature of the calculation and the expansion of the simulated universe. SWIFT currently uses a single fixed but time-variable softening length for all the particles. Many useful external potentials are also available, such as galaxy haloes or stratified boxes that are used in idealised problems. SWIFT implements a standard LCDM cosmology background expansion and solves the equations in a comoving frame; equations of state of dark-energy evolve with scale-factor. The structure of the code allows implementation for modified-gravity solvers or self-interacting dark matter schemes to be implemented. Many hydrodynamics schemes are implemented in SWIFT and the software allows users to add their own.

[ascl:2309.003]
Swiftbat: Utilities for handing BAT instrument data from the Neil Gehrels Swift Observatory

Swiftbat retrieves, analyzes, and displays data from NASA's Swift spacecraft, especially data from the Swift Burst Alert Telescope (BAT). All BAT data are available from the Swift data archive; however, a few routines in this library use data access methods not available to the general public and thus are useful only to Swift team members. The package also installs a command-line program 'swinfo' that provides Swift Information such as what the MET (onboard-clock) time is, where Swift was pointing, and whether a specific source was above the horizon and/or in the field of view.

[submitted]
Swiftest

Swiftest is a software package designed to model the long-term dynamics of system of bodies in orbit around a dominant central body, such a planetary system around a star, or a satellite system around a planet. The main body of the program is written in Modern Fortran, taking advantage of the object-oriented capabilities included with Fortran 2003 and the parallel capabilities included with Fortran 2008 and Fortran 2018. Swiftest also includes a Python package that allows the user to quickly generate input, run simulations, and process output from the simulations. Swiftest uses a NetCDF output file format which makes data analysis with the Swiftest Python package a streamlined and flexible process for the user. Building off a strong legacy, including its predecessors Swifter and Swift, Swiftest takes the next step in modeling the dynamics of planetary systems by improving the performance and ease of use of software, and by introducing a new collisional fragmentation model. Currently, Swiftest includes the four main symplectic integrators included in its predecessors: WHM, RMVS, HELIO, and SyMBA. In addition, Swiftest also contains the Fraggle model for generating products of collisional fragmentation.

[submitted]
SWIFTGalaxy

SWIFTGalaxy provides a software abstraction of simulated galaxies produced by the SWIFT smoothed particle hydrodynamics code. It extends the SWIFTSimIO module and is tailored to analyses of particles belonging to individual simulated galaxies. It inherits from and extends the functionality of the SWIFTDataset. It understands the output of halo finders and therefore which particles belong to a galaxy, and its integrated properties. The particles occupy a coordinate frame that is enforced to be consistent, such that particles loaded on-the-fly will match e.g. rotations and translations of particles already in memory. Intuitive masking of particle datasets is also enabled. Finally, some utilities to make working in cylindrical and spherical coordinate systems more convenient are also provided.

[ascl:1112.018]
SwiftVis: Data Analysis & Visualization For Planetary Science

SwiftVis is a tool originally developed as part of a rewrite of Swift (ascl:1303.001) to be used for analysis and plotting of simulations performed with Swift and Swifter. The extensibility built into the design has allowed us to make SwiftVis a general purpose analysis and plotting package customized to be usable by the planetary science community at large. SwiftVis is written in Java and has been tested on Windows, Linux, and Mac platforms. Its graphical interface allows users to do complex analysis and plotting without having to write custom code.

[ascl:2012.022]
SWIGLAL: Access LALSuite libraries with Python and Octave scripts

SWIGLAL, a wrapper for and component of the LALSuite (ascl:2012.021) gravitational wave detection and analysis libraries, which are primarily written in C, makes LALSuite routines directly accessible to Python and Octave scripts.

[ascl:1606.001]
SWOC: Spectral Wavelength Optimization Code

SWOC (Spectral Wavelength Optimization Code) determines the wavelength ranges that provide the optimal amount of information to achieve the required science goals for a spectroscopic study. It computes a figure-of-merit for different spectral configurations using a user-defined list of spectral features, and, utilizing a set of flux-calibrated spectra, determines the spectral regions showing the largest differences among the spectra.

[ascl:2110.014]
swordfish: Information yield of counting experiments

Swordfish studies the information yield of counting experiments. It implements at its core a rather general version of a Poisson point process with background uncertainties described by a Gaussian random field, and provides easy access to its information geometrical properties. Based on this information, a number of common and less common tasks can be performed. Swordfish allows quick and accurate forecasts of experimental sensitivities without time-intensive Monte Carlos, mock data generation and likelihood maximization. It can:

- calculate the expected upper limit or discovery reach of an instrument;

- derive expected confidence contours for parameter reconstruction;

- visualize confidence contours as well as the underlying information metric field;

- calculate the information flux, an effective signal-to-noise ratio that accounts for background systematics and component degeneracies; and

- calculate the Euclideanized signal which approximately maps the signal to a new vector which can be used to calculate the Euclidean distance between points.

[ascl:1707.007]
swot: Super W Of Theta

SWOT (Super W Of Theta) computes two-point statistics for very large data sets, based on “divide and conquer” algorithms, mainly, but not limited to data storage in binary trees, approximation at large scale, parellelization (open MPI), and bootstrap and jackknife resampling methods “on the fly”. It currently supports projected and 3D galaxy auto and cross correlations, galaxy-galaxy lensing, and weighted histograms.

[ascl:2302.016]
swyft: Scientific simulation-based inference at scale

swyft implements Truncated Marginal Neural Radio Estimation (TMNRE), a Bayesian parameter inference technique for complex simulation data. The code improves performance by estimating low-dimensional marginal posteriors rather than the joint posteriors of distributions, while also targeting simulations to targets of observational interest via an indicator function. The use of local amortization permits statistical checks, enabling validation of parameters that cannot be performed using sampling-based methods. swyft is also based on stochastic simulations, mapping parameters to observational data, and incorporates a simulator manager.

[ascl:1904.001]
sxrbg: ROSAT X-Ray Background Tool

The ROSAT X-Ray Background Tool (sxrbg) calculates the average X-ray background count rate and statistical uncertainty in each of the six standard bands of the ROSAT All-Sky Survey (RASS) diffuse background maps (R1, R2, R4, R5, R6, R7) for a specified astronomical position and a search region consisting of either a circle with a specified radius or an annulus with specified inner and outer radii centered on the position. The values returned by the tool are in units of 10^-6 counts/second/arcminute^2. sxrbg can also create a count-rate-based spectrum file which can be used with XSpec (ascl:9910.005) to calculate fluxes and offers support for counts statistics (cstat), an alternative method for generating a background spectrum. HEASoft (ascl:1408.004) is a prerequisite for building. The code is in the public domain.

[ascl:1806.019]
SYGMA: Modeling stellar yields for galactic modeling

SYGMA (Stellar Yields for Galactic Modeling Applications) follows the ejecta of simple stellar populations as a function of time to model the enrichment and feedback from simple stellar populations. It is the basic building block of the galaxy code One-zone Model for the Evolution of GAlaxies (OMEGA, ascl:1806.018) and is part of the NuGrid Python Chemical Evolution Environment (NuPyCEE, ascl:1610.015). Stellar yields of AGB and massive stars are calculated with the same nuclear physics and are provided by the NuGrid collaboration.

[ascl:2203.018]
sympy2c: Generating fast C/C++ functions and ODE solvers from symbolic expressions

The Python package sympy2c allows creation and compilation of fast C/C++ based extension modules from symbolic representations. It can create fast code for the solution of high dimensional ODEs, or numerical evaluation of integrals where sympy fails to compute an anti-derivative. Based on the symbolic formulation of a stiff ODE, sympy2c analyzes sparsity patterns in the Jacobian matrix of the ODE, and generates loop-less fast code by unrolling loops in the internally used LU factorization algorithm and by avoiding unnecessary computations involving known zeros.

[ascl:1308.008]
SYN++: Standalone SN spectrum synthesis

SYN++ is a standalone SN spectrum synthesis program. It is a rewrite of the original SYNOW (ascl:1010.055) code in modern C++. It offers further enhancements, a new structured input control file format, and the atomic data files have been repackaged and are more complete than those of SYNOW.

[ascl:1308.007]
SYNAPPS: Forward-modeling of supernova spectroscopy data sets

SYNAPPS is a spectrum fitter embedding a highly parameterized synthetic SN spectrum calculation within a parallel asynchronous optimizer. This open-source code is aimed primarily at the problem of systematically interpreting large sets of SN spectroscopy data.

[submitted]
synchrofit: Python-based synchrotron spectral fitting

The synchrofit (synchrotron fitter) package implements a reduced dimensionality parameterisation of standard synchrotron spectrum models, and provides fitting routines applicable for active galactic nuclei and supernova remnants. The Python code includes the Jaffe-Parola model (JP), Kardashev-Pacholczyk model (KP), and continuous injection models (CI/KGJP) for both constant or Maxwell-Boltzmann magnetic field distributions. An adaptive maximum likelihood algorithm is invoked to fit these models to multi-frequency radio observations; the adaptive mesh is customisable for either optimal precision or computational efficiency. Functions are additionally provided to plot the fitted spectral model with its confidence interval, and to derive the spectral age of the synchrotron emitting particles.

[ascl:1302.014]
SYNMAG Photometry: Catalog-level Matched Colors of Extended Sources

Bundy, Kevin; Hogg, David W.; Higgs, Tim D.; Nichol, Robert C.; Yasuda, Naoki; Masters, Karen L.; Lang, Dustin; Wake, David A.

SYNMAG is a tool for producing synthetic aperture magnitudes to enable fast matched photometry at the catalog level without reprocessing imaging data. Aperture magnitudes are the most widely tabulated flux measurements in survey catalogs; obtaining reliable, matched photometry for galaxies imaged by different observatories represents a key challenge in the era of wide-field surveys spanning more than several hundred square degrees. Methods such as flux fitting, profile fitting, and PSF homogenization followed by matched-aperture photometry are all computationally expensive. An alternative solution called "synthetic aperture photometry" exploits galaxy profile fits in one band to efficiently model the observed, point-spread-function-convolved light profile in other bands and predict the flux in arbitrarily sized apertures.

[ascl:1010.055]
SYNOW: A Highly Parameterized Spectrum Synthesis Code for Direct Analysis of SN Spectra

SYNOW is a highly parameterized spectrum synthesis code used primarily for direct (empirical) analysis of SN spectra. The code is based on simple assumptions : spherical symmetry; homologous expansion; a sharp photosphere that emits a blackbody continuous spectrum; and line formation by resonance scattering, treated in the Sobolev approximation. Synow does not do continuum transport, it does not solve rate equations, and it does not calculate ionization ratios. Its main function is to take line multiple scattering into account so that it can be used in an empirical spirit to make line identifications and estimate the velocity at the photosphere (or pseudo-photosphere) and the velocity interval within which each ion is detected. these quantities provide constraints on the composition structure of the ejected matter.

[ascl:1811.001]
synphot: Synthetic photometry using Astropy

Synphot simulates photometric data and spectra, observed or otherwise. It can incorporate the user's filters, spectra, and data, and use of a pre-defined standard star (Vega), bandpass, or extinction law. synphot can also construct complicated composite spectra using different models, simulate observations, and compute photometric properties such as count rate, effective wavelength, and effective stimulus. It can manipulate a spectrum by, for example, applying redshift, or normalize it to a given flux value in a given bandpass. Synphot can also sample a spectrum at given wavelengths, plot a quick-view of a spectrum, and perform repetitive operations such as simulating the observations of multiple type of sources through multiple bandpasses. Synphot understands Astropy (ascl:1304.002) models and units and is an Astropy affiliated package. Support for HST and JWST is available through the extension stsynphot (ascl:2010.003).

[ascl:1109.022]
Synspec: General Spectrum Synthesis Program

Synspec is a user-oriented package written in FORTRAN for modeling stellar atmospheres and for stellar spectroscopic diagnostics. It assumes an existing model atmosphere, calculated previously with Tlusty or taken from the literature (for instance, from the Kurucz grid of models). The opacity sources (continua, atomic and molecular lines) are fully specified by the user. An arbitrary stellar rotation and instrumental profile can be applied to the synthetic spectrum.

[ascl:1212.010]
Synth3: Non-magnetic spectrum synthesis code

Synth3 is a non-magnetic spectrum synthesis code. It works with model atmospheres in Kurucz format and VALD Sf line lists and features element stratification, molecular equilibrium and individual microturbulence for each line. Disk integration can be done with s3di which is included in the archive. Synth3 computes spectra emergent from the stellar atmospheres with a depth-dependent chemical composition if depth-dependent abundance is provided in the input model atmosphere file.

[ascl:2307.014]
Synthetic LISA: Simulator for LISA-like gravitational-wave observatories

Synthetic LISA simulates the LISA science process at the level of scientific and technical requirements. The code generates synthetic time series of the LISA fundamental noises, as filtered through all the TDI observables, and provides a streamlined module to compute the TDI responses to gravitational waves, according to a full model of TDI, including the motion of the LISA array, and the temporal and directional dependence of the armlengths.

[ascl:2209.014]
SyntheticISOs: Synthetic Population of Interstellar Objects

Synthetic Population of Interstellar Objects generates a synthetic population of interstellar objects (orbits and sizes) in arbitrary volume of space around the Sun. The only necessary assumption is that the population of ISOs in the interstellar space (far from any massive body) is homogeneous and isotropic. The assumed distribution of interstellar velocities of ISOs has to be provided as an input. This distribution can be defined analytically, but also in a discrete form. The algorithm, based on the multivariate inverse transform sampling method, is implemented in Python.

[ascl:2401.010]
SYSNet: Neural Network modeling of imaging systematics in galaxy surveys

The Feed Forward Neural Network SYSNet models the relationship between the imaging maps, such as stellar density and the observed galaxy density field, in order to mitigate the systematic effects and to make a robust galaxy clustering measurements. The cost function is Mean Squared Error and a L2 regularization term, and the optimization algorithm is Adaptive Moment (ADAM).

[ascl:1210.018]
Systemic Console: Advanced analysis of exoplanetary data

Systemic Console is a tool for advanced analysis of exoplanetary data. It comprises a graphical tool for fitting radial velocity and transits datasets and a library of routines for non-interactive calculations. Among its features are interactive plotting of RV curves and transits, combined fitting of RV and transit timing (primary and secondary), interactive periodograms and FAP estimation, and bootstrap and MCMC error estimation. The console package includes public radial velocity and transit data.

[ascl:1304.018]
SZpack: Computation of Sunyaev-Zeldovich (SZ) signals

SZpack is a numerical library which allows fast and precise computation of the Sunyaev-Zeldovich (SZ) signal for hot, moving clusters of galaxies. Both explicit numerical integration as well as approximate representation of the SZ signals can be obtained. Variations of the electron temperature and bulk velocity along the line-of-sight can be included. SZpack allows very fast and precise (<~0.001% at frequencies h nu <~ 30kT_g and electron temperature kTe ~ 75 keV) computation and its accuracy practically eliminates uncertainties related to more expensive numerical evaluation of the Boltzmann collision term. It furthermore cleanly separates kinematic corrections from scattering physics, effects that previously have not been clarified.

[ascl:1511.006]
T-Matrix: Codes for Computing Electromagnetic Scattering by Nonspherical and Aggregated Particles

The T-Matrix package includes codes to compute electromagnetic scattering by homogeneous, rotationally symmetric nonspherical particles in fixed and random orientations, randomly oriented two-sphere clusters with touching or separated components, and multi-sphere clusters in fixed and random orientations. All codes are written in Fortran-77. LAPACK-based, extended-precision, Gauss-elimination- and NAG-based, and superposition codes are available, as are double-precision superposition, parallelized double-precision, double-precision Lorenz-Mie codes, and codes for the computation of the coefficients for the generalized Chebyshev shape.

[ascl:1609.001]
T-PHOT: PSF-matched, prior-based, multiwavelength extragalactic deconfusion photometry

Merlin, E.; Fontana, A.; Ferguson, H. C.; Dunlop, J. S.; Elbaz, D.; Bourne, N.; Bruce, V. A.; Buitrago, F.; Castellano, M.; Schreiber, C.; Grazian, A.; McLure, R. J.; Okumura, K.; Shu, X.; Wang, T.; Amorín, R.; Boutsia, K.; Cappelluti, N.; Comastri, A.; Derriere, S.; Faber, S. M.; Santini, P.

T-PHOT extracts accurate photometry from low-resolution images of extragalactic fields, where the blending of sources can be a serious problem for accurate and unbiased measurement of fluxes and colors. It gathers data from a high-resolution image of a region of the sky and uses the source positions and morphologies to obtain priors for the photometric analysis of the lower resolution image of the same field. T-PHOT handles different types of datasets as input priors, including a list of objects that will be used to obtain cutouts from the real high-resolution image, a set of analytical models (as .fits stamps), and a list of unresolved, point-like sources, useful for example for far-infrared wavelength domains. T-PHOT yields accurate estimations of fluxes within the intrinsic uncertainties of the method when systematic errors are taken into account (which can be done using a flagging code given in the output), and handles multiwavelength optical to far-infrared image photometry. T-PHOT was developed as part of the ASTRODEEP project (www.astrodeep.eu).

[ascl:1906.008]
T-RECS: Tiered Radio Extragalactic Continuum Simulation

Bonaldi, Anna; Bonato, Matteo; Galluzzi, Vincenzo; Harrison, Ian; Massardi, Marcella; Kay, Scott; De Zotti, Gianfranco; Brown, Michael L.

T-RECS produces radio sources catalogs with user-defined frequencies, area and depth. It models two main populations of radio galaxies, Active Galactic Nuclei (AGNs) and Star-Forming Galaxies (SFGs), and corresponding sub-populations. T-RECS is not computationally demanding and can be run multiple times, using the same catalog inputs, to project the simulated sky onto different fields.

[ascl:1403.014]
T(dust) as a function of sSFR

This IDL code returns the dust temperature of a galaxy from its redshift, SFR and stellar mass; it can also predict the observed monochromatic fluxes of the galaxy. These monochromatic fluxes correspond to those of a DH SED template with the appropriate dust temperature and the appropriate normalization. Dust temperatures and fluxes predictions are only valid and provided in the redshift, stellar mass, SSFR and wavelength ranges 0 < z < 2.5, Mstar> 10^10 Msun, 10^-11 < SSFR[yr-1]< 10^-7 and 30um < lambda_rest < 2mm.

[ascl:1210.006]
TA-DA: A Tool for Astrophysical Data Analysis

TA-DA is a pre-compiled IDL widget-based application which greatly simplifies and improves the analysis of stellar photometric data in comparison with theoretical models and allows the derivation of stellar parameters from multi-band photometry. It is flexible and can address a number of problems, from the interpolation of stellar models or sets of stellar physical parameters in general to the computation of synthetic photometry in arbitrary filters or units. It also analyzes observed color-magnitude diagrams and allows a Bayesian derivation of stellar parameters (and extinction) based on multi-band data.

[ascl:1303.010]
TAC-maker: Transit Analytical Curve maker

TAC-maker allows for rapid and interactive calculation of synthetic planet transits by numerical computations of the integrals, allowing the use of an arbitrary limb-darkening law of the host star. This advantage together with the practically arbitrary precision of the calculations makes the code a valuable tool for the continuously increasing photometric precision of ground-based and space observations.

Would you like to view a random code?