Results 2201-2300 of 2617 (2565 ASCL, 52 submitted)
SPAM searches for imprints of Hu-Sawicki f(R) gravity on the rotation curves of the SPARC (Spitzer Photometry and Accurate Rotation Curves) sample using the MCMC sampler emcee (ascl:1303.002). The code provides attributes for inspecting the MCMC chains and translating names of parameters to indices. The SPAM package also contains plotting scripts.
SPAM is a extension to AIPS for reducing high-resolution, low-frequency radio interferometric observations. Direction-dependent ionospheric calibration and image-plane ripple suppression are among the features that help to make high-quality sub-GHz images. Data reductions are captured in well-tested Python scripts that execute AIPS tasks directly (mostly during initial data reduction steps), call high-level functions that make multiple AIPS or ParselTongue calls, and require few manual operations.
SPAMCART generates synthetic spectral energy distributions and intensity maps from smoothed particle hydrodynamics simulation snapshots. It follows discrete luminosity packets as they propagate through a density field, and computes the radiative equilibrium temperature of the ambient dust from their trajectories. The sources can be extended and/or embedded, and discrete and/or diffuse. The density is not mapped on to a grid, and therefore the calculation is performed at exactly the same resolution as the hydrodynamics. The code strictly adheres to Kirchhoff's law of radiation. The algorithm is based on the Lucy Monte Carlo radiative transfer method and is fairly simple to implement, as it uses data structures that are already constructed for other purposes in modern particle codes
The Seismic Propagation through Active Regions and Convection (SPARC) code was developed by S. Hanasoge. The acoustic wavefield in SPARC is simulated by numerically solving the linearised 3-D Euler equations in Cartesian geometry (e.g., see Hanasoge, Duvall and Couvidat (2007)). Spatial derivatives are calculated using sixth-order compact finite differences (Lele,1992) and time evolution is achieved through the repeated application of an optimized second-order five-stage Runge-Kutta scheme (Hu, 1996). Periodic horizontal boundaries are used.
SpArcFiRe takes as input an image of a galaxy in FITS, JPG, or PNG format, identifies spiral arms, and extracts structural information about the spiral arms. Pixels in each arm segment are listed, enabling image analysis on each segment. The automated method also performs a least-squares fit of a logarithmic spiral arc to the pixels in that segment, giving per-arc parameters, such as the pitch angle, arm segment length, and location, and outputs images showing the steps SpArcFire took to detect arm segments.
SPARK (Software Package for Astronomical Reduction with KMOS) reduces data from the K-band Multi Object Spectrograph (KMOS) for the VLT. In many cases, science data can be processed using a single recipe; alternately, all functions this recipe provides can be performed using other recipes provided as tools. Among the functions the recipes provide are sky subtraction, cube reconstruction with the application of flexure corrections, dividing out the telluric spectrum, applying an illumination correction, aligning the cubes, and then combinging them. The result is a set of files which contain the combined datacube and associated noise cube for each of the 24 integral field unit (IFUs). The pipeline includes simple error propagation.
SparseBLS uses the Box-fitting Least Squares (BLS) algorithm to detect transiting exoplanets in photometric data. SparseBLS does not bin data into phase bins and does not use a phase grid. Because its detection efficiency does not depend on the transit phase, it is significantly faster than BLS for sparse data and is well-suited for large photometric surveys producing unevenly-sampled sparse light curves, such as Gaia.
SparsePZ uses sparse basis representation to fully represent individual photometric redshift probability density functions (PDFs). This approach requires approximately half the parameters for the same multi-Gaussian fitting accuracy, and has the additional advantage that an entire PDF can be stored by using a 4-byte integer per basis function. Only 10-20 points per galaxy are needed to reconstruct both the individual PDFs and the ensemble redshift distribution, N(z), to an accuracy of 99.9 per cent when compared to the one built using the original PDFs computed with a resolution of δz = 0.01, reducing the required storage of 200 original values by a factor of 10-20. This basis representation can be directly extended to a cosmological analysis, thereby increasing computational performance without losing resolution or accuracy.
SPARTA analyzes periodically-variable spectroscopic observations. Intended for common astronomical uses, SPARTA facilitates analysis of single- and double-lined binaries, high-precision radial velocity extraction, and periodicity searches in complex, high dimensional data. It includes two modules, UNICOR and USuRPER. UNICOR analyzes spectra using 1-d CCF. It includes maximum-likelihood analysis of multi-order spectra and detection of systematic shifts. USuRPER (Unit Sphere Representation PERiodogram) is a phase-distance correlation (PDC) based periodogram and is designed for very high-dimensional data such as spectra.
SPARTA is a post-processing framework for particle-based cosmological simulations. The code is written in pure, MPI-parallelized C and is optimized for high performance. The main purpose of SPARTA is to understand the formation of structure in a dynamical sense, namely by analyzing the trajectories (or orbits) of dark matter particles around their halos. Within this framework, the user can add analysis modules that operate on individual trajectories or entire halos. The initial goal of SPARTA was to compute the splashback radius of halos, but numerous other applications have been implemented as well, including spherical overdensity calculations and tracking subhalos via their constituent particles.
SpcAudace processes long slit spectra with automated pipelines and performs astrophysical analysis of the latter data. These powerful pipelines do all the required steps in one pass: standard preprocessing, masking of bad pixels, geometric corrections, registration, optimized spectrum extraction, wavelength calibration and instrumental response computation and correction. Both high and low resolution long slit spectra are managed for stellar and non-stellar targets. Many types of publication-quality figures can be easily produced: pdf and png plots or annotated time series plots. Astrophysical quantities can be derived from individual or large amount of spectra with advanced functions: from line profile characteristics to equivalent width and periodogram. More than 300 documented functions are available and can be used into TCL scripts for automation. SpcAudace is based on Audela open source software.
SpDust is an IDL program that evaluates the spinning dust emissivity for user-provided environmental conditions. A new version of the code became available in March, 2010.
The DEEP2 DEIMOS Data Reduction Pipeline ("spec2d") is an IDL-based, automated software package designed to reduce Keck/DEIMOS multi-slit spectroscopic observations, collected as part of the DEEP2 Galaxy Redshift Survey. The pipeline is best suited for handling data taken with the 1200 line/mm grating tilted towards the red (lambda_c ~ 7800Å). The spec2d reduction package takes the raw DEIMOS data as its input and produces a variety of outputs including 2-d slit spectra and 1-d object spectra.
Specdre performs spectroscopy data reduction and analysis. General features of the package include data cube manipulation, arc line calibration, resampling and spectral fitting. Particular care is taken with error propagation, including tracking covariance. SPECDRE is distributed as part of the Starlink software collection (ascl:1110.012).
SpecPro is an interactive program for viewing and analyzing spectra, particularly in the context of modern imaging surveys. In addition to displaying the 1D and 2D spectrum, SpecPro can simultaneously display available stamp images as well as the spectral energy distribution of a source. This extra information can help significantly in assessing a spectrum.
Specstack creates stacked spectra using a simple algorithm with sigma-clipping to combine the spectra of galaxies in the rest-frame into a single averaged spectrum. Though written originally for galaxy spectra, it also works for other types of objects. It is written in Python and is started from the command-line.
Studies of astrophysical non-LTE media require the combination of atomic and molecular spectroscopic and collisional data often described differently in various databases. SPECTCOL is a tool that implements VAMDC standards, retrieve relevant information from different databases such as CDMS, HITRAN, BASECOL, and can upload local files. All transfer of data between the client and the databases use the VAMDC-XSAMS schema. The spectroscopic and collisional information is combined and useful outputs (ascii or xsams) are provided for the study of the interstellar medium.
Spectra calculates the power spectrum of a time series equally spaced or not based on the Spectral Correlation Coefficient (Ferraz-Mello 1981, Astron. Journal 86 (4), 619). It is very efficient for detection of low frequencies.
Spectractor extracts spectra from slitless spectrophotometric images and measures the atmospheric transmission on the line of sight if standard stars are targeted. It has been optimized on CTIO images but can be configured to analyze any kind of slitless data that contains the order 0 and the order 1 of a spectrum. In particular, it can be used to estimate the atmospheric transmission of the Vera Rubin Observatory site using the dedicated Auxiliary Telescope.
Spectral-cube provides an easy way to read, manipulate, analyze, and write data cubes with two positional dimensions and one spectral dimension, optionally with Stokes parameters. It is a versatile data container for building custom analysis routines. It provides a uniform interface to spectral cubes, robust to the wide range of conventions of axis order, spatial projections, and spectral units that exist in the wild, and allows easy extraction of cube sub-regions using physical coordinates. It has the ability to create, combine, and apply masks to datasets and is designed to work with datasets too large to load into memory, and provide basic summary statistic methods like moments and array aggregates.
SPECTRE's chief purpose is the manipulation of single-order spectra, and it performs many of the tasks contained in such IRAF routines as "splot" and "rv". It is not meant to replace the much more general capabilities of IRAF, but does some functions in a manner that some might find useful. A brief list of SPECTRE tasks are: spectrum smoothing; equivalent width calculation; continuum rectification; noise spike excision; and spectrum comparison. SPECTRE was written to manipulate coude spectra, and thus is probably most useful for working on high dispersion spectra. Echelle spectra can be gathered from various observatories, reduced to singly-dimensioned spectra using IRAF, then written out as FITS files, thus becoming accessible to SPECTRE. Three different spectra may be manipulated and displayed simultaneously. SPECTRE, written in standard FORTRAN77, can be used only with the SM graphics package.
SpectRes efficiently resamples spectra and their associated uncertainties onto an arbitrary wavelength grid. The Python function works with any grid of wavelength values, including non-uniform sampling, and preserves the integrated flux. This may be of use for binning data to increase the signal to noise ratio, obtaining synthetic photometry, or resampling model spectra to match the sampling of observational data.
SPECTRUM ((C) Richard O. Gray, 1992-2008) is a stellar spectral synthesis program which runs on a number of platforms, including most flavors of UNIX and LINUX. It will also run under Windwos 9x/ME/NT/2000/XP using the Cygwin tools or the distributed Windows binaries. The code for SPECTRUM has been written in the "C" language. SPECTRUM computes the LTE synthetic spectrum given a stellar atmosphere model. SPECTRUM can use as input the fully blanketed stellar atmosphere models of Robert Kurucz including the new models of Castelli and Kurucz, but any other stellar atmosphere model which can be cast into the format of Kurucz's models can be used as well. SPECTRUM can be programmed with "command-line switches" to give a number of different outputs. In the default mode, SPECTRUM computes the stellar-disk-integrated normalized-intensity spectrum, but in addition, SPECTRUM will compute the absolute monochromatic flux from the stellar atmosphere or the specific intensity from any point on the stellar surface.
Specutils provides a basic interface for the loading, manipulation, and common forms of analysis of spectroscopic data. Its generic data containers and accompanying modules can be used to build a particular scientific workflow or higher-level analysis tool. It is an AstroPy (ascl:1304.002) affiliated package, and SpecViz (ascl:1902.011), which is built on top of Specutils, provides a visual, interactive interface to its analysis capabilities.
Specview is a tool for 1-D spectral visualization and analysis of astronomical spectrograms. Written in Java, it is capable of reading all the Hubble Space Telescope spectral data formats as well as data from several other instruments (such as IUE, FUSE, ISO, FORS and SDSS), preview spectra from MAST, and data from generic FITS and ASCII tables. It can read data from Virtual Observatory servers, and read and write spectrogram data in Virtual Observatory SED format. It can also read files in the SPC Galactic format used in the chemistry field. Once ingested, data can be plotted and examined with a large selection of custom settings. Specview supports instrument-specific data quality handling, flexible spectral units conversions, custom plotting attributes, plot annotations, tiled plots, hardcopy to JPEG files and PostScript file or printer, etc. Specview can be used to build wide-band SEDs, overplotting or combining data from the same astronomical source taken with different instruments and/or spectral bands. Data can be further processed with averaging, splicing, detrending, and Fourier filtering tools. Specview has a spectral model fitting capability that enables the user to work with multi-component models (including user-defined models) and fit models to data.
SpecViz interactively visualizes and analyzes 1D astronomical spectra. It reads data from FITS and ASCII tables and allows spectra to be easily plotted and examined. It supports instrument-specific data quality handling, flexible spectral units conversions, custom plotting attributes, plot annotations, tiled plots, among other features. SpecViz includes a measurement tool for spectral lines for performing and recording measurements and a model fitting capability for creating simple (e.g., single Gaussian) or multi-component models (e.g., multiple Gaussians for emission and absorption lines in addition to regions of flat continua). SpecViz is built on top of the Specutils (ascl:1902.012) Astropy-affiliated python library, providing a visual, interactive interface to the analysis capabilities in that library.
SPECX is a general purpose line data reduction system. It can read and write FITS data cubes but has specialist support for the GSD format data from the James Clerk Maxwell Telescope. It includes commands to store and retrieve intermediate spectra in storage registers and perform the fitting and removal of polynomial, harmonic and Gaussian baselines.
SPECX can filter and edit spectra and list and display spectra on a graphics terminal. It is able to perform Fourier transform and power spectrum calculations, process up to eight spectra (quadrants) simultaneously with either the same or different center, and assemble a number of reduced individual spectra into a map file and contour or greyscale any plane or planes of the resulting cube.
Two versions of SPECX are distributed. Version 6.x is the VMS and Unix version and is distributed as part of the Starlink software collection. Version 7.x is a complete rewrite of SPECX distributed for Windows.
SPEGID (Single-Pulse Event Group IDentification) identifies astrophysical pulse candidates as trial single-pulse event groups (SPEGs) by first applying Density Based Spatial Clustering of Applications with Noise (DBSCAN) on trial single-pulse events and then merging the clusters that fall within the expected DM (Dispersion Measure) and time span of astrophysical pulses. SPEGID also calculates the peak score for each SPEG in the S/N versus DM space to identify the expected peak-like shape in the signal-to-noise (S/N) ratio versus DM curve of astrophysical pulses. Additionally, SPEGID groups SPEGs that appear at a consistent DM and therefore are likely emitted from the same source. After running SPEGID, periocity.py can be used to find (or verify) the underlying periodicity among a group of SPEGs (i.e., astrophysical pulse candidates).
spex_to_xspec takes the output from the collisional ionisation equilibrium model in the SPEX spectral modelling and fitting package (ascl:1308.014), and converts it into a form usable by the XSPEC spectral fitting package (ascl:9910.005). For a list of temperatures it computes the line strengths and continuum spectra using SPEX. These are collated and written into an APEC-format table model which can be loaded into Xspec. By allowing SPEX models to be loaded into XSPEC, the program allows easy comparison between the results of the SPEX and APEC codes.
SPEX is optimized for the analysis and interpretation of high-resolution cosmic X-ray spectra. The software is especially suited for fitting spectra obtained by current X-ray observatories like XMM-Newton, Chandra, and Suzaku. SPEX can fit multiple spectra with different model components simultaneously and handles highly complex models with many free parameters.
SPEX provides a uniform interface suitable for the X-ray spectral analysis of a number of solar (or other) instruments in the X and Gamma Ray energy ranges. Part of the SolarSoft (ascl:1208.013) library, this package is suitable for any datastream which can be placed in the form of response vs interval where the response is usually a counting rate (spectrum) and the interval is normally an accumulation over time. Together with an algorithm which can be used to relate a model input spectrum to the observed response, generally a response matrix, the dataset is amenable to analysis with this package. Currently the data from a large number of instruments, including SMM (HXRBS, GRS Gamma, GRS X1, and GRS X2), Yohkoh (HXT, HXS, GRS, and SXT,) CGRO (BATSE SPEC and BATSE LAD), WIND (TGRS), HIREX, and NEAR (PIN). SPEX's next generation software is available in OSPEX (ascl:2007.018), an object-oriented package that is also part of and dependent on SolarSoft.
Spextool (Spectral EXtraction tool) is an IDL-based data reduction package for SpeX, a medium resolution near-infrared spectrograph on the NASA IRTF. It performs all of the steps necessary to produce spectra ready for analysis and publication including non-linearity corrections, flat fielding, wavelength calibration, telluric correction, flux calibration, and order merging.
We describe a fast tree algorithm for gravitational N-body simulation on SIMD parallel computers. The tree construction uses fast, parallel sorts. The sorted lists are recursively divided along their x, y and z coordinates. This data structure is a completely balanced tree (i.e., each particle is paired with exactly one other particle) and maintains good spatial locality. An implementation of this tree-building algorithm on a 16k processor Maspar MP-1 performs well and constitutes only a small fraction (approximately 15%) of the entire cycle of finding the accelerations. Each node in the tree is treated as a monopole. The tree search and the summation of accelerations also perform well. During the tree search, node data that is needed from another processor is simply fetched. Roughly 55% of the tree search time is spent in communications between processors. We apply the code to two problems of astrophysical interest. The first is a simulation of the close passage of two gravitationally, interacting, disk galaxies using 65,636 particles. We also simulate the formation of structure in an expanding, model universe using 1,048,576 particles. Our code attains speeds comparable to one head of a Cray Y-MP, so single instruction, multiple data (SIMD) type computers can be used for these simulations. The cost/performance ratio for SIMD machines like the Maspar MP-1 make them an extremely attractive alternative to either vector processors or large multiple instruction, multiple data (MIMD) type parallel computers. With further optimizations (e.g., more careful load balancing), speeds in excess of today's vector processing computers should be possible.
SpheCow explores the structure and dynamics of any spherical model for galaxies and dark matter haloes. The lightweight and flexible code automatically calculates the dynamical properties, assuming an isotropic or Osipkov-Merritt anisotropic orbital structure, of any model with either an analytical density profile or an analytical surface density profile as a starting point. SpheCow contains readily usable implementations for many standard models, including the Plummer, Hernquist, NFW, Einasto, Sérsic and Nuker models. The code is easily extendable, allowing new models to be added in a straightforward way. The code is publicly available as a set of C++ routines and as a Python module.
Spheral++ provides a steerable parallel environment for performing coupled hydrodynamical and gravitational numerical simulations. Hydrodynamics and gravity are modeled using particle-based methods (SPH and N-Body). It uses an Adaptive Smoothed Particle Hydrodynamics (ASPH) algorithm, provides a total energy conserving compatible hydro mode, and performs fluid and solid material modeling and damage and fracture modeling in solids.
The Spherical Library provides an efficient and accurate mathematical representation of shapes on the celestial sphere, such as sky coverage and footprints. Shapes of arbitrary complexity and size can be dynamically created from simple building blocks, whose exact area is also analytically computed. This methodology is also perfectly suited for censoring problematic parts of datasets, e.g., bad seeing, satellite trails or diffraction spikes of bright stars.
Spheroid determines the size distribution of polarizing interstellar dust grains based on electromagnetic scattering by spheroidal particles. It contains subroutines to treat the case of complex refractive indices, and also includes checks for some limiting cases.
SPHGR (Smoothed-Particle Hydrodynamics Galaxy Reduction) is a python based open-source framework for analyzing smoothed-particle hydrodynamic simulations. Its basic form can run a baryonic group finder to identify galaxies and a halo finder to identify dark matter halos; it can also assign said galaxies to their respective halos, calculate halo & galaxy global properties, and iterate through previous time steps to identify the most-massive progenitors of each halo and galaxy. Data about each individual halo and galaxy is collated and easy to access.
SPHGR supports a wide range of simulations types including N-body, full cosmological volumes, and zoom-in runs. Support for multiple SPH code outputs is provided by pyGadgetReader (ascl:1411.001), mainly Gadget (ascl:0003.001) and TIPSY (ascl:1111.015).
SPHRAY, a Smoothed Particle Hydrodynamics (SPH) ray tracer, is designed to solve the 3D, time dependent, radiative transfer (RT) equations for arbitrary density fields. The SPH nature of SPHRAY makes the incorporation of separate hydrodynamics and gravity solvers very natural. SPHRAY relies on a Monte Carlo (MC) ray tracing scheme that does not interpolate the SPH particles onto a grid but instead integrates directly through the SPH kernels. Given initial conditions and a description of the sources of ionizing radiation, the code will calculate the non-equilibrium ionization state (HI, HII, HeI, HeII, HeIII, e) and temperature (internal energy/entropy) of each SPH particle. The sources of radiation can include point like objects, diffuse recombination radiation, and a background field from outside the computational volume. The MC ray tracing implementation allows for the quick introduction of new physics and is parallelization friendly. A quick Axis Aligned Bounding Box (AABB) test taken from computer graphics applications allows for the acceleration of the raytracing component. We present the algorithms used in SPHRAY and verify the code by performing all the test problems detailed in the recent Radiative Transfer Comparison Project of Iliev et. al. The Fortran 90 source code for SPHRAY and example SPH density fields are made available online.
SPHYNX addresses subsonic hydrodynamical instabilities and strong shocks; it is Newtonian, grounded on the Euler-Lagrange formulation of the smoothed-particle hydrodynamics technique, and density based. SPHYNX uses an integral approach for estimating gradients, a flexible family of interpolators to suppress pairing instability, and incorporates volume elements to provides better partition of the unity.
The SPICE (Spacecraft Planet Instrument C-matrix [“Camera matrix”] Events) toolkit offers a set of building blocks for constructing tools supporting multi-mission, international space exploration programs and research in planetary science, heliophysics, Earth science, and for observations from terrestrial observatories. It computes many kinds of observation geometry parameters, including the ephemerides, orientations, sizes, and shapes of planets, satellites, comets and asteroids. It can also compute the orientation of a spacecraft, its various moving structures, and an instrument's field-of-view location on a planet's surface or atmosphere. It can determine when a specified geometric event occurs, such as when an object is in shadow or is in transit across another object. The SPICE toolkit is available in FORTRAN 77, ANSI C, IDL, and MATLAB.
SpiceyPy is a Python wrapper for the NAIF C SPICE Toolkit (ascl:1903.015). It is compatible with Python 2 and 3, and was written using ctypes.
SPIDERMAN calculates exoplanet phase curves and secondary eclipses with arbitrary surface brightness distributions in two dimensions. The code uses a geometrical algorithm to solve exactly the area of sections of the disc of the planet that are occulted by the star. Approximately 1000 models can be generated per second in typical use, which makes making Markov Chain Monte Carlo analyses practicable. The code is modular and allows comparison of the effect of multiple different brightness distributions for a dataset.
SPIDERz (SuPport vector classification for IDEntifying Redshifts) applies powerful support vector machine (SVM) optimization and statistical learning techniques to custom data sets to obtain accurate photometric redshift (photo-z) estimations. It is written for the IDL environment and can be applied to traditional data sets consisting of photometric band magnitudes, or alternatively to data sets with additional galaxy parameters (such as shape information) to investigate potential correlations between the extra galaxy parameters and redshift.
spinOS calculates binary orbital elements. Given a set of radial velocity measurements of a spectroscopic binary and/or relative position measurement of an astrometric binary, spinOS fits an orbital model by minimizing a chi squared metric. These routines are neatly packaged in a graphical user interface, developed using tkinter, facilitating use. Minimization is achieved by default using a Levenberg-Marquardt algorithm from lmfit [ascl:1606.014]. A Markov Chain Monte Carlo option is available to sample the posterior probability distribution in order to estimate errors on the orbital elements.
SPInS (Stellar Parameters INferred Systematically) provides the age, mass, and radius of a star, among other parameters, from a set of photometric, spectroscopic, interferometric, and/or asteroseismic observational constraints; it also generates error bars and correlations. Derived from AIMS (ascl:1611.014), it relies on a stellar model grid and uses a Bayesian approach to find the PDF of stellar parameters from a set of classical constraints. The heart of SPInS is a MCMC solver coupled with interpolation within a pre-computed stellar model grid. The code can consider priors such as the IMF or SFR and can characterize single stars or coeval stars, such as members of binary systems or of stellar clusters.
SPIPS (Spectro-Photo-Interferometry of Pulsating Stars) combines radial velocimetry, interferometry, and photometry to estimate physical parameters of pulsating stars, including presence of infrared excess, color excess, Teff, and ratio distance/p-factor. The global model-based parallax-of-pulsation method is implemented in Python. Derived parameters have a high level of confidence; statistical precision is improved (compared to other methods) due to the large number of data taken into account, accuracy is improved by using consistent physical modeling and reliability of the derived parameters is strengthened by redundancy in the data.
Spirality measures spiral arm pitch angles by fitting galaxy images to spiral templates of known pitch. Written in MATLAB, the code package also includes GenSpiral, which produces FITS images of synthetic spirals, and SpiralArmCount, which uses a one-dimensional Fast Fourier Transform to count the spiral arms of a galaxy after its pitch is determined.
SPISEA (Stellar Population Interface for Stellar Evolution and Atmospheres) generates single-age, single-metallicity populations (i.e., star clusters). The software (formerly called PyPopStar) provides control over different parameters, including cluster characteristics (age, metallicity, mass, distance); total extinction, differential extinction, and extinction law; stellar evolution and atmosphere models; stellar multiplicity and Initial Mass Function; and photometric filters. SPISEA can be used to create a cluster isochrone in many filters using different stellar models, generate a star cluster at any age with an unusual IMF and unresolved multiplicity, and make a spectrum of a star cluster in integrated light.
SPLASH (formerly SUPERSPHPLOT) visualizes output from (astrophysical) simulations using the Smoothed Particle Hydrodynamics (SPH) method in one, two and three dimensions. Written in Fortran 90, it uses the PGPLOT graphics subroutine library for plotting. It is based around a command-line menu structure but utilizes the interactive capabilities of PGPLOT to manipulate data interactively in the plotting window. SPLASH is fully interactive; visualizations can be changed rapidly at the touch of a button (e.g. zooming, rotating, shifting cross section positions etc). Data is read directly from the code dump format giving rapid access to results and the visualization is advanced forwards and backwards through timesteps by single keystrokes. SPLASH uses the SPH kernel to render plots of not only density but other physical quantities, giving a smooth representation of the data.
SPLAT-VO is an extension of the SPLAT (Spectral Analysis Tool, ascl:1402.007) graphical tool for displaying, comparing, modifying and analyzing astronomical spectra; it includes facilities that allow it to work as part of the Virtual Observatory (VO). SPLAT-VO comes in two different forms, one for querying and downloading spectra from SSAP servers and one for interoperating with VO tools, such as TOPCAT (ascl:1101.010).
SPLAT is a graphical tool for displaying, comparing, modifying and analyzing astronomical spectra stored in NDF, FITS and TEXT files as well as in NDX format. It can read in many spectra at the same time and then display these as line plots. Display windows can show one or several spectra at the same time and can be interactively zoomed and scrolled, centered on specific wavelengths, provide continuous coordinate readout, produce printable hardcopy and be configured in many ways. Analysis facilities include the fitting of a polynomial to selected parts of a spectrum, the fitting of Gaussian, Lorentzian and Voigt profiles to emission and absorption lines and the filtering of spectra using average, median and line-shape window functions as well as wavelet denoising. SPLAT also supports a full range of coordinate systems for spectra, which allows coordinates to be displayed and aligned in many different coordinate systems (wavelength, frequency, energy, velocity) and transformed between these and different standards of rest (topocentric, heliocentric, dynamic and kinematic local standards of rest, etc). SPLAT is distributed as part of the Starlink (ascl:1110.012) software collection.
Splotch is a light and fast, publicly available, ray-tracer software tool which supports the effective visualization of cosmological simulations data. The algorithm it relies on is designed to deal with point-like data, optimizing the ray-tracing calculation by ordering the particles as a function of their 'depth', defined as a function of one of the coordinates or other associated parameters. Realistic three-dimensional impressions are reached through a composition of the final colour in each pixel properly calculating emission and absorption of individual volume elements.
spops is a database of populations synthesis simulations of spinning black-hole binary systems, together with a python module to query it. Data are obtained with the startrack and precession [ascl:1611.004] numerical codes to consistently evolve binary stars from formation to gravitational-wave detection. spops allows quick exploration of the interplay between stellar physics and black-hole spin dynamics.
SPOTROD is a model for planetary transits of stars with an arbitrary limb darkening law and a number of homogeneous, circular spots on their surface. It facilitates analysis of anomalies due to starspot eclipses, and is a free, open source implementation written in C with a Python API.
SPRITE (Sparse Recovery of InstrumenTal rEsponse) computes a well-resolved compact source image from several undersampled and noisy observations. The algorithm is based on sparse regularization; adding a sparse penalty in the recovery leads to far better accuracy in terms of ellipticity error, especially at low S/N.
The presence of human-made interference mimicking the behavior of celestial radio pulses is a major challenge when searching for radio pulses emitted on millisecond timescales by celestial radio sources such as pulsars and fast radio bursts due to the highly imbalanced samples. Single-pulse Searcher (SpS) reduces the presence of radio interference when processing standard output from radio single-pulse searches to produce diagnostic plots useful for selecting good candidates. The modular software allows modifications for specific search characteristics. LOTAAS Single-pulse Searcher (L-SpS) is an implementation of different features of the software (such as a machine-learning approach) developed for a particular study: the LOFAR Tied-Array All-Sky Survey (LOTAAS).
The SPS software simulates the operation of the Spectral and Photometric Imaging Receiver on-board the ESA’s Herschel Space Observatory. It is coded using the Interactive Data Language (IDL), and produces simulated data at the level-0 stage (non-calibrated data in digitised units). The primary uses for the simulator are to:
The SPT lensing likelihood code, written in Fortran90, performs a Gaussian likelihood based upon the lensing potential power spectrum using a file from CAMB (ascl:1102.026) which contains the normalization required to get the power spectrum that the likelihood call is expecting.
SPTCLASS assigns semi-automatic spectral types to a sample of stars. The main code includes three spectral classification schemes: the first one is optimized to classify stars in the mass range of TTS (K5 or later, hereafter LATE-type scheme); the second one is optimized to classify stars in the mass range of IMTTS (F late to K early, hereafter Gtype scheme), and the third one is optimized to classify stars in the mass range of HAeBe (F5 or earlier, hereafter HAeBe scheme). SPTCLASS has an interactive module that allows the user to select the best result from the three schemes and analyze the input spectra.
SSE is a rapid single-star evolution (SSE) code; these analytical formulae cover all phases of evolution from the zero-age main-sequence up to and including remnant phases. It is valid for masses in the range 0.1-100 Msun and metallicity can be varied. The SSE package contains a prescription for mass loss by stellar winds. It also follows the evolution of rotational angular momentum for the star.
sslf is a simple, effective and useful spectral line finder for 1D data. It utilizes the continuous wavelet transform from SciPy, which is a productive way to find even weak spectral lines.
SSMM (Slotted Symbolic Markov Modeling) reduces time-domain stellar variable observations to classify stellar variables. The method can be applied to both folded and unfolded data, and does not require time-warping for waveform alignment. Written in Matlab, the performance of the supervised classification code is quantifiable and consistent, and the rate at which new data is processed is dependent only on the computational processing power available.
The ssos pipeline detects and identifies known and unknown Solar System Objects (SSOs) in astronomical images. ssos requires at least 3 images with overlapping field-of-views in the sky taken within a reasonable amount of time (e.g., 2 hours, 1 night). SSOs are detected mainly by judging the apparent motion of all sources in the images. The pipeline serves as a wrapper for the SExtractor (ascl:1010.064) and SCAMP (ascl:1010.063) software suites and allows different source extraction strategies to be chosen. All sources in the images are subject to a highly configurable filter pipeline. ssos is a versatile, light-weight, and easy-to-use software for surveys or PI-observation campaigns lacking a dedicated SSO detection pipeline.
SSSpaNG is a data-driven Gaussian Process model of the spectra of APOGEE red clump stars, whose parameters are inferred using Gibbs sampling. By pooling information between stars to infer their covariance it permits clear identification of the correlations between spectral pixels. Harnessing this correlation structure, a complete spectrum for each red clump star can be inferred, inpainting missing regions and de-noising by a factor of at least 2-3 for low-signal-to-noise stars.
STACKER stacks sources in interferometric data, i.e., averaging emission from different sources. The library allows stacking to be done directly on visibility data as well as in the image domain. The code is in format of a CASA (ascl:1107.013) task and implements uv- and image-stacking algorithms; it also provides several useful tasks for stacking related data processing. It allows introduction and stacking of random sources to estimate bias and noise, and also allows removal of a model of bright sources from the data.
Stagger is an astrophysical MHD code actively used to model star formation. It is equipped with a multi-frequency radiative transfer module and a comprehensive equation of state module that includes a large number of atomic and molecular species, to be able to compute realistic 3-D models of the near-surface layers of stars. The current version of the code allows a discretization that explicitly conserves mass, momentum, energy, and magnetic flux. The tensor formulation of the viscosity ensures that the viscous force is insensitive to the coordinate system orientation, thereby avoiding artificial grid-alignment.
Stan facilitates statistical inference at the frontiers of applied statistics and provides both a modeling language for specifying complex statistical models and a library of statistical algorithms for computing inferences with those models. These components are exposed through interfaces in environments such as R, Python, and the command line.
Database management is an increasingly important part of astronomical data analysis. Astronomers need easy and convenient ways of storing, editing, filtering, and retrieving data about data. Commercial databases do not provide good solutions for many of the everyday and informal types of database access astronomers need. The Starbase database system with simple data file formatting rules and command line data operators has been created to answer this need. The system includes a complete set of relational and set operators, fast search/index and sorting operators, and many formatting and I/O operators. Special features are included to enhance the usefulness of the database when manipulating astronomical data. The software runs under UNIX, MSDOS and IRAF.
STARBLADE (STar and Artefact Removal with a Bayesian Lightweight Algorithm from Diffuse Emission) separates superimposed point-like sources from a diffuse background by imposing physically motivated models as prior knowledge. The algorithm can also be used on noisy and convolved data, though performing a proper reconstruction including a deconvolution prior to the application of the algorithm is advised; the algorithm could also be used within a denoising imaging method. STARBLADE learns the correlation structure of the diffuse emission and takes it into account to determine the occurrence and strength of a superimposed point source.
Starburst99 is a comprehensive set of model predictions for spectrophotometric and related properties of galaxies with active star formation. The models are presented in a homogeneous way for five metallicities between Z = 0.040 and 0.001 and three choices of the initial mass function. The age coverage is 10^6 to 10^9 yr. Spectral energy distributions are used to compute colors and other quantities.
StarburstPy is a python wrapper for Starburst99 (ascl:1104.003). The code contains methods for setting all inputs, running Starburst99, and reading output data into python dictionaries.
StarcNet (STAR Cluster classification NETwork) classifies star clusters from galaxy images taken by the Hubble Space Telescope (HST); it uses a convolutional neural network (CNN) trained to classify five-band galaxy images into four morphological classes. Written in PyTorch, StarcNet runs using mosaics (.fits files with the galaxy photometric information) and catalogs (.tab files with object coordinates), and includes the option to also download the galaxy mosaics from a single .tar.gz file per galaxy (as from the Legacy ExtraGalactic UV Survey).
StarCrash is a parallel fortran code based on Smoothed Particle Hydrodynamics (SPH) techniques to calculate the 3-d evolution of self-gravitating fluid systems. The code in particularly suited to the study of stellar interactions, such as mergers of binary star systems and stellar collisions. The StarCrash code comes with several important features, including:
stardate measures precise stellar ages by combining isochrone fitting with gyrochronology (rotation-based ages) to increase the precision of stellar ages on the main sequence. The best possible ages provided by stardate will be for stars with rotation periods, though ages can also be predicted for stars without rotation periods. stardate is an extension to isochrones that incorporates gyrochronology and the code reverts back to isochrones when no rotation period is provided.
StarFinder is an IDL code for the deep analysis of stellar fields, designed for Adaptive Optics well-sampled images with high and low Strehl ratio. The Point Spread Function is extracted directly from the frame, to take into account the actual structure of the instrumental response and the atmospheric effects. The code is written in IDL language and organized in the form of a self-contained widget-based application, provided with a series of tools for data visualization and analysis. A description of the method and some applications to Adaptive Optics data are presented.
StarFISH is a suite of programs designed to determine the star formation history (SFH) of a stellar population, given multicolor stellar photometry and a library of theoretical isochrones. It constructs a library of synthetic color-magnitude diagrams from the isochrones, which includes the effects of extinction, photometric errors and completeness, and binarity. A minimization routine is then used to determine the linear combination of synthetic CMDs that best matches the observed photometry. The set of amplitudes modulating each synthetic CMD describes the star formation history of the observed stellar population.
Starfish is a set of tools used for spectroscopic inference. It robustly determines stellar parameters using high resolution spectral models and uses Markov Chain Monte Carlo (MCMC) to explore the full posterior probability distribution of the stellar parameters. Additional potential applications include other types of spectra, such as unresolved stellar clusters or supernovae spectra.
Traditionally, a simulation of a dense stellar system required choosing an initial model, running an integrator, and analyzing the output. Almost all of the effort went into writing a clever integrator that could handle binaries, triples and encounters between various multiple systems efficiently. Recently, the scope and complexity of these simulations has increased dramatically, for three reasons: 1) the sheer size of the data sets, measured in Terabytes, make traditional 'awking and grepping' of a single output file impractical; 2) the addition of stellar evolution data brings qualitatively new challenges to the data reduction; 3) increased realism of the simulations invites realistic forms of 'SOS': Simulations of Observations of Simulations, to be compared directly with observations. We are now witnessing a shift toward the construction of archives as well as tailored forms of visualization including the use of virtual reality simulators and planetarium domes, and a coupling of both with budding efforts in constructing virtual observatories. This review describes these new trends, presenting Starlab as the first example of a full software environment for realistic large-scale simulations of dense stellar systems.
The study of stellar populations in galaxies is entering a new era with the availability of large and high quality databases of both observed galactic spectra and state-of-the-art evolutionary synthesis models. The power of spectral synthesis can be investigated as a mean to estimate physical properties of galaxies. Spectral synthesis is nothing more than the decomposition of an observed spectrum in terms of a superposition of a base of simple stellar populations of various ages and metallicities, producing astrophysically interesting output such as the star-formation and chemical enrichment histories of a galaxy, its extinction and velocity dispersion. This is what the STARLIGHT spectral synthesis code does.
Starlink Figaro is an independently-maintained fork of Figaro (ascl:1203.013) that runs in the Starlink software environment (ascl:1110.012). It is a general-purpose data reduction package targeted mainly at optical/IR spectroscopy. It uses the NDF data format and the ADAM libraries for parameters and messaging.
Starlink has many applications within it to meet a variety of needs; it includes:
STARMAN is a stellar photometry package designed for the reduction of data from imaging systems. Its main components are crowded-field photometry programs, aperture photometry programs, a star finding program, and a CCD reduction program.
Image and table handling are served by a large number of programs which have a general use in photometry and other types of work. The package is a coherent whole, for use in the entire process of stellar photometry from raw images to the final standard-system magnitudes and their plotting as color-magnitude and color-color diagrams. It was distributed as part of the Starlink software collection (ascl:1110.012).
StarPy derives the quenching star formation history (SFH) of a single galaxy through the Bayesian Markov Chain Monte Carlo method code emcee (ascl:1303.002). The sample function implements the emcee EnsembleSampler function for the galaxy colors input. Burn-in is run and calculated for the length specified before the sampler is reset and then run for the length of steps specified. StarPy provides the ability to use the look-up tables provided or creating your own.
STARRY computes light curves for various applications in astronomy: transits and secondary eclipses of exoplanets, light curves of eclipsing binaries, rotational phase curves of exoplanets, light curves of planet-planet and planet-moon occultations, and more. By modeling celestial body surface maps as sums of spherical harmonics, STARRY does all this analytically and is therefore fast, stable, and differentiable. Coded in C++ but wrapped in Python, STARRY is easy to install and use.
We have developed a detailed stellar evolution code capable of following the simultaneous evolution of both stars in a binary system, together with their orbital properties. To demonstrate the capabilities of the code we investigate potential progenitors for the Type IIb supernova 1993J, which is believed to have been an interacting binary system prior to its primary exploding. We use our detailed binary stellar evolution code to model this system to determine the possible range of primary and secondary masses that could have produced the observed characteristics of this system, with particular reference to the secondary. Using the luminosities and temperatures for both stars (as determined by Maund et al. 2004) and the remaining mass of the hydrogen envelope of the primary at the time of explosion, we find that if mass transfer is 100 per cent efficient the observations can be reproduced by a system consisting of a 15 solar mass primary and a 14 solar mass secondary in an orbit with an initial period of 2100 days. With a mass transfer efficiency of 50 per cent, a more massive system consisting of a 17 solar mass primary and a 16 solar mass secondary in an initial orbit of 2360 days is needed. We also investigate some of the uncertainties in the evolution, including the effects of tidal interaction, convective overshooting and thermohaline mixing.
The 3D grid-based Monte Carlo code STaRS (Sejong Radiative Transfer through Raman and Rayleigh Scattering with atomic hydrogen) traces radiative transfer through Raman and Rayleigh scattering. This can be used to investigate line formation of Raman-scattered features in a thick neutral region illuminated by a strong far-UV emission source. Favorable conditions for Raman scattering with atomic hydrogen are easily met in symbiotic stars, young planetary nebulae, and active galactic nuclei.
The Matlab starsense_algorithms package evaluates the performance of various star sensors through the implementation of centroiding, geometric voting and QUEST algorithms. The physical parameters of a star sensor are parametrized and by changing these parameters, performance estimators such as sky coverage, memory requirement, and timing requirements can be estimated for the selected star sensor.
Smoothed Particle Hydrodynamics (SPH) is a Lagrangian particle method that approximates a continuous fluid as discrete nodes, each carrying various parameters such as mass, position, velocity, pressure, and temperature. In an SPH simulation the resolution scales with the particle density; StarSmasher is able to handle both equal-mass and equal number-density particle models. StarSmasher solves for hydro forces by calculating the pressure for each particle as a function of the particle's properties - density, internal energy, and internal properties (e.g. temperature and mean molecular weight). The code implements variational equations of motion and libraries to calculate the gravitational forces between particles using direct summation on NVIDIA graphics cards. Using a direct summation instead of a tree-based algorithm for gravity increases the accuracy of the gravity calculations at the cost of speed. The code uses a cubic spline for the smoothing kernel and an artificial viscosity prescription coupled with a Balsara Switch to prevent unphysical interparticle penetration. The code also implements an artificial relaxation force to the equations of motion to add a drag term to the calculated accelerations during relaxation integrations. Initially called StarCrash, StarSmasher was developed originally by Rasio.
STATCONT determines the continuum emission level in line-rich spectral data by inspecting the intensity distribution of a given spectrum by using different statistical approaches. The sigma-clipping algorithm provides the most accurate continuum level determination, together with information on the uncertainty in its determination; this uncertainty is used to correct the final continuum emission level. In general, STATCONT obtains accuracies of < 10 % in the continuum determination, and < 5 % in most cases. The main products of the software are the continuum emission level, together with its uncertainty, and data cubes containing only spectral line emission, i.e. continuum-subtracted data cubes. STATCONT also includes the option to estimate the spectral index or variation of the continuum emission with frequency.
statpl estimates the parameter of power-law distributed data and calculates goodness-of-fit tests for them. Many objects studied in astronomy follow a power-law distribution function (DF), for example the masses of stars or star clusters. Such data is often analyzed by generating a histogram and fitting a straight line to it. The parameters obtained in this way can be severely biased, and the properties of the underlying DF, such as its shape or a possible upper limit, are difficult to extract. statpl is an (effectively) bias-free estimator for the exponent and the upper limit.
STECKMAP stands for STEllar Content and Kinematics via Maximum A Posteriori likelihood. It is a tool for interpreting galaxy spectra in terms of their stellar populations through the derivation of their star formation history, age-metallicity relation, kinematics and extinction. The observed spectrum is projected onto a temporal sequence of models of single stellar populations, so as to determine a linear combination of these models that best fits the observed spectrum. The weights of the various components of this linear combination indicate the stellar content of the population. This procedure is regularized using various penalizing functions. The principles of the method are detailed in Ocvirk et al. 2006.
STELLA is a one-dimensional multi-group radiation hydrodynamics code. STELLA incorporates implicit hydrodynamics coupled to a multi-group non-equilibrium radiative transfer for modeling SN II-L light curves. The non-equilibrium description of radiation is crucial for this problem since the presupernova envelope may be of low mass and very dilute. STELLA implicitly treats time dependent equations of the angular moments of intensity averaged over a frequency bin. Local thermodynamic equilibrium is assumed to determine the ionization levels of materials.
stella creates and trains a neural network to identify stellar flares. Within stella, users can simulate flares as a training set, run a neural network, and feed in their own data to the neural network model. The software returns a probability at each data point as to whether that data point is part of a flare; the code can also characterize the flares identified.
stellaR accesses and manipulates publicly available stellar evolutionary tracks and isochrones from the Pisa low-mass database. It retrieves and plots the required calculations from CDS, constructs by interpolation tracks or isochrones of compositions different to the ones available in the database, constructs isochrones for age not included in the database, and extracts relevant evolutionary points from tracks or isochrones.
Cosmic ray electrons scatter on the photon fields around stars, including the sun, to create gamma rays by the inverse Compton effect. Stellarics computes the spectrum and angular distribution of this emission. The software also includes general-purpose routines for inverse Compton scattering on a given electron spectrum, for example for interstellar or astrophysical source modelling.
stellarWakes uses stellar kinematic data to search for dark matter (DM) subhalos through their gravitational perturbations to the stellar phase-space distribution.
StePar computes the stellar atmospheric parameters Teff, log g, [Fe/H], and ξ of FGK-type stars using the Equivalent Width (EW) method. The code implements a grid of MARCS model atmospheres and uses the MOOG radiative transfer code (ascl:1202.009) and TAME (ascl:1503.003). StePar uses a Downhill Simplex minimization algorithm, running it twice for any given star, to compute the stellar atmospheric parameters.
stepped_luneburg investigates the scattered light properties of a Luneburg lens approximated as a series of concentric shells with discrete refractive indices. The optical Luneburg lens has promising applications for low-cost, continuous all-sky monitoring to obtain transit light curves of bright, nearby stars. This code implements a stack-based algorithm that tracks all reflected and refracted rays generated at each optical interface of the lens as described by Snell's law. The Luneburg lens model parameters, such as number of lens layers, the power-law that describes the refractive indices, the number of incident rays, and the initial direction of the incident wavefront can be altered to optimize lens performance. The stepped_luneburg module can be imported within the Python environment or used with scripting, and it is accompanied by two other modules, enc_int and int_map, that help the user to determine the resolving power of the lens and the strength of scattered light haloes for the purpose of quality assessment.
StePS (Stereographically Projected Cosmological Simulations) compactifies the infinite spatial extent of the Universe into a finite sphere with isotropic boundary conditions to simulate the evolution of the large-scale structure. This eliminates the need for periodic boundary conditions, which are a numerical convenience unsupported by observation and which modifies the law of force on large scales in an unrealistic fashion. StePS uses stereographic projection for space compactification and naive O(N2) force calculation; this arrives at a correlation function of the same quality more quickly than standard (tree or P3M) algorithms with similar spatial and mass resolution. The N2 force calculation is easy to adapt to modern graphics cards, hence StePS can function as a high-speed prediction tool for modern large-scale surveys.
Would you like to view a random code?