Results 1901-1950 of 2331 (2295 ASCL, 36 submitted)
SkyMaker is a program that simulates astronomical images. It accepts object lists in ASCII generated by the Stuff program to produce realistic astronomical fields. SkyMaker is part of the EFIGI development project.
The general-purpose nuclear reaction network SkyNet evolves the abundances of nuclear species under the influence of nuclear reactions. SkyNet can be used to compute the nucleosynthesis evolution in all astrophysical scenarios where nucleosynthesis occurs. Any list of isotopes can be evolved and SkyNet supports various different types of nuclear reactions. SkyNet is modular, permitting new or existing physics, such as nuclear reactions or equations of state, to be easily added or modified.
SkyNet is an efficient and robust neural network training code for machine learning. It is able to train large and deep feed-forward neural networks, including autoencoders, for use in a wide range of supervised and unsupervised learning applications, such as regression, classification, density estimation, clustering and dimensionality reduction. SkyNet is implemented in C/C++ and fully parallelized using MPI.
The SkyView Virtual telescope provides access to survey datasets ranging from radio through the gamma-ray regimes. Over 100 survey datasets are currently available. The SkyView library referenced here is used as the basis for the SkyView web site (at http://skvyiew.gsfc.nasa.gov) but is designed for individual use by researchers as well.
SkyView's approach to access surveys is distinct from most other toolkits. Rather than providing links to the original data, SkyView attempts to immediately re-render the source data in the user-requested reference frame, projection, scaling, orientation, etc. The library includes a set of geometry transformation and mosaicking tools that may be integrated into other applications independent of SkyView.
SL1M deconvolves radio synthesis images based on direct inversion of the measured visibilities that can deal with the non-coplanar base line effect and can be applied to telescopes with direction dependent gains. The code is more computationally demanding than some existing methods, but is highly parallelizable and scale well to clusters of CPUs and GPUs. The algorithm is also extremely flexible, allowing the solution of the deconvolution problem on arbitrarily placed pixels.
SLALIB is a library of routines that make accurate and reliable positional-astronomy applications easier to write. Most SLALIB routines are concerned with astronomical position and time, but a number have wider trigonometrical, numerical or general applications. A Fortran implementation of SLALIB under GPL licensing is available as part of Starlink (ascl:1110.012).
SlicerAstro extends 3D Slicer, a multi-platform package for visualization and medical image processing, to provide a 3-D interactive viewer with 3-D human-machine interaction features, based on traditional 2-D input/output hardware, and analysis capabilities.
The semi-spectral linear MHD (SLiM) code follows the interaction of linear waves through an inhomogeneous three-dimensional solar atmosphere. The background model allows almost arbitrary perturbations of density, temperature, sound speed as well as magnetic and velocity fields. The code is useful in understanding the helioseismic signatures of various solar features, including sunspots.
Slim performs lossless compression on binary data files. Written in C++, it operates very rapidly and achieves better compression on noisy physics data than general-purpose tools designed primarily for text.
slimplectic is a python implementation of a numerical integrator that uses a fixed time-step variational integrator formalism applied to the principle of stationary nonconservative action. It allows nonconservative effects to be included in the numerical evolution while preserving the major benefits of normally conservative symplectic integrators, particularly the accurate long-term evolution of momenta and energy. slimplectic is appropriate for exploring cosmological or celestial N-body dynamics problems where nonconservative interactions, e.g. dynamical friction or dissipative tides, can play an important role.
SLOPES computes six least-squares linear regression lines for bivariate datasets of the form (x_i,y_i) with unknown population distributions. Measurement errors, censoring (nondetections) or other complications are not treated. The lines are: the ordinary least-squares regression of y on x, OLS(Y|X); the inverse regression of x on y, OLS(X_Y); the angular bisector of the OLS lines; the orthogonal regression line; the reduced major axis, and the mean-OLS line. The latter four regressions treat the variables symmetrically, while the first two regressions are asymmetrical. Uncertainties for the regression coefficients of each method are estimated via asymptotic formulae, bootstrap resampling, and bivariate normal simulation. These methods, derivation of the regression coefficient uncertainties, and discussions of their use are provided in three papers listed below. The user is encouraged to read and reference these studies.
Stellar Locus Regression (SLR) is a simple way to calibrate colors at the 1-2% level, and magnitudes at the sub-5% level as limited by 2MASS, without the traditional use of standard stars. With SLR, stars in any field are "standards." This is an entirely new way to calibrate photometry. SLR exploits the simple fact that most stars lie along a well defined line in color-color space called the stellar locus. Cross-match point-sources in flattened images taken through different passbands and plot up all color vs color combinations, and you will see the stellar locus with little effort. SLR calibrates colors by fitting these colors to a standard line. Cross-match with 2MASS on top of that, and SLR will deliver calibrated magnitudes as well.
The effects of stochasticity on the luminosities of stellar populations are an often neglected but crucial element for understanding populations in the low mass or low star formation rate regime. To address this issue, we present SLUG, a new code to "Stochastically Light Up Galaxies". SLUG synthesizes stellar populations using a Monte Carlo technique that treats stochastic sampling properly including the effects of clustering, the stellar initial mass function, star formation history, stellar evolution, and cluster disruption. This code produces many useful outputs, including i) catalogs of star clusters and their properties, such as their stellar initial mass distributions and their photometric properties in a variety of filters, ii) two dimensional histograms of color-magnitude diagrams of every star in the simulation, iii) and the photometric properties of field stars and the integrated photometry of the entire simulated galaxy. After presenting the SLUG algorithm in detail, we validate the code through comparisons with starburst99 in the well-sampled regime, and with observed photometry of Milky Way clusters. Finally, we demonstrate the SLUG's capabilities by presenting outputs in the stochastic regime.
SMART is an IDL-based software tool, developed by the IRS Instrument Team at Cornell University, that allows users to reduce and analyze Spitzer data from all four modules of the Infrared Spectrograph, including the peak-up arrays. The software is designed to make full use of the ancillary files generated in the Spitzer Science Center pipeline so that it can either remove or flag artifacts and corrupted data and maximize the signal-to-noise ratio in the extraction routines. It can be run in both interactive and batch modes. SMART includes visualization tools for assessing data quality, basic arithmetic operations for either two-dimensional images or one-dimensional spectra, extraction of both point and extended sources, and a suite of spectral analysis tools.
SMARTIES calculates the optical properties of oblate and prolate spheroidal particles, with comparable capabilities and ease-of-use as Mie theory for spheres. This suite of MATLAB codes provides a fully documented implementation of an improved T-matrix algorithm for the theoretical modelling of electromagnetic scattering by particles of spheroidal shape. Included are scripts that cover a range of scattering problems relevant to nanophotonics and plasmonics, including calculation of far-field scattering and absorption cross-sections for fixed incidence orientation, orientation-averaged cross-sections and scattering matrix, surface-field calculations as well as near-fields, wavelength-dependent near-field and far-field properties, and access to lower-level functions implementing the T-matrix calculations, including the T-matrix elements which may be calculated more accurately than with competing codes.
Spectroscopy Made Easy (SME) is IDL software and a compiled external library that fits an observed high-resolution stellar spectrum with a synthetic spectrum to determine stellar parameters. The SME external library is available for Mac, Linux, and Windows systems. Atomic and molecular line data formatted for SME may be obtained from VALD. SME can solve for empirical log(gf) and damping parameters, using an observed spectrum of a star (usually the Sun) as a constraint.
SMERFS (Stochastic Markov Evaluation of Random Fields on the Sphere) creates large realizations of random fields on the sphere. It uses a fast algorithm based on Markov properties and fast Fourier Transforms in 1d that generates samples on an n X n grid in O(n2 log n) and efficiently derives the necessary conditional covariance matrices.
SMILE is interactive software for studying a variety of 2D and 3D models, including arbitrary potentials represented by a basis-set expansion, a spherical-harmonic expansion with coefficients being smooth functions of radius (splines), or a set of fixed point masses. Its main features include:
SMILI uses sparse sampling techniques and other regularization methods for interferometric imaging. The python-interfaced library is mainly designed for very long baseline interferometry, and has been under the active development primarily for the Event Horizon Telescope (EHT).
SMMOL (Spherical Multi-level MOLecular line radiative transfer) is a molecular line radiative transfer code that uses Accelerated Lambda Iteration to solve the coupled level population and line transfer problem in spherical geometry. The code uses a discretized grid and a ray tracing methodology. SMMOL is designed for high optical depth regimes and can cope with maser emission as long as the spatial-velocity sampling is fine enough.
SMURF reduces submillimeter single-dish continuum and heterodyne data. It is mainly targeted at data produced by the James Clerk Maxwell Telescope but data from other telescopes have been reduced using the package. SMURF is released as part of the bundle that comprises Starlink (ascl:1110.012) and most of the packages that use it. The two key commands are MAKEMAP for the creation of maps from sub millimeter continuum data and MAKECUBE for the creation of data cubes from heterodyne array instruments. The software can also convert data from legacy JCMT file formats to the modern form to allow it to be processed by MAKECUBE. SMURF is a core component of the ORAC-DR (ascl:1310.001) data reduction pipeline for JCMT.
SNANA is a general analysis package for supernova (SN) light curves that contains a simulation, light curve fitter, and cosmology fitter. The software is designed with the primary goal of using SNe Ia as distance indicators for the determination of cosmological parameters, but it can also be used to study efficiencies for analyses of SN rates, estimate contamination from non-Ia SNe, and optimize future surveys. Several SN models are available within the same software architecture, allowing technical features such as K-corrections to be consistently used among multiple models, and thus making it easier to make detailed comparisons between models. New and improved light-curve models can be easily added. The software works with arbitrary surveys and telescopes and has already been used by several collaborations, leading to more robust and easy-to-use code. This software is not intended as a final product release, but rather it is designed to undergo continual improvements from the community as more is learned about SNe.
SNAPDRAGONS (Stellar Numbers And Parameters Determined Routinely And Generated Observing N-body Systems) is a simplified version of the population synthesis code Galaxia (ascl:1101.007), using a different process to generate the stellar catalog. It splits each N-body particle from the galaxy simulation into an appropriate number of stellar particles to create a mock catalog of observable stars from the N-body model. SNAPDRAGON uses the same isochrones and extinction map as Galaxia.
SNCosmo synthesizes supernova spectra and photometry from SN models, and has functions for fitting and sampling SN model parameters given photometric light curve data. It offers fast implementations of several commonly used extinction laws and can be used to construct SN models that include dust. The SNCosmo library includes supernova models such as SALT2, MLCS2k2, Hsiao, Nugent, PSNID, SNANA and Whalen models, as well as a variety of built-in bandpasses and magnitude systems, and provides convenience functions for reading and writing peculiar data formats used in other packages. The library is extensible, allowing new models, bandpasses, and magnitude systems to be defined using an object-oriented interface.
SNEC (SuperNova Explosion Code) is a spherically-symmetric Lagrangian radiation-hydrodynamics code that follows supernova explosions through the envelope of their progenitor star, produces bolometric (and approximate multi-color) light curve predictions, and provides input to spectral synthesis codes for spectral modeling. SNEC's features include 1D (spherical) Lagrangian Newtonian hydrodynamics with artificial viscosity, stellar equation of state with a Saha solver ionization/recombination, equilibrium flux-limited photon diffusion with OPAL opacities and low-temperature opacities, and prediction of bolometric light curves and multi-color lightcurves (in the blackbody approximation).
We present an algorithm to identify the type of an SN spectrum and to determine its redshift and age. This algorithm, based on the correlation techniques of Tonry & Davis, is implemented in the Supernova Identification (SNID) code. It is used by members of ongoing high-redshift SN searches to distinguish between type Ia and type Ib/c SNe, and to identify "peculiar" SNe Ia. We develop a diagnostic to quantify the quality of a correlation between the input and template spectra, which enables a formal evaluation of the associated redshift error. Furthermore, by comparing the correlation redshifts obtained using SNID with those determined from narrow lines in the SN host galaxy spectrum, we show that accurate redshifts (with a typical error less than 0.01) can be determined for SNe Ia without a spectrum of the host galaxy. Last, the age of an input spectrum is determined with a typical 3-day accuracy, shown here by using high-redshift SNe Ia with well-sampled light curves. The success of the correlation technique confirms the similarity of some SNe Ia at low and high redshifts. The SNID code, which is available to the community, can also be used for comparative studies of SN spectra, as well as comparisons between data and models.
Snoopy is a spectral 3D code that solves the MHD and Boussinesq equations, such as compressibility, particles, and Braginskii viscosity, and several other physical effects. It's useful for turbulence study involving shear and rotation. Snoopy requires the FFTW library (ascl:1201.015), and can run on parallel machine using MPI OpenMP or both at the same time.
The SNooPy package (also known as SNpy), written in Python, contains tools for the analysis of TypeIa supernovae. It offers interactive plotting of light-curve data and models (and spectra), computation of reddening laws and K-corrections, LM non-linear least-squares fitting of light-curve data, and various types of spline fitting, including Diercx and tension. The package also includes a SNIa lightcurve template generator in the CSP passbands, estimates of Milky-Way Extinction, and a module for dealing with filters and spectra.
SNRPy (Super Nova Remnant Python) models supernova remnant (SNR) evolution and is useful for understanding SNR evolution and to model observations of SNR for obtaining good estimates of SNR properties. It includes all phases for the standard path of evolution for spherically symmetric SNRs and includes alternate evolutionary models, including evolution in a cloudy ISM, the fractional energy loss model, and evolution in a hot low-density ISM. The graphical interface takes in various parameters and produces outputs such as shock radius and velocity vs. time, SNR surface brightness profile and spectrum.
SNSEDextend extrapolates core-collapse and Type Ia Spectral Energy Distributions (SEDs) into the UV and IR for use in simulations and photometric classifications. The user provides a library of existing SED templates (such as those in the authors' SN SED Repository) along with new photometric constraints in the UV and/or NIR wavelength ranges. The software then extends the existing template SEDs so their colors match the input data at all phases. SNSEDextend can also extend the SALT2 spectral time-series model for Type Ia SN for a "first-order" extrapolation of the SALT2 model components, suitable for use in survey simulations and photometric classification tools; as the code does not do a rigorous re-training of the SALT2 model, the results should not be relied on for precision applications such as light curve fitting for cosmology.
Supernova Time Delays (SNTD) simulates and measures time delay of multiply-imaged supernovae, and offers an improved characterization of the uncertainty caused by microlensing. Lensing time delays can be determined by fitting the multiple light curves of these objects; measuring these delays provide precise tests of lens models or constraints on the Hubble constant and other cosmological parameters that are independent of the local distance ladder. Fitting the effects of microlensing without an accurate prior often leads to biases in the time delay measurement and over-fitting to the data; this can be mitigated by using a Gaussian Process Regression (GPR) technique to determine the uncertainty due to microlensing. SNTD can produce accurate simulations for wide-field time domain surveys such as LSST and WFIRST.
SOAP (Spot Oscillation And Planet) 2.0 simulates the effects of dark spots and bright plages on the surface of a rotating star, computing their expected radial velocity and photometric signatures. It includes the convective blueshift and its inhibition in active regions.
SOFA (Standards Of Fundamental Astronomy) is a collection of subprograms, in source-code form, that implement official IAU algorithms for fundamental astronomy computations. SOFA offers more than 160 routines for fundamental astronomy, including time scales (including dealing with leap seconds), Earth rotation, sidereal time, precession, nutation, polar motion, astrometry and transforms between various reference systems (e.g. BCRS, ICRS, GCRS, CIRS, TIRS, ITRS). The subprograms are supported by 55 vector/matrix routines, and are available in both Fortran77 and C implementations.
SoFiA is a flexible source finding pipeline designed to detect and parameterise sources in 3D spectral-line data cubes. SoFiA combines several powerful source finding and parameterisation algorithms, including wavelet denoising, spatial and spectral smoothing, source mask optimisation, spectral profile fitting, and calculation of the reliability of detections. In addition to source catalogues in different formats, SoFiA can also generate a range of output data cubes and images, including source masks, moment maps, sub-cubes, position-velocity diagrams, and integrated spectra. The pipeline is controlled by simple parameter files and can either be invoked on the command line or interactively through a modern graphical user interface.
SoFiAX is a web-based platform to merge and interact with the results of parallel execution of SoFiA HI source finding software [ascl:1412.001] and other steps of processing ASKAP Wallaby HI survey data.
Although large volumes of solar data are available for investigation and study, the vast majority of these data remain unlabeled and are therefore not amenable to modern supervised machine learning methods. Having a way to accurately and automatically classify spectra into categories related to the degree of solar activity is highly desirable and will assist and speed up future research efforts in solar physics. At the same time, the large volume of raw observational data is a serious bottleneck for machine learning, requiring powerful computational means that are not at the disposal of many laboratories. Additionally, the raw data communication imposes some restrictions on real time data observations and requires considerable bandwidth and energy for the onboard solar observation systems. To cope with the above mentioned issues, we propose a framework to classify solar activity on compressed data. To this end, we used a labeling scheme from a pre-existing vector quantization technique in conjunction with several machine learning algorithms to categorize Mg II spectra measured by NASA’s small explorer satellite IRIS into several groups characterizing solar activity. Our training data set is a human annotated list of 85 IRIS observations containing 29097 frames in total or equivalently 9 million Mg II spectra. The annotated types of Solar activity are: active region, pre-flare activity, Solar flare, Sunspot and quiet Sun. We used the vector quantization to compress these data and to reduce its complexity before training classifiers. From a host of classifiers, we found that the XGBoost classifier produced the most accurate results on the compressed data, yielding over a 95% prediction rate, and outperforming other ML methods like convolution neural networks, K-nearest neighbors, naive Bayes classifiers and support vector machines. A principle finding of this research is that the classification performance on compressed and uncompressed data is comparable under our particular architecture, implying the possibility of large compression rates for relatively low degrees of information loss.
SolarSoft is a set of integrated software libraries, data bases, and system utilities which provide a common programming and data analysis environment for Solar Physics. The SolarSoftWare (SSW) system is built from Yohkoh, SOHO, SDAC and Astronomy libraries and draws upon contributions from many members of those projects. It is primarily an IDL based system, although some instrument teams integrate executables written in other languages. The SSW environment provides a consistent look and feel at widely distributed co-investigator institutions to facilitate data exchange and to stimulate coordinated analysis. Commonalities and overlap in solar data and analysis goals are exploited to permit application of fundamental utilities to the data from many different solar instruments. The use of common libraries, utilities, techniques and interfaces minimizes the learning curve for investigators who are analyzing new solar data sets, correlating results from multiple experiments or performing research away from their home institution.
SONG computes the non-linear evolution of the Universe in order to predict cosmological observables such as the bispectrum of the Cosmic Microwave Background (CMB). More precisely, it is a second-order Boltzmann code, as it solves the Einstein and Boltzmann equations up to second order in the cosmological perturbations.
SOPHIA (Simulations Of Photo Hadronic Interactions in Astrophysics) solves problems connected to photohadronic processes in astrophysical environments and can also be used for radiation and background studies at high energy colliders such as LEP2 and HERA, as well as for simulations of photon induced air showers. SOPHIA implements well established phenomenological models, symmetries of hadronic interactions in a way that describes correctly the available exclusive and inclusive photohadronic cross section data obtained at fixed target and collider experiments.
SOPHISM models astronomical instrumentation from the entrance of the telescope to data acquisition at the detector, along with software blocks dealing with, for example, demodulation, inversion, and compression. The code performs most analyses done with light in astronomy, such as differential photometry, spectroscopy, and polarimetry. The simulator offers flexibility and implementation of new effects and subsystems, making it user-adaptable for a wide variety of instruments. SOPHISM can be used for all stages of instrument definition, design, operation, and lifetime tracking evaluation.
SOPIE (Sequential Off-Pulse Interval Estimation) provides functions to non-parametrically estimate the off-pulse interval of a source function originating from a pulsar. The technique is based on a sequential application of P-values obtained from goodness-of-fit tests for the uniform distribution, such as the Kolmogorov-Smirnov, Cramér-von Mises, Anderson-Darling and Rayleigh goodness-of-fit tests.
SOPT (Sparse OPTimisation) is a C implementation of the Sparsity Averaging Reweighted Analysis (SARA) algorithm. The approach relies on the observation that natural images exhibit strong average sparsity; average sparsity outperforms state-of-the-art priors that promote sparsity in a single orthonormal basis or redundant frame, or that promote gradient sparsity.
Spin-Orbit Tomography (SOT) is a retrieval technique of a two-dimensional map of an Exo-Earth from time-series data of integrated reflection light. The software provides code for the Bayesian version of the static SOT and dynamic mapping (time-varying mapping) with full Bayesian modeling, and tutorials for L2 and Bayesian SOT are available in jupyter notebooks.
SP_Ace (Stellar Parameters And Chemical abundances Estimator) estimates the stellar parameters Teff, log g, [M/H], and elemental abundances. It employs 1D stellar atmosphere models in Local Thermodynamic Equilibrium (LTE). The code is highly automated and suitable for analyzing the spectra of large spectroscopic surveys with low or medium spectral resolution (R = 2000-20 000). A web service for calculating these values with the software is also available.
The Solar Position Algorithm (SPA) calculates the solar zenith and azimuth angles in the period from the year -2000 to 6000, with uncertainties of +/- 0.0003 degrees based on the date, time, and location on Earth. SPA is implemented in C; in addition to being available for download, an online calculator using this code is available at http://www.nrel.gov/midc/solpos/spa.html.
SpacePy provides data analysis and visualization tools for the space science community. Written in Python, it builds on the capabilities of the NumPy and MatPlotLib packages to make basic data analysis, modeling and visualization easier. It contains modules for handling many complex time formats, obtaining data from the OMNI database, and accessing the powerful Onera library. It contains a library of commonly used empirical relationships, performs association analysis, coordinate transformations, radiation belt modeling, and CDF reading, and creates publication quality plots.
SpaghettiLens allows citizen scientists to model gravitational lenses collaboratively; the software should also be easily adaptable to any other, reasonably similar problem. It lets volunteers execute a computer intensive task that cannot be easily executed client side and relies on citizen scientists collaborating. SpaghettiLens makes survey data available to citizen scientists, manages the model configurations generated by the volunteers, stores the resulting model configuration, and delivers the actual model. A model can be shared and discussed with other volunteers and revised, and new child models can be created, resulting in a branching version tree of models that explore different possibilities. Scientists can choose a collection of models; discussion among volunteers and scientists prune the tree to determine which models will receive further analysis.
SPAM searches for imprints of Hu-Sawicki f(R) gravity on the rotation curves of the SPARC (Spitzer Photometry and Accurate Rotation Curves) sample using the MCMC sampler emcee (ascl:1303.002). The code provides attributes for inspecting the MCMC chains and translating names of parameters to indices. The SPAM package also contains plotting scripts.
SPAM is a extension to AIPS for reducing high-resolution, low-frequency radio interferometric observations. Direction-dependent ionospheric calibration and image-plane ripple suppression are among the features that help to make high-quality sub-GHz images. Data reductions are captured in well-tested Python scripts that execute AIPS tasks directly (mostly during initial data reduction steps), call high-level functions that make multiple AIPS or ParselTongue calls, and require few manual operations.
SPAMCART generates synthetic spectral energy distributions and intensity maps from smoothed particle hydrodynamics simulation snapshots. It follows discrete luminosity packets as they propagate through a density field, and computes the radiative equilibrium temperature of the ambient dust from their trajectories. The sources can be extended and/or embedded, and discrete and/or diffuse. The density is not mapped on to a grid, and therefore the calculation is performed at exactly the same resolution as the hydrodynamics. The code strictly adheres to Kirchhoff's law of radiation. The algorithm is based on the Lucy Monte Carlo radiative transfer method and is fairly simple to implement, as it uses data structures that are already constructed for other purposes in modern particle codes
Would you like to view a random code?