[ascl:1505.020]
rvfit: Radial velocity curves fitting for binary stars or exoplanets
rvfit, developed in IDL 7.0, fits non-precessing keplerian radial velocity (RV) curves for double-line and single-line binary stars or exoplanets. It fits a simple keplerian model to the observed RV and computes the seven parameters (six for a single-line system) from the model. Some parameters can be fixed beforehand if they are known, for instance, if photometric observations are available. The fit is done using an Adaptive Simulated Annealing algorithm optimized for this specific task. Simulated Annealing methods are powerful heuristic algorithms to minimize functions in multiparametric spaces.
[ascl:1210.031]
RVLIN: Fitting Keplerian curves to radial velocity data
The RVLIN package for IDL is a set of routines that quickly fits an arbitrary number of Keplerian curves to radial velocity data. It can handle data from multiple telescopes (i.e. it solves for the offset), constraints on P, e, and time of peri passage, and can incorporate transit timing data. The code handles fixed periods and circular orbits in combination and transit time constraints, including for multiple transiting planets.
[ascl:9912.003]
RVSAO 2.0: Digital Redshifts and Radial Velocities
RVSAO is a set of programs to obtain redshifts and radial velocities from digital spectra. RVSAO operates in the IRAF (Tody 1986, 1993) environment. The heart of the system is xcsao, which implements the cross-correlation method, and is a direct descendant of the system built by Tonry and Davis (1979). emsao uses intelligent heuristics to search for emission lines in spectra, then fits them to obtain a redshift. sumspec shifts and sums spectra to build templates for cross-correlation. linespec builds synthetic spectra given a list of spectral lines. bcvcorr corrects velocities for the motion of the earth. We discuss in detail the parameters necessary to run xcsao and emsao properly. We discuss the reliability and error associated with xcsao derived redshifts. We develop an internal error estimator, and we show how large, stable surveys can be used to develop more accurate error estimators. We develop a new methodology for building spectral templates for galaxy redshifts. We show how to obtain correlation velocities using emission line templates. Emission line correlations are substantially more efficient than the previous standard technique, automated emission line fitting. We compare the use of RVSAO with new methods, which use Singular Value Decomposition and $chi^2$ fitting techniques.
[ascl:2512.012]
RVSNUpy: Spectroscopic redshift measurement by cross-correlation
RVSNUpy measures spectroscopic redshifts using inverse-variance-weighted cross-correlation between an observed spectrum and a library of rest-frame template spectra. It ingests spectra (<i>e.g.</i>, SDSS FITS spectra and MMT/Hectospec raw spectra) and runs single-object or batch redshift measurements through a reusable measurement object initialized with a chosen template set. RVSNUpy reports redshift solutions across templates and identifies a best measurement within its returned results table. The package includes notebooks demonstrating typical workflows for redshift measurement and inspection of cross-correlation outputs.
[ascl:1907.013]
RVSpecFit: Radial velocity and stellar atmospheric parameter fitting
RVSpecFit determines radial velocities and stellar atmospheric parameters from spectra by direct pixel fitting by interpolated stellar templates. The code doesn't require spectrum normalization and can deal with non-flux calibrated spectra. RVSpecFit is able to fit multiple spectra simultaneously.
[ascl:2402.003]
Rwcs: World coordinate system transforms in R
Rwcs offers access to all the projection and distortion systems of WCSLIB (ascl:1108.003) in R. This can be used directly for, for example, pixel lookups, or for higher level general distortion and projection.
[ascl:2503.024]
S3Fit: Simultaneous Spectrum and photometric-SED Fitting code for galaxy observations
S<sup>3</sup>Fit fits spectrum and multi-band photometric Spectral Energy Distribution (SED) simultaneously for analyzing observational data of galaxies. It improves the moderate constraints on properties of continuum models in a pure spectral fitting due to the limited wavelength coverage. The code supports multiple models with multiple components, and can handle complex systems with a mixed contribution of Active Galactic Nucleus (AGN) and its host galaxy in both of continua and emission lines (<i>e.g.</i>, narrow lines and broad outflow lines). The fitting strategy is optimized to enable an efficient solution of the best-fit results for several tens of parameters and model components. S<sup>3</sup>Fit is also extensible for adding functions and components by users such as new band filters, star formation history functions, emission lines, and also types of models.
[ascl:1606.008]
s2: Object oriented wrapper for functions on the sphere
The s2 package can represent any arbitrary function defined on the sphere. Both real space map and harmonic space spherical harmonic representations are supported. Basic sky representations have been extended to simulate full sky noise distributions and Gaussian cosmic microwave background realisations. Support for the representation and convolution of beams is also provided. The code requires HEALPix (ascl:1107.018) and CFITSIO (ascl:1010.001).
[ascl:2404.027]
s2fft: Differentiable and accelerated spherical transforms
S2FFT computes Fourier transforms on the sphere and rotation group using JAX (ascl:2111.002) or PyTorch. It leverages autodiff to provide differentiable transforms, which are also deployable on hardware accelerators (<i>e.g.</i>, GPUs and TPUs). More specifically, S2FFT provides support for spin spherical harmonic and Wigner transforms (for both real and complex signals), with support for adjoint transformations where needed, and comes with different optimisations (precompute or not) that one may select depending on available resources and desired angular resolution <i>L</i>.
[ascl:1110.013]
S2HAT: Scalable Spherical Harmonic Transform Library
Many problems in astronomy and astrophysics require a computation of the spherical harmonic transforms. This is in particular the case whenever data to be analyzed are distributed over the sphere or a set of corresponding mock data sets has to be generated. In many of those contexts, rapidly improving resolutions of both the data and simulations puts increasingly bigger emphasis on our ability to calculate the transforms quickly and reliably.
The scalable spherical harmonic transform library S2HAT consists of a set of flexible, massively parallel, and scalable routines for calculating diverse (scalar, spin-weighted, etc) spherical harmonic transforms for a class of isolatitude sky grids or pixelizations. The library routines implement the standard algorithm with the complexity of O(n^3/2), where n is a number of pixels/grid points on the sphere, however, owing to their efficient parallelization and advanced numerical implementation, they achieve very competitive performance and near perfect scalability. S2HAT is written in Fortran 90 with a C interface. This software is a derivative of the spherical harmonic transforms included in the <a href="http://ascl.net/1107.018">HEALPix package</a> (ascl:1107.018) and is based on both serial and MPI routines of its version 2.01, however, since version 2.5 this software is fully autonomous of HEALPix and can be compiled and run without the HEALPix library.
[ascl:1211.001]
S2LET: Fast wavelet analysis on the sphere
S2LET provides high performance routines for fast wavelet analysis of signals on the sphere. It uses the SSHT code (ascl:2207.034) built on the MW sampling theorem to perform exact spherical harmonic transforms on the sphere. The resulting wavelet transform implemented in S2LET is theoretically exact, <span style="font-style: italic">i.e.</span> a band-limited signal can be recovered from its wavelet coefficients exactly and the wavelet coefficients capture all the information. S2LET also supports the HEALPix sampling scheme, in which case the transforms are not theoretically exact but achieve good numerical accuracy. The core routines of S2LET are written in C and have interfaces in Matlab, IDL and Java. Real signals can be written to and read from FITS files and plotted as Mollweide projections.
[ascl:1103.003]
S2PLOT: Three-dimensional (3D) Plotting Library
We present a new, three-dimensional (3D) plotting library with advanced features, and support for standard and enhanced display devices. The library - S2PLOT - is written in C and can be used by C, C++ and FORTRAN programs on GNU/Linux and Apple/OSX systems. S2PLOT draws objects in a 3D (x,y,z) Cartesian space and the user interactively controls how this space is rendered at run time. With a PGPLOT inspired interface, S2PLOT provides astronomers with elegant techniques for displaying and exploring 3D data sets directly from their program code, and the potential to use stereoscopic and dome display devices. The S2PLOT architecture supports dynamic geometry and can be used to plot time-evolving data sets, such as might be produced by simulation codes. In this paper, we introduce S2PLOT to the astronomical community, describe its potential applications, and present some example uses of the library.
[ascl:2005.009]
s3PCF: Compute the 3-point correlation function in the squeezed limit
s3PCF computes the 3-point correlation function (3PCF) in the squeezed limit given galaxy positions and pair positions. The code is currently written specifically for the Abacus simulations, but the main functionalities can be also easily adapted for other galaxy catalogs with the appropriate properties.
[ascl:2403.008]
s4cmb: Systematics For Cosmic Microwave Background
s4cmb (Systematics For Cosmic Microwave Background) studies the impact of instrumental systematic effects on measurements of CMB experiments based on bolometric detector technology. s4cmb provides a unified framework to simulate raw data streams in the time domain (TODs) acquired by CMB experiments scanning the sky, and to inject in these realistic instrumental systematics effect.
[ascl:1111.003]
Saada: A Generator of Astronomical Database
Saada transforms a set of heterogeneous FITS files or VOtables of various categories (images, tables, spectra, etc.) in a powerful database deployed on the Web. Databases are located on your host and stay independent of any external server. This job doesn’t require writing code. Saada can mix data of various categories in multiple collections. Data collections can be linked each to others making relevant browsing paths and allowing data-mining oriented queries. Saada supports 4 VO services (Spectra, images, sources and TAP) . Data collections can be published immediately after the deployment of the Web interface.
[ascl:1306.001]
SAC: Sheffield Advanced Code
The Sheffield Advanced Code (SAC) is a fully non-linear MHD code designed for simulations of linear and non-linear wave propagation in gravitationally strongly stratified magnetized plasma. It was developed primarily for the forward modelling of helioseismological processes and for the coupling processes in the solar interior, photosphere, and corona; it is built on the well-known VAC platform that allows robust simulation of the macroscopic processes in gravitationally stratified (non-)magnetized plasmas. The code has no limitations of simulation length in time imposed by complications originating from the upper boundary, nor does it require implementation of special procedures to treat the upper boundaries. SAC inherited its modular structure from VAC, thereby allowing modification to easily add new physics.
[submitted]
Sacc: Save All Correlations and Covariances
Zuntz, J.;
Slosar, A.;
Alonso, D.;
Becker, M.;
Broussard, A.;
McClintock, T.;
Nicola, A.;
Miyatake, H.;
Sanchez, J.;
Neveu, J.
SACC (Save All Correlations and Covariances) is a format and reference library for general storage
of summary statistic measurements for the Dark Energy Science Collaboration (DESC) within and from the Large Synoptic Survey Telescope (LSST) project's Dark Energy Science Collaboration.
[ascl:1601.006]
SAGE: Semi-Analytic Galaxy Evolution
Croton, Darren J.;
Stevens, Adam R. H.;
Tonini, Chiara;
Garel, Thibault;
Bernyk, Maksym;
Bibiano, Antonio;
Hodkinson, Luke;
Mutch, Simon J.;
Poole, Gregory B.;
Shattow, Genevieve M.
SAGE (Semi-Analytic Galaxy Evolution) models galaxy formation in a cosmological context. SAGE has been rebuilt to be modular and customizable. The model runs on any dark matter cosmological N-body simulation whose trees are organized in a supported format and contain a minimum set of basic halo properties.
[ascl:2312.028]
SAGE: Stellar Activity Grid for Exoplanets
SAGE corrects the time-dependent impact of stellar activity on transmission spectra. It uses a pixelation approach to model the stellar surface with spots and faculae, while accounting for limb-darkening and rotational line-broadening. The code can be used to evaluate stellar contamination for F to M-type hosts, test various spot sizes and locations, and quantify the impact of limb-darkening. SAGE can also retrieve the properties and distribution of active regions on the stellar surface from photometric monitoring, and connect the photometric variability to the stellar contamination of transmission spectra.
[ascl:2408.004]
Sailfish: GPU-accelerated grid-based astrophysics gas dynamics code
Sailfish simulates accreting binary systems, including binary protostars, post-AGN stellar binaries, mass-transferring X-ray binaries, and double black hole systems. The binary components are "on the grid" rather than excised, and are evolved according to the Kepler two-body problem, modified to account for gravitational wave losses or self-consistent forcing from the orbiting gas. The solvers are shock-capturing and are second order accurate in space and time. Gravity is fully Newtonian. Thermodynamics can be treated using a gamma-law equation of state with a blackbody cooling term, or in the locally isothermal approximation, in which the gas temperature is set to a constant times the local free-fall speed. Sailfish is fully Cartesian and has extensive diagnostic capabilities to facilitate accurate calculations of gas-driven orbital evolution or the extraction of electromagnetic disk signatures. The code is extremely efficient, reaching more than one billion zone updates per second on an NVIDIA A100 GPU, enabling extremely high resolution of complex flows around the binary components.
[ascl:1203.011]
SALT2: Spectral Adaptive Lightcurve Template
SALT (Spectral Adaptive Lightcurve Template) is a package for Type Ia Supernovae light curve fitting. Its main purpose is to provide a distance estimator but it can also be used for photometric redshifts, and spectroscopic + photometric identification. This code is also known by the name snfit.
[ascl:1407.006]
SAMI: Sydney-AAO Multi-object Integral field spectrograph pipeline
Allen, J. T.;
Green, A. W.;
Fogarty, L. M. R.;
Sharp, R.;
Nielsen, J.;
Konstantopoulos, I.;
Taylor, E. N.;
Scott, N.;
Cortese, L.;
Richards, S. N.;
Croom, S.;
Owers, M. S.;
Bauer, A. E.;
Sweet, S. M.;
Bryant, J. J.
The SAMI (Sydney-AAO Multi-object Integral field spectrograph) pipeline reduces data from the Sydney-AAO Multi-object Integral field spectrograph (SAMI) for the SAMI Galaxy Survey. The python code organizes SAMI data and, along with the AAO 2dfdr package, carries out all steps in the data reduction, from raw data to fully calibrated datacubes. The principal steps are: data management, use of 2dfdr to produce row-stacked spectra, flux calibration, correction for telluric absorption, removal of atmospheric dispersion, alignment of dithered exposures, and drizzling onto a regular output grid. Variance and covariance information is tracked throughout the pipeline. Some quality control routines are also included.
[ascl:1504.011]
samiDB: A Prototype Data Archive for Big Science Exploration
samiDB is an archive, database, and query engine to serve the spectra, spectral hypercubes, and high-level science products that make up the SAMI Galaxy Survey. Based on the versatile Hierarchical Data Format (HDF5), samiDB does not depend on relational database structures and hence lightens the setup and maintenance load imposed on science teams by metadata tables. The code, written in Python, covers the ingestion, querying, and exporting of data as well as the automatic setup of an HTML schema browser. samiDB serves as a maintenance-light data archive for Big Science and can be adopted and adapted by science teams that lack the means to hire professional archivists to set up the data back end for their projects.
[ascl:2207.011]
samsam: Scaled Adaptive Metropolis SAMpler
The samsam package provides two samplers, a scaled adaptive metropolis algorithm to robustly obtain samples from a target distribution, and a covariance importance sampling algorithm to efficiently compute the model evidence (or other integrals). It also includes tools to assess the convergence of the sam sampler and a few commonly used prior distributions.
[ascl:2307.030]
SAMUS: Simulator of Asteroid Malformation Under Stress
SAMUS (Simulator of Asteroid Malformation Under Stress) simulates the deformation of minor bodies, assuming that they are homogenous incompressible fluid masses. They are initialized as ellipsoids and the Navier-Stokes equations are interatively solved to investigate the deformation of the body over time. The software is modular and allows for user-defined output functions, size, and trajectories. Structured as a single large class, SAMUS can store variables and handle arbitrary function calls, which eases debugging and investigation, especially for lengthy high-fidelity simulation runs.
[ascl:1605.015]
SAND: Automated VLBI imaging and analyzing pipeline
The Search And Non-Destroy (SAND) is a VLBI data reduction pipeline composed of a set of Python programs based on the AIPS interface provided by ObitTalk. It is designed for the massive data reduction of multi-epoch VLBI monitoring research. It can automatically investigate calibrated visibility data, search all the radio emissions above a given noise floor and do the model fitting either on the CLEANed image or directly on the uv data. It then digests the model-fitting results, intelligently identifies the multi-epoch jet component correspondence, and recognizes the linear or non-linear proper motion patterns. The outputs including CLEANed image catalogue with polarization maps, animation cube, proper motion fitting and core light curves. For uncalibrated data, a user can easily add inline modules to do the calibration and self-calibration in a batch for a specific array.
[ascl:0003.002]
SAOImage DS9: A utility for displaying astronomical images in the X11 window environment
SAOImage DS9 is an astronomical imaging and data visualization application. DS9 supports FITS images and binary tables, multiple frame buffers, region manipulation, and many scale algorithms and colormaps. It provides for easy communication with external analysis tasks and is highly configurable and extensible via XPA and SAMP. DS9 is a stand-alone application. It requires no installation or support files. Versions of DS9 currently exist for Solaris, Linux, MacOSX, and Windows. All versions and platforms support a consistent set of GUI and functional capabilities. DS9 supports advanced features such as multiple frame buffers, mosaic images, tiling, blinking, geometric markers, colormap manipulation, scaling, arbitrary zoom, rotation, pan, and a variety of coordinate systems. DS9 also supports FTP and HTTP access. The GUI for DS9 is user configurable. GUI elements such as the coordinate display, panner, magnifier, horizontal and vertical graphs, button bar, and colorbar can be configured via menus or the command line. DS9 is a Tk/Tcl application which utilizes the SAOTk widget set. It also incorporates the X Public Access (XPA) mechanism to allow external processes to access and control its data, GUI functions, and algorithms.
[ascl:2112.015]
SAPHIRES: Stellar Analysis in Python for HIgh REsolution Spectroscopy
The SAPHIRES (Stellar Analysis in Python for HIgh REsolution Spectroscopy) suite contains functions for analyzing high-resolution stellar spectra. Though most of its functionality is aimed at deriving radial velocities (RVs), the suite also includes capabilities to measure projected rotational velocities (vsini) and determine spectroscopic flux ratios in double-lined binary systems (SB2s). These measurements are made primarily by computing spectral-line broadening functions. More traditional techniques such as Fourier cross-correlation, and two-dimensional cross-correlation (TODCOR) are also included.
[ascl:2507.013]
Sapphire++: Interaction of charged particles with a background plasma simulator
Sapphire++ (Simulating astrophysical plasmas and particles with highly relativistic energies in C++) numerically solves the Vlasov–Fokker–Planck equation for astrophysical applications. It employs a numerical algorithm based on a spherical harmonic expansion of the distribution function, expressing the Vlasov–Fokker–Planck equation as a system of partial differential equations governing the evolution of the expansion coefficients. The code uses the discontinuous Galerkin method in conjunction with implicit and explicit time stepping methods to compute these coefficients, providing significant flexibility in its choice of spatial and temporal accuracy.
[ascl:1210.029]
Sapporo: N-body simulation library for GPUs
Sapporo mimics the behavior of GRAPE hardware and uses the GPU to perform high-precision gravitational N-body simulations. It makes use of CUDA and therefore only works on NVIDIA GPUs. N-body codes currently running on GRAPE-6 can switch to Sapporo by a simple relinking of the library. Sapporo's precision is comparable to that of GRAPE-6, even though internally the GPU hardware is limited to single precision arithmetics. This limitation is effectively overcome by emulating double precision for calculating the distance between particles.
[ascl:2408.015]
SAQQARA: Stochastic gravitational wave background analysis
SAQQARA analyzes stochastic gravitational wave background signals. This Simulation-based Inference (SBI) library is built on top of the swyft code (ascl:2302.016), which implements neural ratio estimation to efficiently access marginal posteriors for all parameters of interest. Simulation-based inference combined with implicit marginalization (over nuisance parameters) has been shown to be well suited for SGWB data analysis.
[ascl:1907.005]
SARA-PPD: Preconditioned primal-dual algorithm for radio-interferometric imaging
SARA-PPD is a proof of concept MATLAB implementation of an acceleration strategy for a recently proposed primal-dual distributed algorithm. The algorithm optimizes resolution by accounting for the correct noise statistics, leverages natural weighting in the definition of the minimization problem for image reconstruction, and optimizes sensitivity by enabling accelerated convergence through a preconditioning strategy incorporating sampling density information. This algorithm offers efficient processing of large-scale data sets that will be acquired by next generation radio-interferometers such as the Square Kilometer Array.
[ascl:1904.020]
SARAH: SUSY and non-SUSY model builder and analyzer
SARAH builds and analyzes SUSY and non-SUSY models. It calculates all vertices, mass matrices, tadpoles equations, one-loop corrections for tadpoles and self-energies, and two-loop RGEs for a given model. SARAH writes model files for a variety of other software packages for dark matter studies, includes many SUSY and non-SUSY models, and makes implementing new models efficient and straightforward. Written in Mathematica, SARAH can also use output from Vevacious (ascl:1904.019) to check for the global minimum for a given model and parameter point.
[ascl:1404.004]
SAS: Science Analysis System for XMM-Newton observatory
The Science Analysis System (SAS) is an extensive suite of software tasks developed to process the data collected by the XMM-Newton Observatory. The SAS extracts standard (spectra, light curves) and/or customized science products, and allows reproductions of the reduction pipelines run to get the PPS products from the ODFs files. SAS includes a powerful and extensive suite of FITS file manipulation packages based on the Data Access Layer library.
[ascl:2302.013]
SASHIMI-C: Semi-Analytical SubHalo Inference ModelIng for Cold Dark Matter
SASHIMI-C calculates various subhalo properties efficiently using semi-analytical models for cold dark matter (CDM), providing a full catalog of dark matter subhalos in a host halo with arbitrary mass and redshift. Each subhalo is characterized by its mass and density profile both at accretion and at the redshift of interest, accretion redshift, and effective number (or weight) corresponding to that particular subhalo. SASHIMI-C computes the subhalo mass function without making any assumptions such as power-law functional forms; the only assumed power law is that for the primordial power spectrum predicted by inflation. The code is not limited to numerical resolution nor to Poisson shot noise, and its results are well in agreement with those from numerical N-body simulations.
[ascl:2302.010]
SASHIMI-W: Semi-Analytical SubHalo Inference ModelIng for Warm Dark Matter
SASHIMI-W calculates various subhalo properties efficiently using semi-analytical models for warm dark matter (WDM); the code is based on the extended Press-Schechter formalism and subhalos' tidal evolution prescription. The calculated constraints are independent of physics of galaxy formation and free from numerical resolution and the Poisson noise, and its results are well in agreement with those from numerical N-body simulations.
[ascl:1707.002]
SASRST: Semi-Analytic Solutions for 1-D Radiative Shock Tubes
SASRST, a small collection of Python scripts, attempts to reproduce the semi-analytical one-dimensional equilibrium and non-equilibrium radiative shock tube solutions of <a href="https://ui.adsabs.harvard.edu/abs/2007ShWav..16..445L">Lowrie & Rauenzahn (2007)</a> and <a href="https://ui.adsabs.harvard.edu/abs/2008ShWav..18..129L">Lowrie & Edwards (2008)</a>, respectively. The included code calculates the solution for a given set of input parameters and also plots the results using Matplotlib. This software was written to provide validation for numerical radiative shock tube solutions produced by a radiation hydrodynamics code.
[ascl:2103.005]
satcand: Orbital stability and tidal migration constraints for KOI exomoon candidates
satcand applies theoretical constraints of orbital stability and tidal migration to KOI exomoon candidates. The package can evaluate the tidal migration within a Sun-Earth-Moon system, plot angular velocity over time, and calculate the migration time scale (T1) and the total migration time scale, among other things. In addition to the theoretical constraints, observational constraints can be applied.
[ascl:2203.011]
SATCHEL: Pipeline to search for long-period exoplanet signals
SATCHEL (Search Algorithm for Transits in the Citizen science Hunt for Exoplanets in Lightcurves) searches for individual signals of interest in time-series data classified through crowdsourcing. The pipeline was built for the purpose of finding long-period exoplanet transit signals in Kepler photometric time-series data, but may be adapted for searches for any kind of one-dimensional signals in crowdsourced classifications.
[ascl:2303.016]
SatGen: Semi-analytical satellite galaxy and dark matter halo generator
SatGen generates satellite-galaxy populations for host halos of desired mass and redshift. It combines halo merger trees, empirical relations for galaxy-halo connection, and analytic prescriptions for tidal effects, dynamical friction, and ram-pressure stripping. It emulates zoom-in cosmological hydrosimulations in certain ways and outperforms simulations regarding statistical power and numerical resolution.
[ascl:1309.005]
SATMC: SED Analysis Through Monte Carlo
SATMC is a general purpose, MCMC-based SED fitting code written for IDL and Python. Following Bayesian statistics and Monte Carlo Markov Chain algorithms, SATMC derives the best fit parameter values and returns the sampling of parameter space used to construct confidence intervals and parameter-parameter confidence contours. The fitting may cover any range of wavelengths. The code is designed to incorporate any models (and potential priors) of the user's choice. The user guide lists all the relevant details for including observations, models and usage under both IDL and Python.
[ascl:2507.014]
SAUSERO: Software to AUtomatize in a Simple Environment the Reduction of Osiris+
SAUSERO (Software to AUtomatize in a Simple Environment the Reduction of Osiris+) processes raw science frames to address noise, cosmetic defects, and pixel heterogeneity, preparing them for photometric analysis for OSIRIS+ (Gran Telescopio Canarias). Correcting these artifacts is a critical prerequisite for reliable scientific analysis. The software applies observation-specific reduction steps, ensuring optimized treatment for different data types. Developed with a focus on simplicity and efficiency, SAUSERO streamlines the reduction pipeline, enabling researchers to obtain calibrated data ready for photometric studies.
[ascl:2306.003]
SAVED21cm: Global 21cm signal extraction pipeline
SAVED21cm extracts the 21cm signal from the simulated mock observation for the Radio Experiment for the Analysis of Cosmic Hydrogen (REACH). Though built for the REACH experiment, this 21cm signal extraction pipeline can in principle can be utilized for any global 21cm experiment. The toolkit is based on a pattern recognition framework using the Singular Value Decomposition (SVD) of the 21cm and foreground training set. SAVED21cm finds the patterns in the training sets and properly models the chromatic distortions with a better basis than the polynomials.
[ascl:1601.012]
SavGolFilterCov: Savitzky Golay filter for data with error covariance
A Savitzky–Golay filter is often applied to data to smooth the data without greatly distorting the signal; however, almost all data inherently comes with noise, and the noise properties can differ from point to point. This python script improves upon the traditional Savitzky-Golay filter by accounting for error covariance in the data. The inputs and arguments are modeled after scipy.signal.savgol_filter.
[ascl:1904.015]
SBGAT: Small Bodies Geophysical Analysis Tool
SBGAT (Small Body Geophysical Analysis Tool) generates simulated data originating from small bodies shape models, combined with advanced shape-modification properties. It uses polyhedral shape models from which can be computed mass properties such as volume, center of mass, and inertia, synthetic observations such as lightcurves and radar, and which can be used within dynamical models, such as spherical harmonics and polyhedron gravity modeling. SBGAT can generate spherical harmonics expansions from constant-density polyhedra (and export them to JSON) and evaluate the spherical harmonics expansions. It can also generate YORP coefficients, multi-threaded Polyhedron Gravity Model gravity and potential evaluations, and synthetic light-curve and radar observations for single/primary asteroids.
SBGAT has two distinct packages: a dynamic library SBGAT Core that contains the data structure and algorithm backbone of SBGAT, and SBGAT Gui, which wraps the former inside a VTK, Qt user interface to facilitate user/data interaction. SBGAT Core can be used without the SBGAT Gui wrapper.
[ascl:2306.002]
sbi: Simulation-based inference toolkit
Simulation-based inference is the process of finding parameters of a simulator from observations. The PyTorch package sbi performs simulation-based inference by taking a Bayesian approach to return a full posterior distribution over the parameters, conditional on the observations. This posterior can be amortized (<i>i.e.</i> useful for any observation) or focused (<i>i.e.</i>tailored to a particular observation), with different computational trade-offs. The code offers a simple interface for one-line posterior inference.
[ascl:2506.016]
SBI++: Simulation-based (likelihood-free) inference for astronomical applications
SBI++ is a complete methodology based on simulation-based (likelihood-free) inference that is customized for astronomical applications. Specifically, the code retains the fast inference speed of ∼1 sec for objects in the observational training set distribution, and additionally permits parameter inference outside of the trained noise and data at ~1 min per object. The package includes scripts for training and implementing SBI++ and is dependent on sbi (ascl:2306.002).
[ascl:1907.014]
sbpy: Small-body planetary astronomy
Mommert, Michael;
Kelley, Michael S. P.;
de Val-Borro, Miguel;
Li, Jian-Yang;
Guzman, Giannina;
Sipőcz, Brigitta;
Ďurech, Josef;
Granvik, Mikael;
Grundy, Will;
Moskovitz, Nick;
Penttilä, Antti;
Samarasinha, Nalin
sbpy, an Astropy affiliated package, supplements functionality provided by Astropy (ascl:1304.002) with functions and methods that are frequently used for planetary astronomy with a clear focus on asteroids and comets. It offers access tools for various databases for orbital and physical data, spectroscopy analysis tools and models, photometry models for resolved and unresolved observations, ephemerides services, and other tools useful for small-body planetary astronomy.
[ascl:1010.063]
SCAMP: Automatic Astrometric and Photometric Calibration
Astrometric and photometric calibrations have remained the most tiresome step in the reduction of large imaging surveys. SCAMP has been written to address this problem. The program efficiently computes accurate astrometric and photometric solutions for any arbitrary sequence of FITS images in a completely automatic way. SCAMP is released under the GNU General Public License.
[ascl:2002.006]
ScamPy: Sub-halo Clustering and Abundance Matching Python interface
ScamPy "paints" an observed population of cosmological objects on top of the DM-halo/subhalo hierarchy obtained from DM-only simulations. The method combines the Halo Occupation Distribution (HOD) method with sub-halo abundance matching (SHAM); the two processes together are dubbed Sub-halo clustering and abundance matching (SCAM). The procedure requires applying the two methods in sequence; by applying the HOD scheme, the host sub-haloes are selected, and the SHAM algorithm associates an observable property of choice to each sub-halo. The provided python interface allows users to load and populate DM halos and sub-halos obtained by FoF and SUBFIND algorithms applied to DM snapshots at any redshift. The software is highly-optimized and flexible.
[ascl:1209.012]
Scanamorphos: Maps from scan observations made with bolometer arrays
Scanamorphos is an IDL program to build maps from scan observations made with bolometer arrays. Scanamorphos can post-process scan observations performed with the Herschel photometer arrays. This post-processing mainly consists in subtracting the total low-frequency noise (both its thermal and non-thermal components), masking cosmic ray hit residuals, and projecting the data onto a map. Although it was developed for Herschel, it is also applicable with minimal adjustment to scan observations made with other bolometer arrays provided they entail sufficient redundancy; it was successfully applied to P-Artemis, an instrument operating on the APEX telescope. Scanamorphos does not assume any particular noise model and does not apply any Fourier-space filtering to the data. It is an empirical tool using only the redundancy built in the observations, taking advantage of the fact that each portion of the sky is sampled at multiple times by multiple bolometers. The user is allowed to optionally visualize and control results at each intermediate step, but the processing is fully automated.
[ascl:1803.003]
scarlet: Source separation in multi-band images by Constrained Matrix Factorization
<small>SCARLET</small> performs source separation (aka "deblending") on multi-band images. It is geared towards optical astronomy, where scenes are composed of stars and galaxies, but it is straightforward to apply it to other imaging data. Separation is achieved through a constrained matrix factorization, which models each source with a Spectral Energy Distribution (SED) and a non-parametric morphology, or multiple such components per source. The code performs forced photometry (with PSF matching if needed) using an optimal weight function given by the signal-to-noise weighted morphology across bands. The approach works well if the sources in the scene have different colors and can be further strengthened by imposing various additional constraints/priors on each source. Because of its generic utility, this package provides a stand-alone implementation that contains the core components of the source separation algorithm. However, the development of this package is part of the LSST Science Pipeline; the meas_deblender package contains a wrapper to implement the algorithms here for the LSST stack.
[ascl:2208.003]
Scatfit: Scattering fits of time domain radio signals (Fast Radio Bursts or pulsars)
Scatfit models observed burst signals of impulsive time domain radio signals (<i> e.g.</i>, Fast Radio Bursts (FRBs) or pulsars pulses), which usually are convolution products of various effects, and fits them to the experimental data. It includes several models for scattering and instrumental effects. The code loads the experimental time domain radio data, cleans them, fits an aggregate scattering model to the data, and robustly estimates the model parameters and their uncertainties. Additionally, scatfit determines the scaling of the scattering time with frequency, <i>i.e.</i> the scattering index, and the scattering-corrected dispersion measure of the burst or pulse.
[ascl:2505.018]
SCATTERING: Solve the coupled equations for a given scattering system
The SCATTERING code solves the coupled equations for a given scattering system, provides the scattering S-matrix elements, and calculates the state-to-state cross-sections. Its approach is different from codes such as MOLSCAT (ascl:1206.004) or Hibridon (ascl:2505.020), as SCATTERING solves coupled equations in the body-fixed (BF) frame, where the coupling matrix exhibits a predominantly block-diagonal structure with blocks interconnected by centrifugal terms. This significantly reduces computational time and memory requirements.
[ascl:1505.008]
SCEPtER: Stellar CharactEristics Pisa Estimation gRid
SCEPtER (Stellar CharactEristics Pisa Estimation gRid) estimates the stellar mass and radius given a set of observable quantities; the results are obtained by adopting a maximum likelihood technique over a grid of pre-computed stellar models. The code is quite flexible since different observables can be used, depending on their availability, as well as different grids of models.
[ascl:2306.060]
SCF-FDPS: Disk-halo systems simulator
The fast N-body code SCF-FDPS (Self-Consistent Field-Framework for Developing Particle Simulators) simulates disk-halo systems. It combines a self-consistent field (SCF) code, which provides scalability, and a tree code that is parallelized using the Framework for Developing Particle Simulators (FDPS) (ascl:1604.011). SCF-FDPS handles a wide variety of halo profiles and can be used to study extensive dynamical problems of disk-halo systems.
[ascl:2103.013]
schNell: Fast calculation of N_ell for GW anisotropies
schNell computes basic map-level noise properties for generic networks of gravitational wave interferometers, primarily the noise power spectrum "N_ell", but this lightweight python module that can also be used for, for example, antenna patterns, overlap functions, and inverse variance maps, among other tasks. The code has three main classes; detectors contain information about each individual detector of the network, such as their positions, noise properties, and orientation. NoiseCorrelations describes the noise-level correlation between pairs of detectors, and the MapCalculators class combines a list of Detectors into a network (potentially together with a NoiseCorrelation object) and computes the corresponding map-level noise properties arising from their correlations.
[ascl:1907.001]
schwimmbad: Parallel processing pools interface
schwimmbad provides a uniform interface to parallel processing pools and enables switching easily between local development (e.g., serial processing or with multiprocessing) and deployment on a cluster or supercomputer (via, e.g., MPI or JobLib). The utilities provided by schwimmbad require that tasks or data be “chunked” and that code can be “mapped” onto the chunked tasks.
[ascl:2202.007]
SciCatalog: Tools for scientific data catalogs
SciCatalog handles catalogs of scientific data in a way that is easily extensible, including the ability to create nicely formatted AASTex deluxe tables for use in AAS Publishing manuscripts. It handles catalogs of values, their positive and negative uncertainties, and references for those values with methods for easily adding columns and changing values. The catalog is also backed up every time it is loaded under the assumption that it is about to be modified.
[ascl:1311.001]
SciDB: Open Source DMAS for Scientific Research
SciDB is a DMAS (Data Management and Analytics Software System) optimized for data management of big data and for big analytics. SciDB is organized around multidimensional array storage, a generalization of relational tables, and is designed to be scalable up to petabytes and beyond. Complex analytics are simplified with SciDB because arrays and vectors are first-class objects with built-in optimized operations. Spatial operators and time-series analysis are easy to express. Interfaces to common scientific tools like R as well as programming languages like C++ and Python are provided.
[ascl:2011.019]
Scintools: Pulsar scintillation data tools
SCINTOOLS (SCINtillation TOOLS) simulates and analyzes pulsar scintillation data. The code can be used for processing observed dynamic spectra, computing secondary spectra and ACFs, measuring scintillation arcs, simulating dynamic spectra, and modeling pulsar transverse velocities through scintillation arcs or diffractive timescales.
[ascl:2306.013]
SCONCE-SCMS: Spherical and conic cosmic web finders with extended SCMS algorithms
SCONCE-SCMS detects cosmic web structures, primarily cosmic filaments and the associated cosmic nodes, from a collection of discrete observations with the extended subspace constrained mean shift (SCMS) algorithms on the unit (hyper)sphere (in most cases, the 2D (RA,DEC) celestial sphere), and the directional-linear products space (most commonly, the 3D (RA,DEC,redshift) light cone).
The subspace constrained mean shift (SCMS) algorithm is a gradient ascent typed method dealing with the estimation of local principal curves, more widely known as density ridges. The one-dimensional density ridge traces over the curves where observational data are highly concentrated and thus serves as a natural model for cosmic filaments in our Universe. Modeling cosmic filaments as density ridges enables efficient estimation by the kernel density estimator (KDE) and the subsequent SCMS algorithm in a statistically consistent way. While the standard SCMS algorithm can identify the density ridges in any "flat" Euclidean space, it exhibits large bias in estimating the density ridges on the data space with a non-linear curvature. The extended SCMS algorithms used in SCONCE-SCMS are adaptive to the spherical and conic geometries and resolve the estimation bias of the standard SCMS algorithm on a 2D (RA,DEC) celestial sphere or 3D (RA,DEC,redshift) light cone.
[ascl:2503.037]
SCONE: Supernova Classification with a Convolutional Neural Network
SCONE (Supernova Classification with a Convolutional Neural Network) classifies supernovae (SNe) by type using multi-band photometry data (lightcurves) using a convolutional neural networks. SCONE takes in supernova (SN) photometry data in the format output by SNANA simulations, separated into two types of files: metadata and observation data. Photometric data is pre-processed via 2D Gaussian process regression, which smooths over irregular sampling rates between filters and also allows SCONE to be independent of the filter set on which it was trained.
[submitted]
ScopeSim
An attempt at creating a common pythonic framework for visual and infrared telescope instrument data simulators.
[submitted]
ScopeSim Instrument Reference Database
A reference database for astronomical instrument and telescope characteristics for all types of visual and infrared systems. Instrument packages are used in conjunction with the ScopeSim instrument data simulator.
[submitted]
ScopeSim Templates
Templates and helper functions for creating on-sky Source description objects for the ScopeSim instrument data simulation engine.
[ascl:2209.005]
SCORE: Shape COnstraint REstoration
The Shape COnstraint REstoration algorithm (SCORE) is a proximal algorithm based on sparsity and shape constraints to restore images. Its main purpose is to restore images while preserving their shape information. It can, for example, denoise a galaxy image by instanciating SCORE and using its denoise method and then visualize the results, and can deconvolve multiple images with different parameter values.
[ascl:2112.003]
SCORPIO: Sky COllector of galaxy Pairs and Image Output
The Python package SCORPIO retrieves images and associated data of galaxy pairs based on their position, facilitating visual analysis and data collation of multiple archetypal systems. The code ingests information from SDSS, 2MASS and WISE surveys based on the available bands and is designed for studies of galaxy pairs as natural laboratories of multiple astrophysical phenomena for, among other things, tidal force deformation of galaxies, pressure gradient induced star formation regions, and morphological transformation.
[ascl:1601.003]
SCOUSE: Semi-automated multi-COmponent Universal Spectral-line fitting Engine
Henshaw, J. D.;
Longmore, S. N.;
Kruijssen, J. M. D.;
Davies, B.;
Bally, J.;
Barnes, A.;
Battersby, C.;
Burton, M.;
Cunningham, M. R.;
Dale, J. E.;
Ginsburg, A.;
Immer, K.;
Jones, P. A.;
Kendrew, S.;
Mills, E. A. C.;
Molinari, S.;
Moore, T. J. T.;
Ott, J.;
Pillai, T.;
Rathborne, J.;
Schilke, P.;
Schmiedeke, A.;
Testi, L.;
Walker, D.;
Walsh, A.;
Zhang, Q.
The Semi-automated multi-COmponent Universal Spectral-line fitting Engine (SCOUSE) is a spectral line fitting algorithm that fits Gaussian files to spectral line emission. It identifies the spatial area over which to fit the data and generates a grid of spectral averaging areas (SAAs). The spatially averaged spectra are fitted according to user-provided tolerance levels, and the best fit is selected using the Akaike Information Criterion, which weights the chisq of a best-fitting solution according to the number of free-parameters. A more detailed inspection of the spectra can be performed to improve the fit through an iterative process, after which SCOUSE integrates the new solutions into the solution file.
[ascl:2303.011]
Scri: Manipulate time-dependent functions of spin-weighted spherical harmonics
Scri manipulates time-dependent functions of spin-weighted spherical harmonics. It implements the BMS transformations of the most common gravitational waveforms, including the Newman-Penrose quantity ψ4, the Bondi news function, the shear spin coefficient σ, and the transverse-traceless metric perturbation h, as well as the remaining Newman-Penrose quantities ψ0 through ψ3.
[ascl:2204.013]
SCRIPT: Semi-numerical Code for ReIonization with PhoTon-conservation
SCRIPT (Semi-numerical Code for ReIonization with PhoTon-conservation) generates the ionization field during the epoch of cosmological reionization using a photon-conserving algorithm. The code depends on density and velocity files obtained using a N-body simulation, which can be generated with a 2LPT code such as MUSIC (ascl:1311.011).
[ascl:2202.018]
Sculptor: Interactive modeling of astronomical spectra
Sculptor manipulates, models and analyzes spectroscopic data; the code facilitates reproducible scientific results and easy to inspect model fits. A built-in graphical user interface around LMFIT (ascl:1606.014) offers interactive control to set up and combine multiple spectral models to fully fit the spectrum of choice. Alternatively, all core functionality can be scripted to facilitate the design of spectral fitting and analysis pipelines.
[ascl:2002.001]
SDAR: Slow-Down Algorithmic Regularization code for solving few-body problems
SDAR (Slow-Down Algorithmic Regularization) simulates the long-term evolution of few-body systems such as binaries and triples. The algorithm used provides a few orders of magnitude faster performance than the classical N-body method. The secular evolution of hierarchical systems, <i>e.g.</i> Kozai-Lidov oscillation, can be well reproduced. The code is written in the C++ language and can be used either as a stand-alone tool or a library to be plugged in other N-body codes. The high precision of the floating point to 62 digits is also supported.
[ascl:2408.001]
SDR: Sharpened Dimensionality Reduction
Sharpened dimensionality reduction (SDR) sharpens original data before dimensionality reduction to create visually segregrated sample clusters. user-guided labeling. Each distinct cluster can then be labeled and used to further analyze an otherwise unlabeled data set. Written in C++, SDR scales well with large high-dimensional data.
[submitted]
SDSS Dual Active Nuclei Galaxy Detection Pipeline
Dual Active Nuclei Galaxies (DAGNs) are rare occurrences in the sky. Until now, most AGNs have been described to be found serendipitously, or by manual observation. In recent years, there has been an increasing interest in such dual AGNs and their astrophysical properties. Their study is important to the understanding of galaxy formation, star formation and these objects are the precursors to Gravitational Wave Sources.
Hence, we have devised a pipeline, that along with systematic data collection, can detect such dual AGN candidates. A novel algorithm 'Graph-Boosted Gradient Ascent' has been devised to detect whether an R-band image of a galaxy is a potential candidate for a DAGN or not. The pipeline can be cloned to a user's machine, and by joining the AstrIRG_DAGN group on SciServer, astronomers can collectively contribute to the mining of DAGNs.
[ascl:2012.015]
seaborn: Statistical data visualization
Waskom, Michael;
Botvinnik, Olga;
Gelbart, Maoz;
Ostblom, Joel;
Hobson, Paul;
Lukauskas, Saulius;
Gemperline, David C;
Augspurger, Tom;
Halchenko, Yaroslav;
Warmenhoven, Jordi;
Cole, John B.;
De Ruiter, Julian;
Vanderplas, Jake;
Hoyer, Stephan;
Pye, Cameron;
Miles, Alistair;
Swain, Corban;
Meyer, Kyle;
Martin, Marcel;
Bachant, Pete Quintero, Eric;
Kunter, Gero;
Villalba, Santi;
Brian;
Fitzgerald, Clark;
Evans, C. G.;
Williams, Mike Lee;
O'Kane, Drew;
Yarkoni, Tal;
Brunner, Thomas
Seaborn provides a high-level interface for drawing attractive statistical graphics. Written in Python, it builds on matplotlib and integrates closely with pandas data structures. Its plotting functions operate on dataframes and arrays containing whole datasets and internally perform the necessary semantic mapping and statistical aggregation to produce informative plots. Its dataset-oriented, declarative API allows the user to focus on what the different elements of the plots mean, rather than on the details of how to draw them.
[ascl:1210.012]
SearchCal: The JMMC Evolutive Search Calibrator Tool
SearchCal builds an evolutive catalog of stars suitable as calibrators within any given user-defined angular distance and magnitude around a scientific target. SearchCal can select suitable bright calibration stars (V ≤ 10; K ≤ 5.0) for obtaining the ultimate precision of current interferometric instruments like the VLTI and faint calibration stars up to K ~ 15 around the scientific target. Star catalogs available at the CDS are searched via web requests and provide the useful astrometric and photometric informations for selecting calibrators. The missing photometries are computed with an accuracy of about 0.1 mag. The stellar angular diameter is estimated with a precision of about 10% through newly determined surface-brightness versus color-index relations based on the I, J, H and K magnitudes. For each star the squared visibility is computed taking into account the central wavelength and the maximum baseline of the predicted observations.
[ascl:1201.003]
SeBa: Stellar and binary evolution
SeBa is the stellar and binary evolution module, fully integrated into the kira N-body integrator, for <a href="https://ascl.net/1010.076">Starlab</a> (ascl:1010.076), although it can also be used as a stand-alone module for non-dynamical applications. Due to the interaction between stellar evolution and stellar dynamics, it is difficult to solve for the evolution of both systems in a completely self-consistent way. The trajectories of stars are computed using a block timestep scheme, as described earlier. Stellar and binary evolution is updated at fixed intervals (every 1/64 of a crossing time, typically a few thousand years). Any feedback between the two systems may thus experience a delay of at most one timestep. Internal evolution time steps may differ for each star and binary, and depend on binary period, perturbations due to neighbors, and the evolutionary state of the star. Time steps in this treatment vary from several milliseconds up to (at most) a million years.
[ascl:1101.001]
Second-order Tight-coupling Code
Prior to recombination photons, electrons, and atomic nuclei rapidly scattered and behaved, almost, like a single tightly-coupled photon-baryon plasma. In order to solve the cosmological perturbation equations during that time, Cosmic Microwave Background (CMB) codes use the so-called tight-coupling approximation in which the problematic terms (i.e. the source of the stiffness) are expanded in inverse powers of the Thomson Opacity. Most codes only keep the terms linear in the inverse Thomson Opacity. We have developed a second-order tight-coupling code to test the validity of the usual first-order tight-coupling code. It is based on the publicly available code <a href="http://ascl.net/1102.026">CAMB</a>.
This code is no longer available.
[ascl:1909.003]
SecularMultiple: Hierarchical multiple system secular evolution model
SecularMultiple computes the secular (orbit-averaged) gravitational dynamics of hierarchical multiple systems composed of nested binary orbits (simplex-type systems) with any configuration and any number of bodies. A particle can represent a binary or a body. The structure of the system is determined by linking to other particles with the attributes child1 and child2, and tidal interactions and relativistic corrections are included in an ad hoc fashion. SecularMultiple also includes routines for external perturbations such as flybys and supernovae.
[ascl:2008.013]
SEDBYS: Spectral Energy Distribution Builder for Young Stars
SEDBYS (Spectral Energy Distribution Builder for Young Stars) provides command-line tools and uses existing functions from standard packages such as Astropy (ascl:1304.002) to collate archival photometric and spectroscopic data. It also builds and inspects SEDS, and automatically collates the necessary software references.
[ascl:2011.014]
SEDkit: Spectral energy distribution construction and analysis tools
SEDkit constructs and analyzes simple spectral energy distributions (SED). This collection of pure Python modules creates individual SEDs or SED catalogs from spectra and/or photometry and calculates fundamental parameters (fbol, Mbol, Lbol, Teff, mass, log(g)).
[ascl:1901.008]
SEDobs: Observational spectral energy distribution simulation
SEDobs uses state-of-the-art theoretical galaxy SEDs (spectral energy distributions) to create simulated observations of distant galaxies. It used BC03 and M05 theoretical models and allows the user to configure the simulated observation that are needed. For a given simulated galaxy, the user is able to simulate multi-spectral and multi-photometric observations.
[ascl:2012.013]
sedop: Optimize discrete versions of common SEDs
sedop is a Monte-Carlo minimization code designed to optimally construct spectral energy distributions (SEDs) for sources of ultraviolet and X-ray radiation employed in numerical simulations of reionization and radiative feedback.
[ascl:1905.026]
SEDPY: Modules for storing and operating on astronomical source spectral energy distribution
SEDPY performs a variety of tasks for astronomical spectral energy distributions. It can generate synthetic photometry through any filter, provides detailed modeling of extinction curves, and offers basic aperture photometry algorithms. SEDPY can also store and interpolate model SEDs, convolve absolute or apparent fluxes, and calculate rest-frame magnitudes.
[ascl:1607.020]
SEEK: Signal Extraction and Emission Kartographer
SEEK (Signal Extraction and Emission Kartographer) processes time-ordered-data from single dish radio telescopes or from the simulation pipline HIDE (ascl:1607.019), removes artifacts from Radio Frequency Interference (RFI), automatically applies flux calibration, and recovers the astronomical radio signal. With its companion code HIDE (ascl:1607.019), it provides end-to-end simulation and processing of radio survey data.
[ascl:2303.003]
SeeKAT: Localizer for transients detected in tied-array beams
SeeKAT is a Python implementation of a novel maximum-likelihood estimation approach to localizing transients and pulsars detected in multiple MeerKAT tied-array beams at once to (sub-)arcsecond precision. It reads in list of detections (RA, Dec, S/N) and the beam PSF and computes a covariance matrix of the S/N value ratios, assuming 1-sigma Gaussian errors on each measurement. It models the aggregate beam response by arranging beam PSFs appropriately relative to each other and calculates a likelihood distribution of obtaining the observed S/N in each beam according to the modeled response. In addition, SeeKAT can plot the likelihood function over RA and Dec with 1-sigma uncertainty, overlaid on the beam coordinates and sizes.
[ascl:1411.007]
segueSelect: SDSS/SEGUE selection function modelling
The Python package segueSelect automatically models the SDSS/SEGUE selection fraction -- the fraction of stars with good spectra -- as a continuous function of apparent magnitude for each plate. The selection function can be determined for any desired sample cuts in signal-to-noise ratio, u-g, r-i, and E(B-V). The package requires Pyfits (ascl:1207.009) and, for coordinate transformations, galpy (ascl:1411.008). It can calculate the KS probability that the spectropscopic sample was drawn from the underlying photometric sample with the model selection function, plot the cumulative distribution function in r-band apparent magnitude of the spectroscopic sample (red) and the photometric sample+selection-function-model for this plate, and, if galpy is installed, can transform velocities into the Galactic coordinate frame. The code can also determine the selection function for SEGUE K stars.
[ascl:2110.019]
SELCIE: Screening Equations Linearly Constructed and Iteratively Evaluated
SELCIE (Screening Equations Linearly Constructed and Iteratively Evaluated) investigates the chameleon model that arises from screening a scalar field introduced in some modified gravity models that is coupled to matter. The code provides tools to construct user defined meshes by utilizing the GMSH mesh generation software. These tools include constructing shapes whose boundaries are defined by some function or by constructing it out of basis shapes such as circles, cones and cylinders. The mesh can also be separated into subdomains, each of which having its own refinement parameters. These meshes can then be converted into a format that is compatible with the finite element software FEniCS. SELCIE uses FEniCS (ascl:2110.018) with a nonlinear solving method (Picard or Newton method) to solve the chameleon equation of motion for some parameters and density distribution. These density distributions are constructed by having the density profile of each subdomain being set by a user defined function, allowing for extremely customizable setups that are easy to implement.
[ascl:2301.006]
Self-cal: Optical/IR long-baseline interferometry
Self-cal produces radio-interferometric images of an astrophysical object. The code is an adaptation of the self-calibration algorithm to optical/infrared long-baseline interferometry, especially to make use of differential phases and differential visibilities. It works together with the Mira image reconstruction software and has been used mainly on VLTI data. Self-cal, written in Yorick, is also available as part of fitsOmatic (ascl:2301.005).
[ascl:1504.009]
Self-lensing binary code with Markov chain
The self-lensing binary code with Markov chain code was used to analyze the self-lensing binary system KOI-3278. It includes the MCMC modeling and the key figures.
[ascl:2012.003]
Sengi: Interactive viewer for spectral outputs from stellar population synthesis models
Sengi enables online viewing of the spectral outputs of stellar population synthesis (SPS) codes. Typical SPS codes require significant disk space or computing resources to produce spectra for simple stellar populations with arbitrary parameters, making it difficult to present their results in an interactive, web-friendly format. Sengi uses Non-negative Matrix Factorisation (NMF) and bilinear interpolation to estimate output spectra for arbitrary values of stellar age and metallicity; this reduces the disk requirements and computational expense, allowing Sengi to serve the results in a client-based Javascript application.
[ascl:1807.026]
SENR: Simple, Efficient Numerical Relativity
SENR (Simple, Efficient Numerical Relativity) provides the algorithmic framework that combines the C codes generated by NRPy+ (ascl:1807.025) into a functioning numerical relativity code. It is part of the numerical relativity code package SENR/NRPy+. The package extends previous implementations of the BSSN reference-metric formulation to a much broader class of curvilinear coordinate systems, making it suitable for modeling physical configurations with approximate or exact symmetries, such as modeling black hole dynamics.
[ascl:1811.004]
SEP: Source Extraction and Photometry
SEP (Source Extraction and Photometry) makes the core algorithms of Source Extractor (ascl:1010.064) available as a library of standalone functions and classes. These operate directly on in-memory arrays (no FITS files or configuration files). The code is derived from the Source Extractor code base (written in C) and aims to produce results compatible with Source Extractor whenever possible. SEP consists of a C library with no dependencies outside the standard library and a Python module that wraps the C library in a Pythonic API. The Python wrapper operates on NumPy arrays with NumPy as its only dependency. It is generated using Cython.
From Source Extractor, SEP includes background estimation, image segmentation (including on-the-fly filtering and source deblending), aperture photometry in circular and elliptical apertures, and source measurements such as Kron radius, "windowed" position fitting, and half-light radius. It also adds the following features that are not available in Source Extractor: optimized matched filter for variable noise in source extraction; circular annulus and elliptical annulus aperture photometry functions; local background subtraction in shape consistent with aperture in aperture photometry functions; exact pixel overlap mode in all aperture photometry functions; and masking of elliptical regions on images.
[ascl:1404.005]
SER: Subpixel Event Repositioning Algorithms
Subpixel Event Repositioning (SER) techniques significantly improve the already unprecedented spatial resolution of Chandra X-ray imaging with the Advanced CCD Imaging Spectrometer (ACIS). Chandra CCD SER techniques are based on the premise that the impact position of events can be refined, based on the distribution of charge among affected CCD pixels. Unlike ACIS SER models that are restricted to corner split (3- and 4-pixel) events and assume that such events take place at the split pixel corners, this IDL code uses two-pixel splits as well, and incorporates more realistic estimates of photon impact positions.
[ascl:1102.010]
SEREN: A SPH code for star and planet formation simulations
SEREN is an astrophysical Smoothed Particle Hydrodynamics code designed to investigate star and planet formation problems using self-gravitating hydrodynamics simulations of molecular clouds, star-forming cores, and protostellar disks.
SEREN is written in Fortran 95/2003 with a modular philosophy for adding features into the code. Each feature can be easily activated or deactivated by way of setting options in the Makefile before compiling the code. This has the added benefit of allowing unwanted features to be removed at the compilation stage resulting in a smaller and faster executable program. SEREN is written with OpenMP directives to allow parallelization on shared-memory architecture.
[ascl:1312.001]
SERPent: Scripted E-merlin Rfi-mitigation PipelinE for iNTerferometry
SERPent is an automated reduction and RFI-mitigation procedure that uses the SumThreshold methodology. It was originally developed for the LOFAR pipeline. SERPent is written in Parseltongue, enabling interaction with the Astronomical Image Processing Software (<a href="http://ascl.net/9911.003">AIPS</a>) program. Moreover, SERPent is a simple "out of the box" Python script, which is easy to set up and is free of compilers.
[ascl:1304.009]
Sérsic: Exact deprojection of Sérsic surface brightness profiles
Sérsic is an implementation of the exact deprojection of Sérsic surface brightness profiles described in Baes and Gentile (2011). This code depends on the <a href="http://mpmath.googlecode.com">mpmath python library</a> for an implementation of the Meijer G function required by the Baes and Gentile (hereafter B+G) formulas for rational values of the Sérsic index. Sérsic requires rational Sérsic indices, but any irrational number can be approximated arbitrarily well by some rational number. The code also depends on <a href="http://scipy.org">scipy</a>, but the dependence is mostly for testing. The implementation of the formulas and the formulas themselves have undergone comprehensive testing.