ASCL.net

Astrophysics Source Code Library

Making codes discoverable since 1999

Browsing Codes

Results 3501-3600 of 4,021 (3911 ASCL, 110 submitted)

Order
Title Date
 
Mode
Abstract Compact
Per Page
50 100 250 All
 
Random
[ascl:2601.017] Cesam2k20: One-dimensional stellar structure and evolution modeling
Cesam2k20 computes one-dimensional stellar structure and evolution models under the assumption of hydrostatic equilibrium. The code evolves stellar models from pre-main sequence through advanced evolutionary stages using modular numerical schemes and up-to-date microphysics. Cesam2k20 builds on the CESAM (ascl:1010.059) stellar evolution code and incorporates developments from related tools such as ADIPLS (ascl:1109.002) for asteroseismic applications. It includes treatments of processes such as chemical mixing, angular momentum transport, and rotation within a unified framework. Cesam2k20 supports stellar modeling studies in both classical and asteroseismic contexts.
[ascl:1010.059] CESAM: A Free Code for Stellar Evolution Calculations
The Cesam code is a consistent set of programs and routines which perform calculations of 1D quasi-hydrostatic stellar evolution including microscopic diffusion of chemical species and diffusion of angular momentum. The solution of the quasi-static equilibrium is performed by a collocation method based on piecewise polynomials approximations projected on a B-spline basis; that allows stable and robust calculations, and the exact restitution of the solution, not only at grid points, even for the discontinuous variables. Other advantages are the monitoring by only one parameter of the accuracy and its improvement by super-convergence. An automatic mesh refinement has been designed for adjusting the localisations of grid points according to the changes of unknowns. For standard models, the evolution of the chemical composition is solved by stiffly stable schemes of orders up to four; in the convection zones mixing and evolution of chemical are simultaneous. The solution of the diffusion equation employs the Galerkin finite elements scheme; the mixing of chemicals is then performed by a strong turbulent diffusion. A precise restoration of the atmosphere is allowed for.
[ascl:1610.002] CERES: Collection of Extraction Routines for Echelle Spectra
The Collection of Extraction Routines for Echelle Spectra (CERES) constructs automated pipelines for the reduction, extraction, and analysis of echelle spectrograph data. This modular code includes tools for handling the different steps of the processing: CCD reductions, tracing of the echelle orders, optimal and simple extraction, computation of the wave-length solution, estimation of radial velocities, and rough and fast estimation of the atmospheric parameters. The standard output of pipelines constructed with CERES is a FITS cube with the optimally extracted, wavelength calibrated and instrumental drift-corrected spectrum for each of the science images. Additionally, CERES includes routines for the computation of precise radial velocities and bisector spans via the cross-correlation method, and an automated algorithm to obtain an estimate of the atmospheric parameters of the observed star.
[ascl:1308.015] Ceph_code: Cepheid light-curves fitting
Ceph_code fits multi-band Cepheid light-curves using templates derived from OGLE observations. The templates include short period stars (<10 day) and overtone stars.
[ascl:1906.021] centerRadon: Center determination code in stellar images
centerRadon finds the center of stars based on Radon Transform to sub-pixel precision. For a coronagraphic image of a star, it starts from a given location, then for each sub-pixel position, it interpolates the image and sums the pixels along different angles, creating a cost function. The center of the star is expected to correspond with where the cost function maximizes. The default values are set for the STIS coronagraphic images of the Hubble Space Telescope by summing over the diagonals (i.e., 45° and 135°), but it can be generally applied to other high-contrast imaging instruments with or without Adaptive Optics systems such as HST-NICMOS, P1640, or GPI.
[ascl:2302.005] celmech: Sandbox for celestial mechanics calculations
celmech provides a variety of analytical and semianalytical tools for celestial mechanics and dynamical astronomy. The package interfaces closely with the REBOUND N-body integrator (ascl:1110.016), thus facilitating comparisons between calculation results and direct N-body integrations. celmech can isolate the contribution of particular resonances to a system's dynamical evolution, and can develop simple analytical models with the minimum number of terms required to capture a particular dynamical phenomenon.
[ascl:1612.016] CELib: Software library for simulations of chemical evolution
CELib (Chemical Evolution Library) simulates chemical evolution of galaxy formation under the simple stellar population (SSP) approximation and can be used by any simulation code that uses the SSP approximation, such as particle-base and mesh codes as well as semi-analytical models. Initial mass functions, stellar lifetimes, yields from type II and Ia supernovae, asymptotic giant branch stars, and neutron star mergers components are included and a variety of models are available for use. The library allows comparisons of the impact of individual models on the chemical evolution of galaxies by changing control flags and parameters of the library.
[ascl:1602.011] Celestial: Common astronomical conversion routines and functions
The R package Celestial contains common astronomy conversion routines, particularly the HMS and degrees schemes, and a large range of functions for calculating properties of different cosmologies (as used by the <a href="http://cosmocalc.icrar.org/">cosmocalc</a> website). This includes distances, ages, growth rate/factor and densities (e.g., Omega evolution and critical energy density). It also includes functions for calculating thermal properties of the CMB and Planck's equations and virial properties of halos in different cosmologies, and standard NFW and weak-lensing formulas and low level orbital routines for calculating Roche properties, Vis-Viva and free-fall times.
[ascl:2310.001] celerite2: Fast and scalable Gaussian Processes in one dimension
celerite2 is a re-write of celerite (ascl:1709.008), an algorithm for fast and scalable Gaussian Process (GP) Regression in one dimension. celerite2 improves numerical stability and integration with various machine learning frameworks. The implementation includes interfaces in Python and C++, with full support for PyMC (ascl:1610.016) and JAX (ascl:2111.002).
[ascl:1709.008] celerite: Scalable 1D Gaussian Processes in C++, Python, and Julia
celerite provides fast and scalable Gaussian Process (GP) Regression in one dimension and is implemented in C++, Python, and Julia. The celerite API is designed to be familiar to users of george and, like george, celerite is designed to efficiently evaluate the marginalized likelihood of a dataset under a GP model. This is then be used alongside a non-linear optimization or posterior inference library for the best results. celerite has been superceded by celerite2 (ascl:2310.001).
[ascl:2305.025] CELEBI: Precision localizations and polarimetric data for fast radio bursts
The Australian Square Kilometre Array Pathfinder (ASKAP) has been enabled by the Commensal Real-time ASKAP Fast Transients Collaboration (CRAFT) to detect Fast Radio Bursts (FRBs) in real-time and save raw antenna voltages containing FRB detections. CELEBI, the CRAFT Effortless Localization and Enhanced Burst Inspection pipeline, extends CRAFT’s existing software to process ASKAP voltages to produce sub-arcsecond precision localizations and polarimetric data at time resolutions as fine as 3 ns of FRB events. CELEBI uses Nextflow (ascl:2305.024) to link together Bash and Python code to perform software correlation, interferometric imaging, and beamforming, thereby making use of common astronomical software packages.
[ascl:2005.017] cdetools: Tools for Conditional Density Estimates
cdetools provides tools for evaluating conditional density estimates and has applications to photometric redshift estimation and likelihood-free cosmological inference. Available in R and Python, it provides functions for computing a so-called CDE loss function for tuning and assessing the quality of individual probability density functions (PDFs) and diagnostic functions that probe the population-level performance of the PDFs.
[ascl:1904.006] CDAWeb: Coordinated Data Analysis Web
CDAWeb (Coordinated Data Analysis Workshop Web) enables viewing essentially any data produced in Common Data Format/CDF with the ISTP/IACG Guidelines and supports interactive plotting of variables from multiple instruments on multiple investigations simultaneously on arbitrary, user-defined time-scales. It also supports data retrieval in both CDF or ASCII format. NASA's GSFC Space Physics Data Facility maintains a publicly available database that includes approximately 600 data variables from Geotail, Wind, Interball, Polar, SOHO, ancilliary spacecraft and ground-based investigations. CDAWeb includes high resolution digital data products that support event correlative science. The system combines the client-server user interface technology of the Web with a powerful set of customized routines based in the COTS Interactive Data Language (IDL) package to leverage the data format standards.
[ascl:1604.009] CCSNMultivar: Core-Collapse Supernova Gravitational Waves
CCSNMultivar aids the analysis of core-collapse supernova gravitational waves. It includes multivariate regression of Fourier transformed or time domain waveforms, hypothesis testing for measuring the influence of physical parameters, and the Abdikamalov et. al. catalog for example use. CCSNMultivar can optionally incorporate additional uncertainty due to detector noise and approximate waveforms from anywhere within the parameter space.
[ascl:1208.006] ccogs: Cosmological Calculations on the GPU
This suite contains two packages for computing cosmological quantities on the GPU: aperture_mass, which calculates the aperture mass map for a given dataset using the filter proposed by Schirmer et al (2007) (an NFW profile with exponential cut-offs at zero and large radii), and angular_correlation, which calculates the 2-pt angular correlation function using data and a flat distribution of randomly generated galaxies. A particular estimator is chosen, but the user has the flexibility to explore other estimators.
[ascl:1901.003] CCL: Core Cosmology Library
The Core Cosmology Library (CCL) computes basic cosmological observables and provides predictions for many cosmological quantities, including distances, angular power spectra, correlation functions, halo bias and the halo mass function through state-of-the-art modeling prescriptions. Fiducial specifications for the expected galaxy distributions for the Large Synoptic Survey Telescope (LSST) are also included, together with the capability of computing redshift distributions for a user-defined photometric redshift model. Predictions for correlation functions of galaxy clustering, galaxy-galaxy lensing and cosmic shear are within a fraction of the expected statistical uncertainty of the observables for the models and in the range of scales of interest to LSST. CCL is written in C and has a python interface.
[ascl:1707.004] CCFpams: Atmospheric stellar parameters from cross-correlation functions
CCFpams allows the measurement of stellar temperature, metallicity and gravity within a few seconds and in a completely automated fashion. Rather than performing comparisons with spectral libraries, the technique is based on the determination of several cross-correlation functions (CCFs) obtained by including spectral features with different sensitivity to the photospheric parameters. Literature stellar parameters of high signal-to-noise (SNR) and high-resolution HARPS spectra of FGK Main Sequence stars are used to calibrate the stellar parameters as a function of CCF areas.
[ascl:1511.013] CCDtoRGB: RGB image production from three-band astronomical images
CCDtoRGB produces red‐green‐blue (RGB) composites from three‐band astronomical images, ensuring an object with a specified astronomical color has a unique color in the RGB image rather than burnt‐out white stars. Use of an arcsinh stretch shows faint objects while simultaneously preserving the structure of brighter objects in the field, such as the spiral arms of large galaxies.
[ascl:1510.007] ccdproc: CCD data reduction software
Ccdproc is an affiliated package for the AstroPy package for basic data reductions of CCD images. The ccdproc package provides many of the necessary tools for processing of ccd images built on a framework to provide error propagation and bad pixel tracking throughout the reduction process.
[ascl:1403.021] CCDPACK: CCD Data Reduction Package
CCDPACK contains programs to debias, remove dark current, flatfield, register, resample and normalize data from single- or multiple-CCD instruments. The basic reduction stages can be set up using an X based GUI that controls an automated reduction system so one can to start working without any detailed knowledge of the package (or indeed of CCD reduction). Registration is performed using graphical, script based or automated techniques that keep the amount of work to a minimum. CCDPACK uses the <a href="http://ascl.net/1110.012">Starlink</a> environment (ascl:1110.012).
[ascl:2206.020] CCDLAB: FITS image viewer and data reducer
CCDLAB provides graphical user interface functionality for FITS image viewing and data reduction based on the JPFITS FITS-file interface. It can view, manipulate, and save FITS primary image data and image extensions, view and manipulate FITS image headers, and view FITS Bintable extensions. The code enables batch processing, viewing, and saving of FITS images and searching FITS files on disk. CCDLAB also provides general image reduction techniques, source detection and characterization, and can create World Coordinate Solutions automatically or manually for FITS images.
[ascl:2402.004] CCBH-Numerics: Cosmologically-coupled-black-holes formation mass numerics
CCBH-Numerics (previously called CCBH-PLPP) computes the probability of the existence of a single cosmologically coupled black hole (BH) with a formation mass below a specified threshold for given observational data of binary black holes (BBHs) from gravitational waves. The code uses the unbiased population of BBHs, as given by the power-law-plus-peak (PLPP) profile, as the observational input, and assumes that the detected BBHs are formed from stellar evolution, not primordial BHs. CCBH-Numerics also works with individual data from BBHs and for NSBH pairs as well.
[ascl:2406.009] CBiRd: Bias tracers In Redshift space
CBiRd (Code for Bias tracers In Redshift space) provides correlators in the Effective Field Theory of Large-Scale Structure (EFTofLSS) in a ready-to-use pipeline for cosmological analysis of galaxy-redshift surveys data. It provides a core calculation package (C++BiRd), a Python implementation of a Taylor expansion of the power spectrum around a reference cosmology for efficient evaluation (TBiRd), and libraries to correct for observational systematics. CBiRd also provides MCMC samplers (MCBiRd) for a power spectrum and bispectrum analysis of galaxy-redshift surveys data based on emcee (ascl:1303.002), and can provide an earlybird pass to explore the cosmos with LSS surveys.
[ascl:2404.001] cbeam: Coupled-mode propagator for slowly-varying waveguides
cbeam models the propagation of guided light through slowly-varying few-mode waveguides using the coupled-mode theory (CMT). When compared with more general numerical methods for waveguide simulation, such as the finite-differences beam propagation method (FD-BPM), numerical implementations of the CMT can be much more computationally efficient. Written in Python and Julia, the package provides a Pythonic class structure to define waveguides, with simple classes for directional couplers and photonic lanterns already provided. cbeam also doubles as a finite-element eigenmode solver.
[ascl:1904.012] CausticFrog: 1D Lagrangian Simulation Package
CausticFrog models the reaction of a system of orbiting particles to instantaneous mass loss. It applies to any spherically symmetric potential, and follows the radial evolution of shells of mass. CausticFrog tracks the inner and outer edge of each shell, whose radius evolves as a test particle. The amount of mass in each shell is fixed but multiple shells can overlap leading to higher densities.
[submitted] Caustic Mass Estimator for Galaxy Clusters
The caustic technique is a powerful method to infer cluster mass profiles to clustrocentric distances well beyond the virial radius. It relies in the measure of the escape velocity of the sistem using only galaxy redshift information. This method was introduced by Diaferio & Geller (1997) and Diaferio (1999). This code allows the caustic mass estimation for galaxy clusters, as well as outlier identification as a side effect. However, a pre-cleaning of interlopers is recommended, using e.g., the shifting-gapper technique.
[ascl:2108.007] catwoman: Transit modeling Python package for asymmetric light curves
catwoman models asymmetric transit lightcurves. Written in Python, it calculates light curves for any radially symmetric stellar limb darkening law, and where planets are modeled as two semi-circles of different radii. Catwoman is built on the batman library (ascl:1510.002) and uses its integration algorithm.
[ascl:1810.013] catsHTM: Catalog cross-matching tool
The catsHTM package quickly accesses and cross-matches large astronomical catalogs that have been reformatted into the HDF5-based file format. It performs efficient cone searches at resolutions from a few arc-seconds to degrees within a few milliseconds time, cross-match numerous catalogs, and can do general searches.
[ascl:2007.024] CaTffs: Calcium triplet indexes
CaTffs predicts the strength of calcium triplet indices (CaT*, PaT and CaT) on the basis of empirical fitting functions and performs required interpolations between the different local functions. Together with the indices predictions, the program also computes the random errors associated to such predictions resulting from the covariance matrices of the fits (for the indices CaT* and PaT). This ensures a reliable error index estimation for any combination of input atmospheric parameters.
[ascl:1206.008] Catena: Ensemble of stars orbit integration
Catena integrates the orbits of an ensemble of stars using the chain-regularization method (Mikkola & Aarseth) with an embedded Runge-Kutta integration method of 9(8)th order (Prince & Dormand).
[ascl:2108.008] CatBoost: High performance gradient boosting on decision trees library
CatBoost is a machine learning method based on gradient boosting over decision trees and can be used for ranking, classification, regression and other machine learning tasks for Python, R, Java, C++. It supports both numerical and categorical features and computation on CPU and GPU, and is fast and scalable. Visualization tools are also included in CatBoost.
[ascl:1804.013] CAT-PUMA: CME Arrival Time Prediction Using Machine learning Algorithms
CAT-PUMA (CME Arrival Time Prediction Using Machine learning Algorithms) quickly and accurately predicts the arrival of Coronal Mass Ejections (CMEs) of CME arrival time. The software was trained via detailed analysis of CME features and solar wind parameters using 182 previously observed geo-effective partial-/full-halo CMEs and uses algorithms of the Support Vector Machine (SVM) to make its predictions, which can be made within minutes of providing the necessary input parameters of a CME.
[ascl:1105.010] CASTRO: Multi-dimensional Eulerian AMR Radiation-hydrodynamics Code
CASTRO is a multi-dimensional Eulerian AMR radiation-hydrodynamics code that includes stellar equations of state, nuclear reaction networks, and self-gravity. Initial target applications for CASTRO include Type Ia and Type II supernovae. CASTRO supports calculations in 1-d, 2-d and 3-d Cartesian coordinates, as well as 1-d spherical and 2-d cylindrical (r-z) coordinate systems. Time integration of the hydrodynamics equations is based on an unsplit version of the piecewise parabolic method (PPM) with new limiters that avoid reducing the accuracy of the scheme at smooth extrema. CASTRO can follow an arbitrary number of isotopes or elements. The atomic weights and amounts of these elements are used to calculate the mean molecular weight of the gas required by the equation of state. CASTRO supports several different approaches to solving for self-gravity. The most general is a full Poisson solve for the gravitational potential. CASTRO also supports a monopole approximation for gravity, and a constant gravity option is also available. The CASTRO software is written in C++ and Fortran, and is based on the BoxLib software framework developed by CCSE.
[ascl:2511.001] CASSL: Conventional and Sub-Conventional Strong Lensing forecasts and simulations
The CASSL pipeline analyzes strong lensing in JWST data. First, the code forecasts galaxy-scale strong lenses for JWST, including Einstein radii down to θ_E=0.02″ systems (generalizable to other telescopes and selection criteria). The code also simulates images of galaxy-scale strong lenses for JWST in the range 0.02″&lt;θ_E&lt;1.5″. These simulations use the VELA hydrodynamical simulations as very realistic lensed source galaxies. CASSL employs empirically motivated parameter distributions to generate realistic datasets of simulated strong lensing images.
[ascl:1402.013] CASSIS: Interactive spectrum analysis
CASSIS (Centre d'Analyse Scientifique de Spectres Infrarouges et Submillimetriques), written in Java, is suited for broad-band spectral surveys to speed up the scientific analysis of high spectral resolution observations. It uses a local spectroscopic database made of the two molecular spectroscopic databases JPL and CDMS, as well as the atomic spectroscopic database NIST. Its tools include a LTE model and the RADEX (ascl:1010.075) model connected to the LAMDA (ascl:1010.077) molecular collisional database. CASSIS can build a line list fitting the various transitions of a given species and to directly produce rotational diagrams from these lists. CASSIS is fully integrated into <a href="http://ascl.net/1111.001">HIPE</a> (ascl:1111.001), the Herschel Interactive Processing Environment, as a plug-in.
[submitted] CASSIS-LTE-Python
CASSIS-LTE-Python is a Python module based on the same LTE algorithm as the one used in CASSIS (ascl:1402.013). Based on the user's description of an observed source, the module builds a function representing the LTE model of the source, and performs the minimization of this function with respect to the observed spectrum (if pointed mode) or spectra (if mapping mode). The source description consists in a series of components arranged in an onion-like structure ; each component is characterized by a size, a velocity, an excitation temperature, a line width and the column densities of the species present in the component. CASSIS-LTE-Python uses the non-linear least-squares minimization package lmfit (ascl:1606.014), so that the user can choose from any minimization algorithm available in lmfit. Other key features include : supporting observations made with multiple telescopes, using constraints on spectroscopic parameters for each species, fitting an entire spectrum or only selected frequency windows, taking into account a continuum in the model for continuum-free observations, possibility to use a spatial mask for data cubes.
[ascl:2009.005] CASI-3D: Convolutional Approach to Structure Identification-3D
CASI-3D identifies signatures of stellar feedback in molecular line spectra, such as 12CO and 13CO, using deep learning. The code is developed from CASI-2D (ascl:1905.023) and exploits the full 3D spectral information.
[ascl:1905.023] CASI-2D: Convolutional Approach to Shell Identification - 2D
CASI-2D (Convolutional Approach to Shell Identification) identifies stellar feedback signatures using data from magneto-hydrodynamic simulations of turbulent molecular clouds with embedded stellar sources and deep learning techniques. Specifically, a deep neural network is applied to dense regression and segmentation on simulated density and synthetic 12 CO observations to identify shells, sometimes referred to as "bubbles," and other structures of interest in molecular cloud data.
[ascl:1912.002] casacore: Suite of C++ libraries for radio astronomy data processing
The casacore package contains the core libraries of the old AIPS++/CASA (ascl:1107.013) package. This split was made to get a better separation of core libraries and applications. CASA is now built on top of Casacore. The system consists of a set of layered libraries (packages) and includes a library (using Boost-Python) that converts the basic Casacore types (e.g., Array, Record) to and from Python. Casacore includes the casa package for core functionality and data types like Array and Record; a scimath package for N-dim functions with auto-differentiation and linear or non-linear fitting; and a tables package for the table data system supporting N-dim arrays with advanced querying. It also includes the measures package to manage values in astronomical reference frames using physical units (Quanta) and the MeasurementSets for storing data in the UV-domain, and also the images package for N-dim images in world coordinates with various analysis operations.
[ascl:1107.013] CASA: Common Astronomy Software Applications
CASA, the Common Astronomy Software Applications package, is being developed with the primary goal of supporting the data post-processing needs of the next generation of radio astronomical telescopes such as ALMA and EVLA. The package can process both interferometric and single dish data. The CASA infrastructure consists of a set of C++ tools bundled together under an iPython interface as a set of data reduction tasks. This structure provides flexibility to process the data via task interface or as a python script. In addition to the data reduction tasks, many post-processing tools are available for even more flexibility and special purpose reduction needs.
[ascl:2207.025] casa_cube: Display and analyze astronomical data cubes
casa_cube provides an interface to data cubes generated by CASA (ascl:1107.013) or Gildas (ascl:1305.010). It performs simple tasks such as plotting given channel maps, moment maps, and line profile in various units, and also corrects for cloud extinction, reconvolves with a beam taper, and permits quick and easy comparisons with models.
[ascl:2103.031] CARTA: Cube Analysis and Rendering Tool for Astronomy
CARTA (Cube Analysis and Rendering Tool for Astronomy) is a image visualization and analysis tool designed for the ALMA, VLA, SKA pathfinders, and the ngVLA. If offers catalog support, shared region analytics, profile smoothing, and spectral line query, and more. CARTA adopts a client-server architecture suitable for visualizing images with large file sizes (GB to TB) easily obtained from ALMA, VLA, or SKA pathfinder observations; computation and data storage are handled by remote enterprise-class servers or clusters with high performance storage, while processed products are sent to clients only for visualization with modern web features, such as GPU-accelerated rendering. This architecture also enables users to interact with the ALMA and VLA science archives by using CARTA as an interface. CARTA provides a desktop version and a server version. The former is suitable for single-user usage with a laptop, a desktop, or a remote server in the "remote" execution mode. The latter is suitable for institution-wide deployment to support multiple users with user authentication and additional server-side features.
[ascl:2103.021] Carsus: Atomic database for astronomy
Carsus manages atomic datasets. It requires Chianti (ascl:9911.004), and can read data from a variety of sources and output them to file formats readable by radiative transfer codes such as TARDIS (ascl:1402.018).
[ascl:2005.007] Carpyncho: VVV Catalog browser toolkit
Carpyncho browses catalogs to search for and characterize time variable data of the Vista Variables in the Via Lactea (VVV) Survey. The stacked pawprint data from the Cambridge Astronomical Science Unit's (CASU) Vista Data Flow System (VDFS) v>= 1.3 catalogs have been crossed matched with the VDFS CASU v1.3 tile catalogs into Parquet files, allowing detection and classification of periodic variables within this dataset.
[ascl:1611.016] Carpet: Adaptive Mesh Refinement for the Cactus Framework
Carpet is an adaptive mesh refinement and multi-patch driver for the Cactus Framework (ascl:1102.013). Cactus is a software framework for solving time-dependent partial differential equations on block-structured grids, and Carpet acts as driver layer providing adaptive mesh refinement, multi-patch capability, as well as parallelization and efficient I/O.
[ascl:1404.009] carma_pack: MCMC sampler for Bayesian inference
carma_pack is an MCMC sampler for performing Bayesian inference on continuous time autoregressive moving average models. These models may be used to model time series with irregular sampling. The MCMC sampler utilizes an adaptive Metropolis algorithm combined with parallel tempering.
[ascl:1505.003] caret: Classification and Regression Training
caret (Classification And REgression Training) provides functions for training and plotting classification and regression models. It contains tools for data splitting, pre-processing, feature selection, model tuning using resampling, and variable importance estimation, as well as other functionality.
[ascl:2406.007] CARDiAC: Anisotropic Redshift Distributions in Angular Clustering
CARDiAC (Code for Anisotropic Redshift Distributions in Angular Clustering) computes the impact of anisotropic redshift distributions on a wide class of angular clustering observables. It supports auto- and cross-correlations of galaxy samples and cosmic shear maps, including galaxy-galaxy lensing. The anisotropy can be present in the mean redshift and/or width of Gaussian distributions, as well as in the fraction of galaxies in each component of multi-modal distributions. Templates of these variations can be provided by the user or simulated internally within the code.
[submitted] Carbonette: Spectral analysis pipeline
The automated infrared spectral-analysis tool Carbonette operates on publicly available Spitzer/IRS (IRSA Enhanced Products) spectra to identify candidate carbon nanotube (CNT) and hydrogenated nanotube (HNT) absorption features. It implements a multi-stage detection and validation framework combining signal-to-noise and Δχ² significance testing, equivalent-width and optical-depth metrics, contamination-avoidance masks (PAH/silicate), LSF resolution checks, and injection–recovery tests for robustness. Outputs include plots and machine-readable tables (CSV/JSON).
[ascl:2006.014] CARACal: Containerized Automated Radio Astronomy Calibration pipeline
CARACal (Containerized Automated Radio Astronomy Calibration, formerly MeerKATHI) reduces radio-interferometric data. Developed originally as an end-to-end continuum- and line imaging pipeline for MeerKAT, it can also be used with other radio telescopes. CARACal reduces large data sets and produces high-dynamic-range continuum images and spectroscopic data cubes. The pipeline is platform-independent and delivers imaging quality metrics to efficiently assess the data quality.
[ascl:2308.009] caput: Utilities for building radio astronomy data analysis pipelines
Caput (Cluster Astronomical Python Utilities) contains utilities for handling large datasets on computer clusters. Written with radio astronomy in mind, the package provides an infrastructure for building, managing and configuring pipelines for data processing. It includes modules for dynamically importing and utilizing mpi4py, in-memory mock-ups of h5py objects, and infrastructure for running data analysis pipelines on computer clusters. Caput features a generic container for holding self-documenting datasets in memory with straightforward syncing to h5py files, and offers specialization for holding time stream data. Caput also includes tools for MPI-parallel analysis and routines for converting between different time representations, dealing with leap seconds, and calculating celestial times.
[ascl:2011.002] CAPTURE: Interferometric pipeline for image creation from GMRT data
CAPTURE (CAsa Pipeline-cum-Toolkit for Upgraded Giant Metrewave Radio Telescope data REduction) produces continuum images from radio interferometric data. Written in Python, it uses CASA (ascl:1107.013) tasks to analyze data obtained by the GMRT. It can produce self-calibrated images in a fully automatic mode or can run in steps to allow the data to be inspected throughout processing.
[ascl:2507.023] Capivara: Scalable spectral-based segmentation package
Capivara implements a spectral-based segmentation method for Integral Field Unit (IFU) data cubes. The code uses hierarchical clustering in the spectral domain, grouping similar spectra to improve the signal-to-noise ratio without compromising astrophysical similarity among regions, and leverages advanced matrix operations via torch for GPU acceleration.
[ascl:1404.011] CAP_LOESS_1D & CAP_LOESS_2D: Recover mean trends from noisy data
CAP_LOESS_1D and CAP_LOESS_2D provide improved implementations of the one-dimensional (Clevelend 1979) and two-dimensional (Cleveland & Devlin 1988) Locally Weighted Regression (LOESS) methods to recover the mean trends of the population from noisy data in one or two dimensions. They include a robust approach to deal with outliers (bad data). The software is available in both IDL and Python versions.
[ascl:1106.017] CAOS: Code for Adaptive Optics Systems
The CAOS "system" (where CAOS stands for Code for Adaptive Optics Systems) is properly said a Problem Solving Environment (PSE). It is essentially composed of a graphical programming interface (the CAOS Application Builder) which can load different packages (set of modules). Current publicly distributed packages are the Software Package CAOS (the original adaptive optics package), the Software Package AIRY (an image-reconstruction-oriented package - AIRY stands for Astronomical Image Restoration with interferometrY), the Software Package PAOLAC (a simple CAOS interface for the analytic IDL code PAOLA developed by Laurent Jolissaint - PAOLAC stands for PAOLA within Caos), and a couple of private packages (not publicly distributed but restricted to the corresponding consortia): SPHERE (especially developed for the VLT planet finder SPHERE), and AIRY-LN (a specialized version of AIRY for the LBT instrument LINC-NIRVANA). Another package is also being developed: MAOS (that stands for Multiconjugate Adaptive Optics Simulations), developed for multi-reference multiconjugate AO studies purpose but still in a beta-version form.
[ascl:2406.004] candl: Differentiable likelihood framework for analyzing CMB power spectrum measurements
candl (CMB Analysis With A Differentiable Likelihood) analyzes CMB power spectrum measurements using a differentiable likelihood framework. It is compatible with JAX (ascl:2111.002), though JAX is optional, allowing for fast and easy computation of gradients and Hessians of the likelihoods, and candl provides interface tools for working with other cosmology software packages, including Cobaya (ascl:1910.019) and MontePython (ascl:1805.027). The package also provides auxiliary tools for common analysis tasks, such as generating mock data, and supports the analysis of primary CMB and lensing power spectrum data.
[ascl:1505.030] CANDID: Companion Analysis and Non-Detection in Interferometric Data
CANDID finds faint companion around star in interferometric data in the <a href="http://www.mrao.cam.ac.uk/research/optical-interferometry/oifits/">OIFITS</a> format. It allows systematically searching for faint companions in OIFITS data, and if not found, estimates the detection limit. The tool is based on model fitting and <a href="http://docs.scipy.org/doc/scipy-0.14.0/reference/generated/scipy.optimize.leastsq.html">Chi2 minimization</a>, with a grid for the starting points of the companion position. It ensures all positions are explored by estimating a-posteriori if the grid is dense enough, and provides an estimate of the optimum grid density.
[ascl:1502.015] Camelus: Counts of Amplified Mass Elevations from Lensing with Ultrafast Simulations
Camelus provides a prediction on weak lensing peak counts from input cosmological parameters. Written in C, it samples halos from a mass function and assigns a profile, carries out ray-tracing simulations, and then counts peaks from ray-tracing maps. The creation of the ray-tracing simulations requires less computing time than N-body runs and the results is in good agreement with full N-body simulations.
[ascl:1605.006] CAMELOT: Cloud Archive for MEtadata, Library and Online Toolkit
CAMELOT facilitates the comparison of observational data and simulations of molecular clouds and/or star-forming regions. The central component of CAMELOT is a database summarizing the properties of observational data and simulations in the literature through pertinent metadata. The core functionality allows users to upload metadata, search and visualize the contents of the database to find and match observations/simulations over any range of parameter space. To bridge the fundamental disconnect between inherently 2D observational data and 3D simulations, the code uses key physical properties that, in principle, are straightforward for both observers and simulators to measure — the surface density (Sigma), velocity dispersion (sigma) and radius (R). By determining these in a self-consistent way for all entries in the database, it should be possible to make robust comparisons.
[ascl:2506.024] CAMEL: Cosmological parameters estimator
CAMEL (Cosmological Analysis with Minuit Exploration of the Likelihood) performs cosmological parameters estimations using best fits, Monte-Carlo Markov Chains, and profile-likelihoods. Widely used in Planck satellite data analysis, by default it employs CLASS (ascl:1106.020) to compute all relevant cosmological quantities, but any other Boltzmann solver can easily be plugged in.
[ascl:1801.007] cambmag: Magnetic Fields in CAMB
cambmag is a modification to CAMB (ascl:1102.026) that calculates the compensated magnetic mode in the scalar, vector and tensor case. Previously CAMB included code only for the vectors. It also corrects for tight-coupling issues and adds in the ability to include massive neutrinos when calculating vector modes.
[ascl:1102.026] CAMB: Code for Anisotropies in the Microwave Background
We present a fully covariant and gauge-invariant calculation of the evolution of anisotropies in the cosmic microwave background (CMB) radiation. We use the physically appealing covariant approach to cosmological perturbations, which ensures that all variables are gauge-invariant and have a clear physical interpretation. We derive the complete set of frame-independent, linearised equations describing the (Boltzmann) evolution of anisotropy and inhomogeneity in an almost Friedmann-Robertson-Walker (FRW) cold dark matter (CDM) universe. These equations include the contributions of scalar, vector and tensor modes in a unified manner. Frame-independent equations for scalar and tensor perturbations, which are valid for any value of the background curvature, are obtained straightforwardly from the complete set of equations. We discuss the scalar equations in detail, including the integral solution and relation with the line of sight approach, analytic solutions in the early radiation dominated era, and the numerical solution in the standard CDM model. Our results confirm those obtained by other groups, who have worked carefully with non-covariant methods in specific gauges, but are derived here in a completely transparent fashion.
[ascl:1105.013] CAMB Sources: Number Counts, Lensing & Dark-age 21cm Power Spectra
We relate the observable number of sources per solid angle and redshift to the underlying proper source density and velocity, background evolution and line-of-sight potentials. We give an exact result in the case of linearized perturbations assuming general relativity. This consistently includes contributions of the source density perturbations and redshift distortions, magnification, radial displacement, and various additional linear terms that are small on sub-horizon scales. In addition we calculate the effect on observed luminosities, and hence the result for sources observed as a function of flux, including magnification bias and radial-displacement effects. We give the corresponding linear result for a magnitude-limited survey at low redshift, and discuss the angular power spectrum of the total count distribution. We also calculate the cross-correlation with the CMB polarization and temperature including Doppler source terms, magnification, redshift distortions and other velocity effects for the sources, and discuss why the contribution of redshift distortions is generally small. Finally we relate the result for source number counts to that for the brightness of line radiation, for example 21-cm radiation, from the sources.
[ascl:2207.015] calviacat: Calibrate star photometry by catalog comparison
calviacat calibrates star photometry by comparison to a catalog, including PanSTARRS 1, ATLAS-RefCat2, and SkyMapper catalogs. Catalog queries are cached so that subsequent calibrations of the same or similar fields can be more quickly executed.
[ascl:2301.001] CALSAGOS: Select cluster members and search, find, and identify substructures
CALSAGOS (Clustering ALgorithmS Applied to Galaxies in Overdense Systems) selects cluster members and searches, finds, and identifies substructures and galaxy groups in and around galaxy clusters using the redshift and position in the sky of the galaxies. The package offers two ways to determine cluster members, ISOMER and CLUMBERI. The ISOMER (Identifier of SpectrOscopic MembERs) function selects the spectroscopic cluster members by defining cluster members as those galaxies with a peculiar velocity lower than the escape velocity of the cluster. The CLUMBERI (CLUster MemBER Identifier) function select the cluster members using a 3D-Gaussian Mixture Modules (GMM). Both functions remove the field interlopers by using a 3-sigma clipping algorithm. CALSAGOS uses the function LAGASU (LAbeller of GAlaxies within SUbstructures) to search, find, and identify substructures and groups in and around a galaxy cluster; this function is based on clustering algorithms (GMM and DBSCAN), which search areas with high density to define a substructure or groups.
[ascl:2106.035] CalPriorSNIa: Effective calibration prior on the absolute magnitude of Type Ia supernovae
CalPriorSNIa quickly computes the effective calibration prior on the absolute magnitude <i>M<sub>B</sub></i> of Type Ia supernovae that corresponds to a given determination of <i>H<sub>0</sub></i>.
[ascl:1210.010] CALCLENS: Curved-sky grAvitational Lensing for Cosmological Light conE simulatioNS
CALCLENS, written in C and employing widely available software libraries, efficiently computes weak gravitational lensing shear signals from large N-body light cone simulations over a curved sky. The algorithm properly accounts for the sky curvature and boundary conditions, is able to produce redshift-dependent shear signals including corrections to the Born approximation by using multiple-plane ray tracing, and properly computes the lensed images of source galaxies in the light cone. The key feature of this algorithm is a new, computationally efficient Poisson solver for the sphere that combines spherical harmonic transform and multgrid methods. As a result, large areas of sky (~10,000 square degrees) can be ray traced efficiently at high-resolution using only a few hundred cores on widely available machines. Coupled with realistic galaxy populations placed in large N-body light cone simulations, CALCLENS is ideally suited for the construction of synthetic weak lensing shear catalogs to be used to test for systematic effects in data analysis procedures for upcoming large-area sky surveys.
[ascl:1505.001] CALCEPH: Planetary ephemeris files access code
CALCEPH accesses binary planetary ephemeris files, including INPOPxx, JPL DExxx ,and SPICE ephemeris files. It provides a C Application Programming Interface (API) and, optionally, a Fortran 77 or 2003 interface to be called by the application. Two groups of functions enable the access to the ephemeris files, single file access functions, provided to make transition easier from the JPL functions, such as PLEPH, to this library, and many ephemeris file at the same time. Although computers have different endianess (order in which integers are stored as bytes in computer memory), CALCEPH can handles the binary ephemeris files with any endianess by automatically swaps the bytes when it performs read operations on the ephemeris file.
[ascl:2501.001] CAFE: Continuum And Feature Extraction tool
CAFE (Continuum And Feature Extraction) fits JWST IFU data; the code is a Python version of the original CAFE IDL software for fitting Spitzer/IRS spectra. The code contains two main tools: (1) the CAFE Region Extraction Tool Automaton (CRETA) and (2) the CAFE spectral fitting tool, or fitter. CRETA performs single-position and full-grid extractions from JWST IFU datasets; that is, from pipeline-processed cubes obtained with the NIRSpec IFU and MIRI MRS instruments. The CAFE fitter uses the spectra extracted by CRETA (or spectra provided by the user) and performs a spectral decomposition of the continuum emission (stellar and/or dust), as well as of a variety of common spectral features (in emission and absorption) present in the near- and mid-IR spectra of galaxies, including prominent, broad emission from small grains and molecules such as Polycyclic Aromatic Hydrocarbons (PAHs). The full dust treatment (size and composition) performed by CAFE allows the dust continuum model components to fit not only spectra from typical star-forming galaxies, but also those from more extreme, heavily dust-obscured starburst galaxies, such as luminous infrared galaxies (LIRGs and ULIRGs), active galactic nuclei (AGN), or very luminous quasars.
[ascl:1807.015] CAESAR: Compact And Extended Source Automated Recognition
CAESAR extracts and parameterizes both compact and extended sources from astronomical radio interferometric maps. The processing pipeline is a series of stages that can run on multiple cores and processors. After local background and rms map computation, compact sources are extracted with flood-fill and blob finder algorithms, processed (selection + deblending), and fitted using a 2D gaussian mixture model. Extended source search is based on a pre-filtering stage, allowing image denoising, compact source removal and enhancement of diffuse emission, followed by a final segmentation. Different algorithms are available for image filtering and segmentation. The outputs delivered to the user include source fitted and shape parameters, regions and contours. Written in C++, CAESAR is designed to handle the large-scale surveys planned with the Square Kilometer Array (SKA) and its precursors.
[ascl:2108.009] caesar-rest: Web service for the caesar source extractor
caesar-rest is a REST-ful web service for astronomical source extraction and classification with the caesar source extractor [ascl:1807.015]. The software is developed in python and consists of containerized microservices, deployable on standalone servers or on a distributed cloud infrastructure. The core component is the REST web application, based on the Flask framework and providing APIs for managing the input data (e.g. data upload/download/removal) and source finding jobs (e.g. submit, get status, get outputs) with different job management systems (Kubernetes, Slurm, Celery). Additional services (AAI, user DB, log storage, job monitor, accounting) enable the user authentication, the storage and retrieval of user data and job information, the monitoring of submitted jobs, and the aggregation of service logs and user data/job stats.
[ascl:1303.017] CADRE: CArma Data REduction pipeline
CADRE, the Combined Array for Millimeter-wave Astronomy (CARMA) data reduction pipeline, gives investigators a first look at a fully reduced set of their data. It runs automatically on all data produced by the telescope as they arrive in the data archive. The pipeline is written in python and uses python wrappers for MIRIAD subroutines for direct access to the data. It applies passband, gain and flux calibration to the data sets and produces a set of continuum and spectral line maps in both MIRIAD and FITS format.
[ascl:2306.037] CADET: X-ray cavity detection tool
The machine learning pipeline CADET (CAvity DEtection Tool) finds and size-estimates arbitrary surface brightness depressions (X-ray cavities) on noisy Chandra images of galaxies. The pipeline is a self-standing Python script and inputs either raw Chandra images in units of counts (numbers of captured photons) or normalized background-subtracted and/or exposure-corrected images. CADET saves corresponding pixel-wise as well as decomposed cavity predictions in FITS format and also preserves the WCS coordinates; it also outputs a PNG file showing decomposed predictions for individual scales.
[ascl:1102.013] Cactus: HPC infrastructure and programming tools
Cactus provides computational scientists and engineers with a collaborative, modular and portable programming environment for parallel high performance computing. Cactus can make use of many other technologies for HPC, such as Samrai, HDF5, PETSc and PAPI, and several application domains such as numerical relativity, computational fluid dynamics and quantum gravity are developing open community toolkits for Cactus.
[ascl:1610.006] C3: Command-line Catalogue Crossmatch for modern astronomical surveys
The Command-line Catalogue Cross-matching (C<sup>3</sup>) software efficiently performs the positional cross-match between massive catalogues from modern astronomical surveys, whose size have rapidly increased in the current data-driven science era. Based on a multi-core parallel processing paradigm, it is executed as a stand-alone command-line process or integrated within any generic data reduction/analysis pipeline. C<sup>3</sup> provides its users with flexibility in portability, parameter configuration, catalogue formats, angular resolution, region shapes, coordinate units and cross-matching types.
[ascl:2312.024] C2-Ray3Dm1D_Helium: Hydrogen + helium version of C2-Ray
C2-Ray3Dm1D_Helium is the hydrogen + helium version of the radiative transfer photo-ionization code C<sup>2</sup>-Ray. It combines the 1D and 3D versions of the code.
[ascl:2312.023] C2-Ray3Dm: 3D version of C2-Ray for multiple sources, hydrogen only
C<sup>2</sup>-Ray3Dm performs time-dependent photo-ionization calculations for 3D multiple sources, and for hydrogen only. Based on C<sup>2</sup>-Ray (ascl:2312.022), it runs under both MPI and OpenMP. The length of subroutines has been reduced to make the code more manageable and easier to read.
[ascl:2312.022] C2-Ray: Time-dependent photo-ionization calculations
C<sup>2</sup>-Ray calculates spherical symmetric time-dependent photo-ionization in 1D with the source at the origin for hydrogen only. The code is explicitly photon-conserving and uses an analytical relaxation solution for the ionization rate equations for each time step, thus enabling integration of the equation of transfer along a ray with fewer cells and time steps than previous methods. It is suitable for coupling radiative transfer to gas and N-body dynamics methods on fixed or adaptive grids. C<sup>2</sup>-Ray is not parallelized but contains an MPI module for compatibility with the 3D version (C<sup>2</sup>-Ray3Dm).
[ascl:1211.005] C-m Emu: Concentration-mass relation emulator
The concentration-mass relation for dark matter-dominated halos is one of the essential results expected from a theory of structure formation. C-m Emu is a simple numerical code for the c-M relation as a function of cosmological parameters for wCDM models generates the best-fit power-law model for each redshift separately and then interpolate between the redshifts. This produces a more accurate answer at each redshift at the minimal cost of running a fast code for every c -M prediction instead of using one fitting formula. The emulator is constructed from 37 individual models, with three nested N-body gravity-only simulations carried out for each model. The mass range covered by the emulator is 2 x 10^{12} M_sun &lt; M &lt;10^{15} M_sun with a corresponding redshift range of z=0 -1. Over this range of mass and redshift, as well as the variation of cosmological parameters studied, the mean halo concentration varies from c ~ 2 to c ~ 8. The distribution of the concentration at fixed mass is Gaussian with a standard deviation of one-third of the mean value, almost independent of cosmology, mass, and redshift over the ranges probed by the simulations.
[ascl:1610.011] BXA: Bayesian X-ray Analysis
BXA connects the nested sampling algorithm <a href="https://ascl.net/1109.006">MultiNest</a> (ascl:1109.006) to the X-ray spectral analysis environments <a href="https://ascl.net/9910.005">Xspec</a> (ascl:9910.005) and <a href="https://ascl.net/1107.005">Sherpa</a> (ascl:1107.005) for Bayesian parameter estimation and model comparison. It provides parameter estimation in arbitrary dimensions and plotting of spectral model vs. the data for best fit, posterior samples, or each component. BXA allows for model selection; it computes the evidence for the considered model, ready for use in computing Bayes factors and is not limited to nested models. It also visualizes deviations between model and data with Quantile-Quantile (QQ) plots, which do not require binning and are more comprehensive than residuals.
[ascl:1806.026] BWED: Brane-world extra dimensions
Braneworld-extra-dimensions places constraints on the size of the AdS5 radius of curvature within the Randall-Sundrum brane-world model in light of the near-simultaneous detection of the gravitational wave event GW170817 and its optical counterpart, the short γ-ray burst event GRB170817A. The code requires a (supplied) patch to the Montepython cosmological MCMC sampler (ascl:1805.027) to sample the posterior distribution of the 4-dimensional parameter space in VBV17 and obtain constraints on the parameters.
[ascl:2306.030] Butterpy: Stellar butterfly diagram and rotational light curve simulator
Butterpy simulates star spot emergence, evolution, decay, and stellar rotational light curves. It tests the recovery of stellar rotation periods using different frequency analysis techniques. Butterpy can simulate light curves of stars with variable activity level, rotation period, spot lifetime, magnetic cycle duration and overlap, spot emergence latitudes, and latitudinal differential rotation shear.
[ascl:1610.010] BurnMan: Lower mantle mineral physics toolkit
BurnMan determines seismic velocities for the lower mantle. Written in Python, BurnMan calculates the isotropic thermoelastic moduli by solving the equations-of-state for a mixture of minerals defined by the user. The user may select from a list of minerals applicable to the lower mantle included or can define one. BurnMan provides choices in methodology, both for the EoS and for the multiphase averaging scheme and the results can be visually or quantitatively compared to observed seismic models.
[ascl:2212.024] Burning Arrow: Black hole massive particles orbit degradation
Burning Arrow determines the destabilization of massive particle circular orbits due to thermal radiation, emitted in X-ray, from the hot accretion disk material. This code requires the radiation forces exerted on the material at the point of interest found by running the code Infinity (ascl:2212.021). Burning Arrow begins by assuming a target particle in the disk that moves in a circular orbit. It then introduces the recorded radiation forces from Infinity code for the target region. The forces are subsequently introduced into the target particle equations of motion and the trajectory is recalculated. Burning Arrow then produces images of the black hole - accretion disk system that includes the degenerated particle trajectories that obey the assorted velocity profiles.
[ascl:2312.003] BUQO: Bayesian Uncertainty Quantification by Optimization
BUQO solves large-scale imaging inverse problems. It leverages probability concentration phenomena and the underlying convex geometry to formulate the Bayesian hypothesis test as a convex problem that is then efficiently solved by using scalable optimization algorithms. This allows scaling to high-resolution and high-sensitivity imaging problems that are computationally unaffordable for other Bayesian computation approaches.
[ascl:1204.003] BUDDA: BUlge/Disk Decomposition Analysis
Budda is a Fortran code developed to perform a detailed structural analysis on galaxy images. It is simple to use and gives reliable estimates of the galaxy structural parameters, which can be used, for instance, in Fundamental Plane studies. Moreover, it has a powerful ability to reveal hidden sub-structures, like inner disks, secondary bars and nuclear rings.
[ascl:2403.004] BTSbot: Automated identification of supernovae with multi-modal deep learning
BTSbot automates real-time identification of bright extragalactic transients in Zwicky Transient Facility (ZTF) data. A multi-modal convolutional neural network, BTSbot provides a bright transient score to individual ZTF detections using their image data and 25 extracted features. The package eliminates the need for daily visual inspection of new transients by automatically identifying and requesting spectroscopic follow-up observations of new bright transient candidates. BTSbot recovers all bright transients in our test split and performs on par with human experts in terms of identification speed (on average, ∼1 hour quicker than scanners).
[ascl:2001.007] BTS: Behind The Spectrum
Behind The Spectrum (BTS) is a fully-automated multiple-component fitter for optically-thin spectra. Written as a python module, the routine uses the first, second and third derivatives to determine thenumber of components in the spectrum. A least-squared fitting routine then determines the best fit with that number of components, checking for over-fitting and over-lapping velocity centroids.
[ascl:2503.032] BT: Blooming Tree hierarchical structure analysis
The Blooming Tree (BT) algorithm identifies clusters, groups, and substructures from galaxy redshift surveys. Based on the hierarchical clustering method, it takes the projected binding energy as the linking length and provides three main analysis approaches to trim the hierarchical tree; 1.) the direct trimming (binding energy, velocity disperion, or eta); 2.) the sigma plateau, when no trimming threshold is appointed; and 3.) the blooming tree. The Blooming Tree algorithm tool works only in the terminal.
[ascl:2309.015] bskit: Bispectra from cosmological simulation snapshots
bskit, built upon the nbodykit (ascl:1904.027) simulation analysis package, measures density bispectra from snapshots of cosmological N-body or hydrodynamical simulations. It can measure auto or cross bispectra in a user-specified set of triangle bins (that is, triplets of 3-vector wavenumbers). Several common sets of bins are also implemented, including all triangle bins for specified k_min and k_max, equilateral triangles between specified k_min and k_max, isosceles triangles, and squeezed isosceles triangles.
[ascl:9904.001] BSGMODEL: The Bahcall-Soneira Galaxy Model
BSGMODEL is used to construct the disk and spheroid components of the Galaxy from which the distribution of visible stars and mass in the Galaxy is calculated. The computer files accessible here are available for export use. The modifications are described in comment lines in the software. The Galaxy model software has been installed and used by different people for a large variety of purposes (see, e. g., the the review "Star Counts and Galactic Structure'', Ann. Rev. Astron. Ap. 24, 577, 1986 ).
[ascl:1303.014] BSE: Binary Star Evolution
BSE is a rapid binary star evolution code. It can model circularization of eccentric orbits and synchronization of stellar rotation with the orbital motion owing to tidal interaction in detail. Angular momentum loss mechanisms, such as gravitational radiation and magnetic braking, are also modelled. Wind accretion, where the secondary may accrete some of the material lost from the primary in a wind, is allowed with the necessary adjustments made to the orbital parameters in the event of any mass variations. Mass transfer occurs if either star fills its Roche lobe and may proceed on a nuclear, thermal or dynamical time-scale. In the latter regime, the radius of the primary increases in response to mass-loss at a faster rate than the Roche-lobe of the star. Prescriptions to determine the type and rate of mass transfer, the response of the secondary to accretion and the outcome of any merger events are in place in BSE.
[ascl:2411.012] BSAVI: Bayesian SAmple VIsualizer for cosmological likelihoods
BSAVI (Bayesian Sample Visualizer) aids likelihood analysis of model parameters where samples from a distribution in the parameter space are used as inputs to calculate a given observable. For example, selecting a range of samples will allow you to easily see how the observables change as you traverse the sample distribution. At the core of BSAVI is the Observable object, which contains the data for a given observable and instructions for plotting it. It is modular, so you can write your own function that takes the parameter values as inputs, and BSAVI will use it to compute observables on the fly. It also accepts tabular data, so if you have pre-computed observables, simply import them alongside the dataset containing the sample distribution to start visualizing. Though BSAVI was developed for use in theoretical cosmology, it can be customized to fit a wide range of visualization needs.
[ascl:1903.004] brutifus: Python module to post-process datacubes from integral field spectrographs
brutifus aids in post-processing datacubes from integral field spectrographs. The set of Python routines in the package handle generic tasks, such as the registration of a datacube WCS solution with the Gaia catalogue, the correction of Galactic reddening, or the subtraction of the nebular/stellar continuum on a spaxel-per-spaxel basis, with as little user interactions as possible. brutifus is modular, in that the order in which the post-processing routines are run is entirely customizable.
[ascl:1407.016] Brut: Automatic bubble classifier
Brut, written in Python, identifies bubbles in infrared images of the Galactic midplane; it uses a database of known bubbles from the Milky Way Project and Spitzer images to build an automatic bubble classifier. The classifier is based on the Random Forest algorithm, and uses the WiseRF implementation of this algorithm.
[ascl:1412.005] BRUCE/KYLIE: Pulsating star spectra synthesizer
BRUCE and KYLIE, written in Fortran 77, synthesize the spectra of pulsating stars. BRUCE constructs a point-sampled model for the surface of a rotating, gravity-darkened star, and then subjects this model to perturbations arising from one or more non-radial pulsation modes. Departures from adiabaticity can be taken into account, as can the Coriolis force through adoption of the so-called traditional approximation. BRUCE writes out a time-sequence of perturbed surface models. This sequence is read in by KYLIE, which synthesizes disk-integrated spectra for the models by co-adding the specific intensity emanating from each visible point toward the observer. The specific intensity is calculated by interpolation in a large temperature-gravity-wavelength-angle grid of pre-calculated intensity spectra.
[ascl:2305.009] breizorro: Image masking tool
Given a FITS image, breizorro creates a binary mask. The software allows the user control various parameters and functions, such as setting a sigma threshold for masking, merging in or subtracting one or more masks or region files, filling holes, applying dilation within a defined radius of pixels, and inverting the mask.
[ascl:2501.009] breads: Broad Repository for Exoplanet Analysis, Discovery, and Spectroscopy
breads (Broad Repository for Exoplanet Analysis, Discovery, and Spectroscopy) provides a toolkit for data analyses in astronomical spectroscopy of exoplanets, in particular frameworks for rigorous forward modeling of observational data to achieve physical inferences with reduced systematic biases. Users choose a data class, a forward model function, and a fitting strategy. Data classes normalize the data format, simplifying reduction across different spectrographs while allowing for specific behaviors of each instrument to also be coded into their own specific class. breads provides specific functionality for modeling data from JWST NIRSpec, Keck OSIRIS, and Keck KPIC, but the underlying mathematical framework is more general.
[ascl:1806.025] BRATS: Broadband Radio Astronomy ToolS
BRATS (Broadband Radio Astronomy ToolS) provides tools for the spectral analysis of broad-bandwidth radio data and legacy support for narrowband telescopes. It can fit models of spectral ageing on small spatial scales, offers automatic selection of regions based on user parameters (e.g. signal to noise), and automatic determination of the best-fitting injection index. It includes statistical testing, including Chi-squared, error maps, confidence levels and binning of model fits, and can map spectral index as a function of position. It also provides the ability to reconstruct sources at any frequency for a given model and parameter set, subtract any two FITS images and output residual maps, easily combine and scale FITS images in the image plane, and resize radio maps.
[ascl:2601.013] BRAINS: BLR Reverberation-mapping Analysis In AGNs with Nested Sampling
BRAINS (BLR Reverberation-mapping Analysis In AGNs with Nested Sampling) dynamically models the broad-line regions of active galactic nuclei using reverberation-mapping and spectro-astrometric observations. It couples flexible geometric and kinematic BLR models with radiative transfer and line-response prescriptions to reproduce observed emission-line light curves, spectra, and spatial signatures. The code employs nested sampling to infer BLR structure and black hole mass, providing posterior distributions for physical and nuisance parameters and enabling rigorous model comparison. Implemented in C and Python, BRAINS supports configurable data sets, model components, and priors, and includes utilities for data preparation, parameter estimation, and visualization of inferred BLR and black hole properties.

Would you like to view a random code?