Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Snakemake aims to reduce the complexity of creating workflows by providing a clean and modern domain specific specification language (DSL) in Python style, together with a fast and comfortable execution environment.
The TDDA Python module provides command-line and Python API support for the overall process of data analysis, through tools that perform reference testing, constraint discovery for data, automatic inference of regular expressions from text data and automatic test generation.
A LEMS simulator written in Python which can be used to run NeuroML2 models.
This package provides miscellaneous tools for data analysis and scientific computing.
Dvc objects provides a filesystem and object-db level abstractions to use in dvc and dvc-data.
vedo is a fast and lightweight python module for scientific analysis and visualization. The package provides a wide range of functionalities for working with three-dimensional meshes and point clouds. It can also be used to generate high quality two-dimensional renderings such as scatter plots and histograms. vedo is based on vtk and numpy.
Pandas is a Python package providing fast, flexible, and expressive data structures designed to make working with structured (tabular, multidimensional, potentially heterogeneous) and time series data both easy and intuitive. It aims to be the fundamental high-level building block for doing practical, real world data analysis in Python.
nibabel is a library that provides read and write access to common neuroimaging file formats, including: ANALYZE (plain, SPM99, SPM2 and later), GIFTI, NIfTI1, NIfTI2, CIFTI-2, MINC1, MINC2, AFNI BRIK/HEAD, ECAT and Philips PAR/REC. In addition, NiBabel also supports FreeSurfer’s MGH, geometry, annotation and morphometry files, and provides some limited support for DICOM.
This package implements a functionality to solve automatic numerical differentiation problems in one or more variables. Finite differences are used in an adaptive manner, coupled with a Richardson extrapolation methodology to provide a maximally accurate result. The user can configure many options like; changing the order of the method or the extrapolation, even allowing the user to specify whether complex-step, central, forward or backward differences are used.
This is the Python package for ECOS: Embedded Cone Solver. ECOS is numerical software for solving convex second-order cone programs (SOCPs).
DecayLanguage implements a language to describe and convert particle decays between digital representations, effectively making it possible to interoperate several fitting programs. Particular interest is given to programs dedicated to amplitude analyses.
python-pydicom is a Python library for reading and writing DICOM medical imaging data. It can read, modify and write DICOM data.
This package provides a Python library for manipulating indices of ndarrays.
Vaex is a high performance Python library for lazy Out-of-Core DataFrames (similar to Pandas), to visualize and explore big tabular datasets. This package provides the core modules of Vaex.
Plotly's Python graphing library makes interactive,publication-quality graphs online. Examples of how to make line plots, scatter plots, area charts, bar charts, error bars, box plots, histograms, heatmaps, subplots, multiple-axes, polar charts, and bubble charts.
Scikit-Optimize, or skopt, is a simple and efficient library to minimize (very) expensive and noisy black-box functions. It implements several methods for sequential model-based optimization. skopt aims to be accessible and easy to use in many contexts.
This package provides tools to easily search and download French data from INSEE and IGN APIs. This data includes more than 150 000 macroeconomic series, a dozen datasets of local french data, numerous sources available on insee.fr, geographical limits of administrative areas taken from IGN as well as key metadata and SIRENE database containing data on all French compagnies.
This package provides encoding and decoding routines that enable the serialization and deserialization of numerical and array data types provided by numpy using the highly efficient msgpack format. Serialization of Python's native complex data types is also supported.
This is a Python package to compute statistical test and add statistical annotations on an existing boxplots and barplots generated by seaborn.
This package provides accelerated simulations and potentials of solids.
Xarray (formerly xray) makes working with labelled multi-dimensional arrays simple, efficient, and fun!
Xarray introduces labels in the form of dimensions, coordinates and attributes on top of raw NumPy-like arrays, which allows for a more intuitive, more concise, and less error-prone developer experience. The package includes a large and growing library of domain-agnostic functions for advanced analytics and visualization with these data structures.
This package implements a functionality to work with Nested sampling, a popular numerical method for Bayesian computation, which simultaneously generates samples from the posterior distribution and an estimate of the Bayesian evidence for a given likelihood and prior. nestcheck provides Python utilities for analysing samples produced by nested sampling, and estimating uncertainties on nested sampling calculations (which have different statistical properties to calculations using other numerical methods).
SALib provides tools for global sensitivity analysis. It contains Sobol', Morris, FAST, DGSM, PAWN, HDMR, Moment Independent and fractional factorial methods.
hepunits collects the most commonly used units and constants in the HEP System of Units, as derived from the basic units originally defined by the CLHEP project.