Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
pynetdicom is a Python package that implements the DICOM networking protocol. It allows the easy creation of DICOM SCUs and SCPs.
This package provides tools to easily search and download French data from INSEE and IGN APIs. This data includes more than 150 000 macroeconomic series, a dozen datasets of local french data, numerous sources available on insee.fr, geographical limits of administrative areas taken from IGN as well as key metadata and SIRENE database containing data on all French compagnies.
Pyzo is a Python IDE focused on interactivity and introspection,which makes it very suitable for scientific computing. Its practical design is aimed at simplicity and efficiency.
It consists of two main components, the editor and the shell, and uses a set of pluggable tools to help the programmer in various ways. Some example tools are source structure, project manager, interactive help, workspace...
This package contains public type stubs for python-pandas, following the convention of providing stubs in a separate package, as specified in PEP 561. The stubs cover the most typical use cases of python-pandas. In general, these stubs are narrower than what is possibly allowed by python-pandas, but follow a convention of suggesting best recommended practices for using python-pandas.
An efficient Python implementation of the Apriori algorithm, which uncovers hidden structures in categorical data
This package provides a Python library for calculating Evapotranspiration using various standard methods.
nibabel is a library that provides read and write access to common neuroimaging file formats, including: ANALYZE (plain, SPM99, SPM2 and later), GIFTI, NIfTI1, NIfTI2, CIFTI-2, MINC1, MINC2, AFNI BRIK/HEAD, ECAT and Philips PAR/REC. In addition, NiBabel also supports FreeSurfer’s MGH, geometry, annotation and morphometry files, and provides some limited support for DICOM.
This package provides accelerated simulations and potentials of solids.
PyZX is a Python tool implementing the theory of ZX-calculus for the creation, visualisation, and automated rewriting of large-scale quantum circuits. PyZX currently allows you to:
Read in quantum circuits in the file format of QASM, Quipper or Quantomatic;
Rewrite circuits into a pseudo-normal form using the ZX-calculus;
Extract new simplified circuits from these reduced graphs;
Visualise the ZX-graphs and rewrites using either Matplotlib, Quantomatic or as a TikZ file for use in LaTeX documents;
Output the optimised circuits in QASM, QC or QUIPPER format.
Dvc data is DVC's data management subsystem.
Snakemake aims to reduce the complexity of creating workflows by providing a clean and modern domain specific specification language (DSL) in Python style, together with a fast and comfortable execution environment.
pykdtree is a kd-tree implementation for fast nearest neighbour search in Python.
Snakemake aims to reduce the complexity of creating workflows by providing a clean and modern domain specific specification language (DSL) in Python style, together with a fast and comfortable execution environment.
This package provides an efficient implementation of Friedman's SuperSmoother based in Python. It makes use of numpy for fast numerical computation.
This package provides functionality to make it easy to make scatter density maps, both for interactive and non-interactive use.
This package provides a stable interface for interactions between Snakemake and its executor plugins.
QuTiP is a library for simulating the dynamics of closed and open quantum systems. It aims to provide numerical simulations of a wide variety of quantum mechanical problems, including those with Hamiltonians and/or collapse operators with arbitrary time-dependence, commonly found in a wide range of physics applications.
Xarray (formerly xray) makes working with labelled multi-dimensional arrays simple, efficient, and fun!
Xarray introduces labels in the form of dimensions, coordinates and attributes on top of raw NumPy-like arrays, which allows for a more intuitive, more concise, and less error-prone developer experience. The package includes a large and growing library of domain-agnostic functions for advanced analytics and visualization with these data structures.
This package provides a Python library for manipulating indices of ndarrays.
spin is a simple interface for common development tasks. It comes with a few common build commands out the box, but can easily be customized per project.
The impetus behind developing the tool was the mass migration of scientific Python libraries (SciPy, scikit-image, and NumPy, etc.) to Meson, after distutils was deprecated. When many of the build and installation commands changed, it made sense to abstract away the nuisance of having to re-learn them.
Pandas is a Python package providing fast, flexible, and expressive data structures designed to make working with structured (tabular, multidimensional, potentially heterogeneous) and time series data both easy and intuitive. It aims to be the fundamental high-level building block for doing practical, real world data analysis in Python.
PyMCubes is an implementation of the marching cubes algorithm to extract iso-surfaces from volumetric data. The volumetric data can be given as a three-dimensional NumPy array or as a Python function f(x, y, z).
Snakemake aims to reduce the complexity of creating workflows by providing a clean and modern domain specific specification language (DSL) in Python style, together with a fast and comfortable execution environment.
Dask is a flexible parallel computing library for analytics. It consists of two components: dynamic task scheduling optimized for computation, and large data collections like parallel arrays, dataframes, and lists that extend common interfaces like NumPy, Pandas, or Python iterators to larger-than-memory or distributed environments. These parallel collections run on top of the dynamic task schedulers.