Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Pandas is a Python package providing fast, flexible, and expressive data structures designed to make working with structured (tabular, multidimensional, potentially heterogeneous) and time series data both easy and intuitive. It aims to be the fundamental high-level building block for doing practical, real world data analysis in Python.
This package provides fast computations for principal component analysis (PCA), SVD, and eigendecompositions via randomized methods
Baycomp is a library for Bayesian comparison of classifiers. Functions in the library compare two classifiers on one or on multiple data sets. They compute three probabilities: the probability that the first classifier has higher scores than the second, the probability that differences are within the region of practical equivalence (rope), or that the second classifier has higher scores.
Scikit-survival is a Python module for survival analysis built on top of scikit-learn. It allows doing survival analysis while utilizing the power of scikit-learn, e.g., for pre-processing or doing cross-validation.
PyMCubes is an implementation of the marching cubes algorithm to extract iso-surfaces from volumetric data. The volumetric data can be given as a three-dimensional NumPy array or as a Python function f(x, y, z).
Dvc objects provides a filesystem and object-db level abstractions to use in dvc and dvc-data.
PyVista is...
Pythonic VTK: a high-level API to the Visualization Toolkit (VTK);
mesh data structures and filtering methods for spatial datasets;
3D plotting made simple and built for large/complex data geometries.
This package provides a Pythonic, well-documented interface exposing VTK's powerful visualization backend to facilitate rapid prototyping, analysis, and visual integration of spatially referenced datasets.
An efficient Python implementation of the Apriori algorithm, which uncovers hidden structures in categorical data
Snakemake aims to reduce the complexity of creating workflows by providing a clean and modern domain specific specification language (DSL) in Python style, together with a fast and comfortable execution environment.
Hist is an analyst-friendly front-end for boost-histogram.
This package provides a Python library for manipulating indices of ndarrays.
The SciPy library is one of the core packages that make up the SciPy stack. It provides many user-friendly and efficient numerical routines such as routines for numerical integration and optimization.
This package provides common functions and classes for Snakemake and its plugins.
This package provides a stable interface for interactions between Snakemake and its software deployment plugins.
Anndata is a package for simple (functional) high-level APIs for data analysis pipelines. In this context, it provides an efficient, scalable way of keeping track of data together with learned annotations and reduces the code overhead typically encountered when using a mostly object-oriented library such as scikit-learn.
Trimesh is a pure Python library for loading and using triangular meshes with an emphasis on watertight surfaces. The goal of the library is to provide a full featured and well tested Trimesh object which allows for easy manipulation and analysis, in the style of the Polygon object in the Shapely library.
This package provides a simple and easy-to-use PID controller.
This package provides a framework for building scientific applications. It aims to bring state of the art software design practices to scientific computing, with the goal of providing a strong skeleton on which to build scientific codes by steering the implementation towards usability and maintainability.
This package implements a functionality to tell whether two images look nearly identical. The image hash algorithms (average, perceptual, difference, wavelet) analyse the image structure on luminance (without color information). The color hash algorithm analyses the color distribution and black & gray fractions (without position information).
Features:
average hashing
perceptual hashing
difference hashing
wavelet hashing
HSV color hashing (colorhash)
crop-resistant hashing
This package provides encoding and decoding routines that enable the serialization and deserialization of numerical and array data types provided by numpy using the highly efficient msgpack format. Serialization of Python's native complex data types is also supported.
Modin uses Ray or Dask to provide an effortless way to speed up your pandas notebooks, scripts, and libraries. Unlike other distributed DataFrame libraries, Modin provides seamless integration and compatibility with existing pandas code.
This is a package meant primarily for documenting histogram indexing and the PlottableHistogram Protocol and any future cross-library standards. It also contains the code for the PlottableHistogram Protocol, to be used in type checking libraries wanting to conform to the protocol. It is not usually a runtime dependency, but only a type checking, testing, and/or docs dependency in support of other libraries.
This package provides utilities and tools for open data science including tools for accessing data sets in Python.
Thi package implements a functionality for mean-preserving interpolation of 1D data (for example, time series) with splines.