Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Datatree is a prototype implementation of a tree-like hierarchical data structure for xarray. Datatree is in the process of being merged upstream into xarray.
This package provides accelerated simulations and potentials of solids.
This package provides encoding and decoding routines that enable the serialization and deserialization of numerical and array data types provided by numpy using the highly efficient msgpack format. Serialization of Python's native complex data types is also supported.
pandarallel allows any Pandas user to take advantage of their multi-core computer, while Pandas uses only one core. pandarallel also offers nice progress bars (available on Notebook and terminal) to get an rough idea of the remaining amount of computation to be done.
Scikit-image is a collection of algorithms for image processing.
This package provides a Python interface to the QDLDL LDL factorization routine for quasi-definite linear system.
vedo is a fast and lightweight python module for scientific analysis and visualization. The package provides a wide range of functionalities for working with three-dimensional meshes and point clouds. It can also be used to generate high quality two-dimensional renderings such as scatter plots and histograms. vedo is based on vtk and numpy.
This package provides a simplified scipy.signal.spectral module to do spectral analysis in Python.
PyQtGraph is a Pure-python graphics library for PyQt5, PyQt6, PySide2 and PySide6. It is intended for use in mathematics, scientific or engineering applications.
Dask is a flexible parallel computing library for analytics. It consists of two components: dynamic task scheduling optimized for computation, and large data collections like parallel arrays, dataframes, and lists that extend common interfaces like NumPy, Pandas, or Python iterators to larger-than-memory or distributed environments. These parallel collections run on top of the dynamic task schedulers.
pyfma provides an implementation of fused multiply-add which computes (x*y) + z with a single rounding. This is useful for dot products, matrix multiplications, polynomial evaluations (e.g., with Horner's rule), Newton's method for evaluating functions, convolutions, artificial neural networks etc.
Dask.distributed is a lightweight library for distributed computing in Python. It extends both the concurrent.futures and dask APIs to moderate sized clusters.
This package provides a Python library for building and analyzing recommender systems that deal with explicit rating data. It was designed with the following purposes in mind:
Provide tools to handle downloaded or user-provided datasets.
Provide ready-to-use prediction algorithms and similarity measures.
Provide a base for creating custom algorithms.
Provide tools to evaluate, analyse and compare algorithm performance.
Provide documentation with precise details regarding library algorithms.
This is a Python implementation of the APTED algorithm,which supersedes the RTED algorithm for computing the tree edit distance.
Baycomp is a library for Bayesian comparison of classifiers. Functions in the library compare two classifiers on one or on multiple data sets. They compute three probabilities: the probability that the first classifier has higher scores than the second, the probability that differences are within the region of practical equivalence (rope), or that the second classifier has higher scores.
This is a Python package to compute statistical test and add statistical annotations on an existing boxplots and barplots generated by seaborn.
Scikit-Optimize, or skopt, is a simple and efficient library to minimize (very) expensive and noisy black-box functions. It implements several methods for sequential model-based optimization. skopt aims to be accessible and easy to use in many contexts.
spin is a simple interface for common development tasks. It comes with a few common build commands out the box, but can easily be customized per project.
The impetus behind developing the tool was the mass migration of scientific Python libraries (SciPy, scikit-image, and NumPy, etc.) to Meson, after distutils was deprecated. When many of the build and installation commands changed, it made sense to abstract away the nuisance of having to re-learn them.
A Snakemake executor plugin for running SLURM jobs.
This package provides a Python package for calculating tissue-specificity metrics for gene expression.
PyMCubes is an implementation of the marching cubes algorithm to extract iso-surfaces from volumetric data. The volumetric data can be given as a three-dimensional NumPy array or as a Python function f(x, y, z).
This package provides miscellaneous tools for data analysis and scientific computing.
This package provides a stable interface for interactions between Snakemake and its storage plugins.
Scikit-survival is a Python module for survival analysis built on top of scikit-learn. It allows doing survival analysis while utilizing the power of scikit-learn, e.g., for pre-processing or doing cross-validation.