Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Scikit-Optimize, or skopt, is a simple and efficient library to minimize (very) expensive and noisy black-box functions. It implements several methods for sequential model-based optimization. skopt aims to be accessible and easy to use in many contexts.
Xarray (formerly xray) makes working with labelled multi-dimensional arrays simple, efficient, and fun!
Xarray introduces labels in the form of dimensions, coordinates and attributes on top of raw NumPy-like arrays, which allows for a more intuitive, more concise, and less error-prone developer experience. The package includes a large and growing library of domain-agnostic functions for advanced analytics and visualization with these data structures.
This package implements sparse arrays of arbitrary dimension on top of numpy and scipy.sparse. Sparse array is a matrix in which most of the elements are zero. python-sparse generalizes the scipy.sparse.coo_matrix and scipy.sparse.dok_matrix layouts, but extends beyond just rows and columns to an arbitrary number of dimensions. Additionally, this project maintains compatibility with the numpy.ndarray interface rather than the numpy.matrix interface used in scipy.sparse. These differences make this project useful in certain situations where scipy.sparse matrices are not well suited, but it should not be considered a full replacement. It lacks layouts that are not easily generalized like compressed sparse row/column(CSR/CSC) and depends on scipy.sparse for some computations.
pyvistaqt is a helper module for pyvista to enable you to plot using Qt by placing a vtk-widget into a background renderer. This can be quite useful when you desire to update your plot in real-time.
python-pandera provides a flexible and expressive API for performing data validation on dataframe-like objects to make data processing pipelines more readable and robust. Dataframes contain information that python-pandera explicitly validates at runtime. This is useful in production-critical data pipelines or reproducible research settings. With python-pandera, you can:
Define a schema once and use it to validate different dataframe types.
Check the types and properties of columns.
Perform more complex statistical validation like hypothesis testing.
Seamlessly integrate with existing data pipelines via function decorators.
Define dataframe models with the class-based API with pydantic-style syntax.
Synthesize data from schema objects for property-based testing.
Lazily validate dataframes so that all validation rules are executed.
Integrate with a rich ecosystem of tools like
python-pydantic,python-fastapiandpython-mypy.
Histoprint uses a mix of terminal color codes and Unicode trickery (i.e. combining characters) to plot overlaying histograms.
This package provides common functions and classes for Snakemake and its plugins.
This package provides functionality to make it easy to make scatter density maps, both for interactive and non-interactive use.
This package provides accelerated simulations and potentials of solids.
AlgoPy provides a functionality to differentiate functions implemented as computer programs by using Algorithmic Differentiation (AD) techniques in the forward and reverse mode.
The forward mode propagates univariate Taylor polynomials of arbitrary order. Hence it is also possible to use AlgoPy to evaluate higher-order derivative tensors. The reverse mode is also known as backpropagation and can be found in similar form in tools like PyTorch. Speciality of AlgoPy is the possibility to differentiate functions that contain matrix functions as +,-,*,/, dot, solve, qr, eigh, cholesky.
This package provides encoding and decoding routines that enable the serialization and deserialization of numerical and array data types provided by numpy using the highly efficient msgpack format. Serialization of Python's native complex data types is also supported.
DecayLanguage implements a language to describe and convert particle decays between digital representations, effectively making it possible to interoperate several fitting programs. Particular interest is given to programs dedicated to amplitude analyses.
Scikit-opt (or sko) is a Python module implementing swarm intelligence algorithms: genetic algorithm, particle swarm optimization, simulated annealing, ant colony algorithm, immune algorithm, artificial fish swarm algorithm.
This package provides utilities for exploratory analysis of large scale genetic variation data.
fgivenx is a Python package for plotting posteriors of functions. It is currently used in astronomy, but will be of use to any scientists performing Bayesian analyses which have predictive posteriors that are functions.
This package allows one to plot a predictive posterior of a function, dependent on sampled parameters. It assumes one has a Bayesian posterior Post(theta|D,M) described by a set of posterior samples theta_i~Post. If there is a function parameterised by theta y=f(x;theta), then this script will produce a contour plot of the conditional posterior P(y|x,D,M) in the (x,y) plane.
This package provides a Python interface to the QDLDL LDL factorization routine for quasi-definite linear system.
Scikit-survival is a Python module for survival analysis built on top of scikit-learn. It allows doing survival analysis while utilizing the power of scikit-learn, e.g., for pre-processing or doing cross-validation.
Baycomp is a library for Bayesian comparison of classifiers. Functions in the library compare two classifiers on one or on multiple data sets. They compute three probabilities: the probability that the first classifier has higher scores than the second, the probability that differences are within the region of practical equivalence (rope), or that the second classifier has higher scores.
This package implements many useful tools for projects involving fuzzy logic, also known as grey logic.
Pandas is a Python package providing fast, flexible, and expressive data structures designed to make working with structured (tabular, multidimensional, potentially heterogeneous) and time series data both easy and intuitive. It aims to be the fundamental high-level building block for doing practical, real world data analysis in Python.
Trimesh is a pure Python library for loading and using triangular meshes with an emphasis on watertight surfaces. The goal of the library is to provide a full featured and well tested Trimesh object which allows for easy manipulation and analysis, in the style of the Polygon object in the Shapely library.
iminuit is a Jupyter-friendly Python interface for the Minuit2 C++ library maintained by CERN's ROOT team.
Minuit was designed to optimize statistical cost functions, for maximum-likelihood and least-squares fits. It provides the best-fit parameters and error estimates from likelihood profile analysis.
Optionally, Iminuit supports SciPy minimizers as alternatives to Minuit's MIGRAD algorithm and Numba accelerated functions.
This is a rewrite of Dask DataFrame that includes query optimization and generally improved organization.
Dask.distributed is a lightweight library for distributed computing in Python. It extends both the concurrent.futures and dask APIs to moderate sized clusters.