Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel search send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Client for prometheus
A list of color names and its values
Color conversion and manipulation with CSS string support
Safer Node.js Buffer API
Streams3, a user-land copy of the stream library from Node.js
Plain color conversion functions
the complete solution for node.js command-line programs
Base stream implementations for winston@3 and up.
Determines if an object can be used as an array
An extensible environment for interactive and reproducible computing, based on the Jupyter Notebook and Architecture.
LDAP Authenticator for JupyterHub
This package provides a spawner for Jupyterhub to spawn notebooks using batch resource managers.
The systemdspawner enables JupyterHub to spawn single-user notebook servers using systemd.
This package includes WrapSpawner and ProfilesSpawner, which provide mechanisms for runtime configuration of spawners. The inspiration for their development was to allow users to select from a range of pre-defined batch job profiles, but their operation is completely generic.
Jupyter telemetry library
JupyterHub: A multi-user server for Jupyter notebooks
The SudoSpawner enables JupyterHub to spawn single-user servers without being root, by spawning an intermediate process via sudo, which takes actions on behalf of the user.
Datasets is a lightweight library providing access to major public datasets (image, audio, text, etc.), as well as enabling efficient data preparation for inspection and ML model evaluation and training.
Accelerate was created for PyTorch users who like to write the training loop of PyTorch models but are reluctant to write and maintain the boilerplate code needed to use multi-GPUs/TPU/fp16. It abstracts exactly and only the boilerplate code related to multi-GPUs/TPU/fp16 and leaves the rest of your code unchanged.
sklearn-compat is a small Python package that help developer writing scikit-learn compatible estimators to support multiple scikit-learn versions.
Ray is a unified framework for scaling AI and Python applications. Ray consists of a core distributed runtime and a set of AI libraries for simplifying ML compute. These are the provided Ray AI libraries:
Data: Scalable datasets for ML;
Train: Distributed training;
Tune: Scalable hyperparameter tuning;
RLlib: Scalable reinforcement learning;
Serve: Scalable and programmable serving.
Ray is a unified framework for scaling AI and Python applications. Ray consists of a core distributed runtime and a set of AI libraries for simplifying ML compute. These are the provided Ray AI libraries:
Data: Scalable datasets for ML;
Train: Distributed training;
Tune: Scalable hyperparameter tuning;
RLlib: Scalable reinforcement learning;
Serve: Scalable and programmable serving.
PyDMD is a Python package designed for Dynamic Mode Decomposition (DMD), a data-driven method used for analyzing and extracting spatiotemporal coherent structures from time-varying datasets. It provides a comprehensive and user-friendly interface for performing DMD analysis, making it a valuable tool for researchers, engineers, and data scientists working in various fields.
Flax is a neural network library for JAX that is designed for flexibility.