Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
This package provides a Qt graphical interface to archiving programs like tar and zip.
Libqtxdg implements the freedesktop.org xdg specifications in Qt.
ScreenGrab is a program for fast creating screenshots, and easily publishing them on internet image hosting services.
lxqt-notificationd is LXQt's implementation of a daemon according to the Desktop Notifications Specification.
LXQt is a lightweight Qt desktop environment.
lxqt-session provides the standard session manager for the LXQt desktop environment.
lxqt-qtplugin is providing a library libqtlxqt to integrate Qt with LXQt.
This package comprises a number of graphic files and themes for LXQt.
GNU M4 is an implementation of the M4 macro language, which features some extensions over other implementations, some of which are required by GNU Autoconf. It is used as a macro processor, which means it processes text, expanding macros as it encounters them. It also has some built-in functions, for example to run shell commands or to do arithmetic.
This package provides easy download of thousands of pretrained models to perform tasks on different modalities such as text, vision, and audio.
These models can be applied on:
Text, for tasks like text classification, information extraction, question answering, summarization, translation, and text generation, in over 100 languages.
Images, for tasks like image classification, object detection, and segmentation.
Audio, for tasks like speech recognition and audio classification.
Transformer models can also perform tasks on several modalities combined, such as table question answering, optical character recognition, information extraction from scanned documents, video classification, and visual question answering.
This package provides APIs to quickly download and use those pretrained models on a given text, fine-tune them on your own datasets and then share them with the community. At the same time, each Python module defining an architecture is fully standalone and can be modified to enable quick research experiments.
Transformers is backed by the three most popular deep learning libraries — Jax, PyTorch and TensorFlow — with a seamless integration between them.
SpaCy is a library for advanced Natural Language Processing in Python and Cython. It comes with pretrained pipelines and currently supports tokenization and training for 70+ languages. It features state-of-the-art speed and neural network models for tagging, parsing, named entity recognition, text classification and more, multi-task learning with pretrained transformers like BERT, as well as a production-ready training system and easy model packaging, deployment and workflow management.
This package is a high-performance inference of OpenAI's Whisper automatic speech recognition (ASR) model, implemented in plain C/C++ without dependencies, with
AVX intrinsics support for x86 architectures
VSX intrinsics support for POWER architectures
Mixed F16 / F32 precision
4-bit and 5-bit integer quantization support
Zero memory allocations at runtime
Support for CPU-only inference
Efficient GPU support for NVIDIA
OpenVINO Support
C-style API
Apache TVM is a compiler stack for deep learning systems. It is designed to close the gap between the productivity-focused deep learning frameworks, and the performance- and efficiency-focused hardware backends. TVM works with deep learning frameworks to provide end to end compilation to different backends
The General Hidden Markov Model library (GHMM) is a C library with additional Python bindings implementing a wide range of types of Hidden Markov Models (HMM) and algorithms: discrete, continuous emissions, basic training, HMM clustering, HMM mixtures.
PyG is a library built upon PyTorch to easily write and train Graph Neural Networks for a wide range of applications related to structured data.
This package provides logging utilities for the SpaCy natural language processing framework.
fastText is a library for efficient learning of word representations and sentence classification.
fastText is a library for efficient learning of word representations and sentence classification.
DeepXDE is a library for scientific machine learning and physics-informed learning. It includes implementations for the PINN (physics-informed neural networks), DeepONet (deep operator network) and MFNN (multifidelity neural network) algorithms.
This package provides a port to Facebook's LLaMA collection of foundation language models. It requires models parameters to be downloaded independently to be able to run a LLaMA model.
ML Collections is a library of Python collections designed for Machine Learning usecases.
This package enables you to deserialize Lua torch-serialized objects from Python.
CTranslate2 is a C++ and Python library for efficient inference with Transformer models.
The project implements a custom runtime that applies many performance optimization techniques such as weights quantization, layers fusion, batch reordering, etc., to accelerate and reduce the memory usage of Transformer models on CPU and GPU.
This package provides Autograd-compatible approximations to the gamma family of functions.