Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Urllib3 supports features left out of urllib and urllib2 libraries. It can reuse the same socket connection for multiple requests, it can POST files, supports url redirection and retries, and also gzip and deflate decoding.
CacheControl is a port of the caching algorithms in httplib2 for use with requests session objects.
Gunicorn ‘Green Unicorn’ is a Python WSGI HTTP Server for UNIX. It’s a pre-fork worker model ported from Ruby’s Unicorn project. The Gunicorn server is broadly compatible with various web frameworks, simply implemented, light on server resources, and fairly speedy.
S3transfer is a Python library for managing Amazon S3 transfers.
This package is a fork of the pickle module (and the supporting C extension) from both Python 3.2 and Python 3.3. The fork adds support for the noload operations used by zodb.
PRAWcore is a low-level communication layer used by PRAW.
python-flask-oidc provides an OpenID Connect extension for Flask.
This package provides an implementation of itertools, builtins, and more for AsyncIO and mixed-type iterables.
Scrapy is a fast high-level web crawling and web scraping framework, used to crawl websites and extract structured data from their pages. It can be used for a wide range of purposes, from data mining to monitoring and automated testing.
aioquic is a library for the QUIC network protocol in Python. It features a minimal TLS 1.3 implementation, a QUIC stack and an HTTP/3 stack.
Pycurl is a lightweight wrapper around libcurl. It provides high-speed transfers via libcurl and frequently outperforms alternatives.
robot_detection is a python module to detect if a given HTTP User Agent is a web crawler. It uses the list of registered robots from http://www.robotstxt.org.
This package provides a script for running Python tests.
This package provides utilities for Google Media Downloads and Resumable Uploads.
Python client library for Google's discovery based APIs
This package integrates webassets with Flask, adding support for merging, minifying and compiling CSS and Javascript files.
This package provides a Python library with classes that mimic pathlib.Path's interface for URIs from different cloud storage services.
Selenium enables web browser automation. It specifically provides infrastructure for the W3C WebDriver specification — a platform and language-neutral coding interface compatible with all major web browsers.
This package implements RetryClient for aiohttp to retry connection to provided endpoint with timeouts logic or use:
ExponentialRetrywith exponential backoffRandomRetryfor random backoffListRetrywith backoff you predefine by listFibonacciRetrywith backoff that looks like fibonacci sequenceJitterRetryexponential retry with a bit of randomness
The JavaScript Object Signing and Encryption (JOSE) technologies - JSON Web Signature (JWS), JSON Web Encryption (JWE), JSON Web Key (JWK), and JSON Web Algorithms (JWA) - collectively can be used to encrypt and/or sign content using a variety of algorithms.
Hupper is an integrated process monitor that will track changes to any imported Python files in sys.modules as well as custom paths. When files are changed the process is restarted.
FastAPI provides a web API framework based on pydantic and starlette.
Socks is a library providing core proxy (SOCKS4, SOCKS5, HTTP tunneling) functionality.
Bind to an ephemeral port, force it into the TIME_WAIT state, and unbind it.