Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
This package provides a CLI tool and Python utility functions for manipulating SQLite databases. It's main features are:
Pipe JSON (or CSV or TSV) directly into a new SQLite database file, automatically creating a table with the appropriate schema.
Run in-memory SQL queries, including joins, directly against data in CSV, TSV or JSON files and view the results.
Configure SQLite full-text search against your database tables and run search queries against them, ordered by relevance.
Run transformations against your tables to make schema changes that SQLite ALTER TABLE does not directly support, such as changing the type of a column.
Extract columns into separate tables to better normalize your existing data.
WiredTiger is an extensible platform for data management. It supports row-oriented storage (where all columns of a row are stored together), column-oriented storage (where columns are stored in groups, allowing for more efficient access and storage of column subsets) and log-structured merge trees (LSM), for sustained throughput under random insert workloads.
This package adds the support of pgvector vectors to Tortoise-ORM as a new type of fields, it helps to filter/order by cosine similarity distances for scementic search using embeddings.
This module creates a temporary SQLite database, deploys a DBIC schema, and then connects to it. This lets you easily test DBIC schema. Since you have a fresh database for every test, you don't have to worry about cleaning up after your tests, ordering of tests affecting failure, etc.
This package is an implementation of psycopg2 using CFFI.
dogpile.cache is a caching API which provides a generic interface to caching backends of any variety, and additionally provides API hooks which integrate these cache backends with the locking mechanism of dogpile.
Yoyo is a database schema migration tool. Migrations are written as SQL files or Python scripts that define a list of migration steps.
Pebble is a LevelDB/RocksDB inspired key-value store focused on performance and internal usage by CockroachDB. Pebble inherits the RocksDB file formats and a few extensions such as range deletion tombstones, table-level bloom filters, and updates to the MANIFEST format. This package provides command line interface (CLI).
The Lightning Memory-Mapped Database (LMDB) is a high-performance transactional database. Unlike more complex relational databases, LMDB handles only key-value pairs (stored as arbitrary byte arrays) and relies on the underlying operating system for caching and locking, keeping the code small and simple. The use of ‘zero-copy’ memory-mapped files combines the persistence of classic disk-based databases with high read performance that scales linearly over multiple cores. The size of each database is limited only by the size of the virtual address space — not physical RAM.
MyCLI is a command line interface for MySQL, MariaDB, and Percona with auto-completion and syntax highlighting.
Peewee is a simple and small ORM (object-relation mapping) tool. Peewee handles converting between pythonic values and those used by databases, so you can use Python types in your code without having to worry. It has built-in support for sqlite, mysql and postgresql. If you already have a database, you can autogenerate peewee models using pwiz, a model generator.
This module tries to split any SQL code, even including non-standard extensions, into the atomic statements it is composed of.
Pebble is a LevelDB/RocksDB inspired key-value store focused on performance and internal usage by CockroachDB. Pebble inherits the RocksDB file formats and a few extensions such as range deletion tombstones, table-level bloom filters, and updates to the MANIFEST format.
python-pyodbc provides a Python DB-API driver for ODBC.
This package provides an Emacs major mode rec-mode for working with GNU Recutils text-based, human-editable databases. It supports editing, navigation, and querying of recutils database files including field and record folding.
UnQLite is an in-process software library which implements a self-contained, serverless, zero-configuration, transactional NoSQL database engine. UnQLite is a document store database similar to Redis, CouchDB, etc., as well as a standard key/value store similar to BerkeleyDB, LevelDB, etc.
This package provides mock helpers for SQLAlchemy that makes it easy to mock an SQLAlchemy session while preserving the ability to do asserts.
Normally Normally SQLAlchemy's expressions cannot be easily compared as comparison on binary expression produces yet another binary expression, but this library provides functions to facilitate such comparisons.
python-sql is a library to write SQL queries, that transforms idiomatic python function calls to well-formed SQL queries.
This package works in conjunction with InflateColumn::DateTime to automatically set update and create date and time based fields in a table.
This module is nearly identical to SQL::Abstract 1.81, and exists to preserve the ability of users to opt into the new way of doing things in later versions according to their own schedules.
It is an abstract SQL generation module based on the concepts used by DBIx::Abstract, with several important differences, especially when it comes to WHERE clauses. These concepts were modified to make the SQL easier to generate from Perl data structures.
The underlying idea is for this module to do what you mean, based on the data structures you provide it. You shouldn't have to modify your code every time your data changes, as this module figures it out.
fastparquet is a Python implementation of the Parquet file format. fastparquet is used implicitly by dask, pandas and intake-parquet. It supports the following compression algorithms:
Gzip
Snappy
Brotli
LZ4
Zstd
LZO (optionally)
Automatically set and update fields with values calculated at runtime. Ipdate or create actions will set the specified columns to the value returned by the callback you specified as a method name or code reference.
Because the many-to-many relationships are not real relationships, they can not be introspected with DBIx::Class. Many-to-many relationships are actually just a collection of convenience methods installed to bridge two relationships. This DBIx::Class component can be used to store all relevant information about these non-relationships so they can later be introspected and examined.
ORC is a self-describing type-aware columnar file format designed for Hadoop workloads. It is optimized for large streaming reads, but with integrated support for finding required rows quickly.