Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Blosc is a high performance compressor optimized for binary data. It has been designed to transmit data to the processor cache faster than the traditional, non-compressed, direct memory fetch approach via a memcpy() system call. Blosc is meant not only to reduce the size of large datasets on-disk or in-memory, but also to accelerate memory-bound computations.
Blosc is a high performance compressor optimized for binary data (i.e. floating point numbers, integers and booleans, although it can handle string data too). It has been designed to transmit data to the processor cache faster than the traditional, non-compressed, direct memory fetch approach via a memcpy() system call. Blosc main goal is not just to reduce the size of large datasets on-disk or in-memory, but also to accelerate memory-bound computations.
C-Blosc2 is the new major version of C-Blosc, and is backward compatible with both the C-Blosc1 API and its in-memory format. However, the reverse thing is generally not true for the format; buffers generated with C-Blosc2 are not format-compatible with C-Blosc1 (i.e. forward compatibility is not supported).
CBOR is a data format whose design goals include the possibility of extremely small code size, fairly small message size, and extensibility without the need for version negotiation. These design goals make it different from earlier binary serializations such as ASN.1 and MessagePack.
Lziprecover is a data recovery tool and decompressor for files in the lzip compressed data format (.lz). It can test the integrity of lzip files, extract data from damaged ones, and repair most files with small errors (up to one single-byte error per member) entirely.
Lziprecover is not a replacement for regular backups, but a last line of defence when even the backups are corrupt. It can recover files by merging the good parts of two or more damaged copies, such as can be easily produced by running ddrescue on a failing device.
This package also includes unzcrash, a tool to test the robustness of decompressors when faced with corrupted input.
gzstream is a small library for providing zlib functionality in a C++ iostream.
Libdeflate is a library for fast, whole-buffer DEFLATE-based compression and decompression. The supported formats are:
DEFLATE (raw)
zlib (a.k.a. DEFLATE with a zlib wrapper)
gzip (a.k.a. DEFLATE with a gzip wrapper)
A data compression/decompression library for embedded/real-time systems.
Among its features are:
Low memory usage (as low as 50 bytes.) It is useful for some cases with less than 50 bytes, and useful for many general cases with less than 300 bytes.
Incremental, bounded CPU use. It can be used to chew on input data in arbitrarily tiny bites. This is a useful property in hard real-time environments.
Can use either static or dynamic memory allocation.
Zip is a compression and file packaging/archive utility. Zip is useful for packaging a set of files for distribution, for archiving files, and for saving disk space by temporarily compressing unused files or directories. Zip puts one or more compressed files into a single ZIP archive, along with information about the files (name, path, date, time of last modification, protection, and check information to verify file integrity). An entire directory structure can be packed into a ZIP archive with a single command.
Zip has one compression method (deflation) and can also store files without compression. Zip automatically chooses the better of the two for each file. Compression ratios of 2:1 to 3:1 are common for text files.
This package provides a simple zip library based on miniz.
Lunzip is a decompressor for files in the lzip compression format (.lz), written as a single small C tool with no dependencies. This makes it well-suited to embedded and other systems without a C++ compiler, or for use in applications such as software installers that need only to decompress files, not compress them. Lunzip is intended to be fully compatible with the regular lzip package.
Lzip is a lossless data compressor with a user interface similar to the one of gzip or bzip2. Lzip decompresses almost as fast as gzip and compresses more than bzip2, which makes it well-suited for software distribution and data archiving. Lzip is a clean implementation of the LZMA algorithm.
ZZipLib is a library based on zlib for accessing zip files.
Plzip is a massively parallel (multi-threaded) lossless data compressor and decompressor that uses the lzip file format (.lz). Files produced by plzip are fully compatible with lzip and can be rescued with lziprecover. On multiprocessor machines, plzip can compress and decompress large files much faster than lzip, at the cost of a slightly reduced compression ratio (0.4% to 2%). The number of usable threads is limited by file size: on files of only a few MiB, plzip is no faster than lzip. Files that were compressed with regular lzip will also not be decompressed faster by plzip, unless the -b option was used: lzip usually produces single-member files which can't be decompressed in parallel.
Miniz is a lossless data compression library that implements the zlib (RFC 1950) and Deflate (RFC 1951) compressed data format specification standards. It supports the most commonly used functions exported by the zlib library.
This library allows reading and writing gzip-compressed JSON catalog files, which can be used to store GPG, PKCS-7 and SHA-256 checksums for each file.
This package provides the reference implementation of Brotli, a generic-purpose lossless compression algorithm that compresses data using a combination of a modern variant of the LZ77 algorithm, Huffman coding and 2nd order context modeling, with a compression ratio comparable to the best currently available general-purpose compression methods. It is similar in speed with deflate but offers more dense compression.
The specification of the Brotli Compressed Data Format is defined in RFC 7932.
Miniz is a lossless data compression library that implements the zlib (RFC 1950) and Deflate (RFC 1951) compressed data format specification standards. It supports the most commonly used functions exported by the zlib library.
Xarchiver is a front-end to various command line archiving tools. It uses GTK+ tool-kit and is designed to be desktop-environment independent. Supported formats are 7z, ARJ, bzip2, gzip, LHA, lzma, lzop, RAR, RPM, DEB, tar, and ZIP. It cannot perform functions for archives, whose archiver is not installed.
Zstandard (zstd) is a lossless compression algorithm that combines very fast operation with a compression ratio comparable to that of zlib. In most scenarios, both compression and decompression can be performed in ‘real time’. The compressor can be configured to provide the most suitable trade-off between compression ratio and speed, without affecting decompression speed.
7-zip is a command-line file compressor that supports a number of archive formats and features self-extracting archives.
xdelta encodes only the differences between two binary files using the VCDIFF algorithm and patch file format described in RFC 3284. It can also be used to apply such patches. xdelta is similar to diff and patch, but is not limited to plain text and does not generate human-readable output.
The existing XZ Utils provide great compression in the .xz file format, but they produce just one big block of compressed data. Pixz instead produces a collection of smaller blocks which makes random access to the original data possible and can compress in parallel. This is especially useful for large tarballs.
The main command is aunpack which extracts files from an archive. The other commands provided are apack (to create archives), als (to list files in archives), and acat (to extract files to standard out). As atool invokes external programs to handle the archives, not all commands may be supported for a certain type of archives.
The zchunk compressed file format allows splitting a file into independent chunks. This makes it possible to retrieve only changed chunks when downloading a new version of the file, and also makes zchunk files efficient over rsync. Along with the library, this package provides the following utilities:
unzckTo decompress a zchunk file.
zckTo compress a new zchunk file, or re-compress an existing one.
zck_delta_sizeTo calculate the difference between two zchunk files.
zck_gen_zdictTo create a dictionary for a zchunk file.
zck_read_headerTo read a zchunk header.
zckdlTo download a zchunk file.