Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
(N)compress provides the original compress and uncompress programs that used to be the de facto UNIX standard for compressing and uncompressing files. These programs implement a fast, simple Lempel-Ziv (LZW) file compression algorithm.
Plzip is a massively parallel (multi-threaded) lossless data compressor and decompressor that uses the lzip file format (.lz). Files produced by plzip are fully compatible with lzip and can be rescued with lziprecover. On multiprocessor machines, plzip can compress and decompress large files much faster than lzip, at the cost of a slightly reduced compression ratio (0.4% to 2%). The number of usable threads is limited by file size: on files of only a few MiB, plzip is no faster than lzip. Files that were compressed with regular lzip will also not be decompressed faster by plzip, unless the -b option was used: lzip usually produces single-member files which can't be decompressed in parallel.
Xarchiver is a front-end to various command line archiving tools. It uses GTK+ tool-kit and is designed to be desktop-environment independent. Supported formats are 7z, ARJ, bzip2, gzip, LHA, lzma, lzop, RAR, RPM, DEB, tar, and ZIP. It cannot perform functions for archives, whose archiver is not installed.
ZZipLib is a library based on zlib for accessing zip files.
The zchunk compressed file format allows splitting a file into independent chunks. This makes it possible to retrieve only changed chunks when downloading a new version of the file, and also makes zchunk files efficient over rsync. Along with the library, this package provides the following utilities:
unzckTo decompress a zchunk file.
zckTo compress a new zchunk file, or re-compress an existing one.
zck_delta_sizeTo calculate the difference between two zchunk files.
zck_gen_zdictTo create a dictionary for a zchunk file.
zck_read_headerTo read a zchunk header.
zckdlTo download a zchunk file.
This package provides a simple zip library based on miniz.
ISA-L is a collection of optimized low-level functions targeting storage applications. ISA-L includes:
Erasure codes: fast block Reed-Solomon type erasure codes for any encode/decode matrix;
CRC: fast implementations of cyclic redundancy check. Six different polynomials supported: iscsi32, ieee32, t10dif, ecma64, iso64, jones64;
Raid: calculate and operate on XOR and P+Q parity found in common RAID implementations;
Compression: fast deflate-compatible data compression;
De-compression: fast inflate-compatible data compression;
igzip: command line application like gzip, accelerated with ISA-L.
Lhasa is a replacement for the Unix LHa tool, for decompressing .lzh (LHA / LHarc) and .lzs (LArc) archives. The backend for the tool is a library, so that it can be reused for other purposes. Lhasa aims to be compatible with as many types of .lzh/lzs archives as possible. It also aims to generate the same output as the (non-free) Unix lha tool, so that it will act as a free drop-in replacement.
This package provides a script to unpack self-extracting archives generated by makeself or mojo without running the possibly untrusted extraction shell script.
Zopfli Compression Algorithm is a compression library programmed in C to perform very good, but slow, deflate or zlib compression. ZopfliCompress supports the deflate, gzip and zlib output formats. This library can only compress, not decompress; existing zlib or deflate libraries can decompress the data.
Snappy is a compression/decompression library. It does not aim for maximum compression, or compatibility with any other compression library; instead, it aims for very high speeds and reasonable compression. For instance, compared to the fastest mode of zlib, Snappy is an order of magnitude faster for most inputs, but the resulting compressed files are anywhere from 20% to 100% bigger.
UnZip is an extraction utility for archives compressed in .zip format, also called "zipfiles".
UnZip lists, tests, or extracts files from a .zip archive. The default behaviour (with no options) is to extract into the current directory, and subdirectories below it, all files from the specified zipfile. UnZip recreates the stored directory structure by default.
7-zip is a command-line file compressor that supports a number of archive formats and features self-extracting archives.
UCL implements a number of compression algorithms that achieve an excellent compression ratio while allowing fast decompression. Decompression requires no additional memory.
Compared to LZO, the UCL algorithms achieve a better compression ratio but decompression is a little bit slower.
XZ Utils is free general-purpose data compression software with high compression ratio. XZ Utils were written for POSIX-like systems, but also work on some not-so-POSIX systems. XZ Utils are the successor to LZMA Utils.
The core of the XZ Utils compression code is based on LZMA SDK, but it has been modified quite a lot to be suitable for XZ Utils. The primary compression algorithm is currently LZMA2, which is used inside the .xz container format. With typical files, XZ Utils create 30 % smaller output than gzip and 15 % smaller output than bzip2.
Archive huge numbers of files, or split massive tar archives into smaller chunks.
LZO is a data compression library which is suitable for data de-/compression in real-time. This means it favours speed over compression ratio.
LZO is written in ANSI C. Both the source code and the compressed data format are designed to be portable across platforms.
GNU sharutils is a package for creating and manipulating shell archives that can be readily emailed. A shell archive is a file that can be processed by a Bourne-type shell to unpack the original collection of files. This package is mostly for compatibility and historical interest.
This package provides the reference implementation of Brotli, a generic-purpose lossless compression algorithm that compresses data using a combination of a modern variant of the LZ77 algorithm, Huffman coding and 2nd order context modeling, with a compression ratio comparable to the best currently available general-purpose compression methods. It is similar in speed with deflate but offers more dense compression.
The specification of the Brotli Compressed Data Format is defined in RFC 7932.
LZFSE is a Lempel-Ziv style data compression algorithm using Finite State Entropy coding. It targets similar compression rates at higher compression and decompression speed compared to Deflate using Zlib.
Miniz is a lossless data compression library that implements the zlib (RFC 1950) and Deflate (RFC 1951) compressed data format specification standards. It supports the most commonly used functions exported by the zlib library.
Zstandard (zstd) is a lossless compression algorithm that combines very fast operation with a compression ratio comparable to that of zlib. In most scenarios, both compression and decompression can be performed in ‘real time’. The compressor can be configured to provide the most suitable trade-off between compression ratio and speed, without affecting decompression speed.
This package provides a parallel implementation of gzip that exploits multiple processors and multiple cores when compressing data.
minizip-ng is a zip manipulation library written in C, forked from the zip manipulation library found in the zlib distribution.