Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Books are "Linear Models with R" published 1st Ed. August 2004, 2nd Ed. July 2014, 3rd Ed. February 2025 by CRC press, ISBN 9781439887332, and "Extending the Linear Model with R" published by CRC press in 1st Ed. December 2005 and 2nd Ed. March 2016, ISBN 9781584884248 and "Practical Regression and ANOVA in R" contributed documentation on CRAN (now very dated).
This package provides a guarded resampling workflow for training and evaluating machine-learning models. When the guarded resampling path is used, preprocessing and model fitting are re-estimated within each resampling split to reduce leakage risk. Supports multiple resampling schemes, integrates with established engines in the tidymodels ecosystem, and aims to improve evaluation reliability by coordinating preprocessing, fitting, and evaluation within supported workflows. Offers a lightweight AutoML-style workflow by automating model training, resampling, and tuning across multiple algorithms, while keeping evaluation design explicit and user-controlled.
Pairwise Hamming distances are computed between the rows of a binary (0/1) matrix using highly optimized C code. The input is an integer matrix where each row represents a binary feature vector and returns a symmetric integer matrix of pairwise distances. Internally, rows are bit-packed into 64-bit words for fast XOR-based comparisons, with hardware-accelerated popcount operations to count differences. OpenMP parallelization ensures efficient performance for large matrices.
Generate privacy-preserving synthetic datasets that mirror structure, types, factor levels, and missingness; export bundles for LLM workflows (data plus JSON schema and guidance); and build fake data directly from SQL database tables without reading real rows. Methods are related to approaches in Nowok, Raab and Dibben (2016) <doi:10.32614/RJ-2016-019> and the foundation-model overview by Bommasani et al. (2021) <doi:10.48550/arXiv.2108.07258>.
Create fake datasets that can be used for prototyping and teaching. This package provides a set of functions to generate fake data for a variety of data types, such as dates, addresses, and names. It can be used for prototyping (notably in shiny') or as a tool to teach data manipulation and data visualization.
Functional clustering aims to group curves exhibiting similar temporal behaviour and to obtain representative curves summarising the typical dynamics within each cluster. A key challenge in this setting is class imbalance, where some clusters contain substantially more curves than others, which can adversely affect clustering performance. While class imbalance has been extensively studied in supervised classification, it has received comparatively little attention in unsupervised clustering. This package implements functional iterative hierarchical clustering ('funIHC'), an adaptation of the iterative hierarchical clustering method originally developed for multivariate data, to the functional data setting. For further details, please see Higgins and Carey (2024) <doi:10.1007/s11634-024-00611-8>.
Reads and writes ARFF files. ARFF (Attribute-Relation File Format) files are like CSV files, with a little bit of added meta information in a header and standardized NA values. They are quite often used for machine learning data sets and were introduced for the WEKA machine learning Java toolbox. See <https://waikato.github.io/weka-wiki/formats_and_processing/arff_stable/> for further info on ARFF and for <http://www.cs.waikato.ac.nz/ml/weka/> for more info on WEKA'. farff gets rid of the Java dependency that RWeka enforces, and it is at least a faster reader (for bigger files). It uses readr as parser back-end for the data section of the ARFF file. Consistency with RWeka is tested on Github and Travis CI with hundreds of ARFF files from OpenML'.
Turn numeric,data.frame,matrix into fraction form.
This package provides functions for creating flashcard decks of terms and definitions. This package creates HTML slides using revealjs that can be viewed in the RStudio viewer or a web browser. Users can create flashcards from either existing built-in decks or create their own from CSV files or vectors of function names.
This package provides analytics directly from R'. It requires: FormShare App': <https://github.com/qlands/FormShare >= 2.22.0> . Analytics plugin: <https://github.com/qlands/formshare_analytics_plugin> . Remote SQL plugin: <https://github.com/qlands/formshare_sql_plugin> .
Computes six functional diversity indices. These are namely, Functional Divergence (FDiv), Function Evenness (FEve), Functional Richness (FRic), Functional Richness intersections (FRic_intersect), Functional Dispersion (FDis), and Rao's entropy (Q) (reviewed in Villéger et al. 2008 <doi:10.1890/07-1206.1>). Provides efficient, modular, and parallel functions to compute functional diversity indices (preprint: <doi:10.32942/osf.io/dg7hw>).
This is a method for Allele-specific DNA Copy Number Profiling using Next-Generation Sequencing. Given the allele-specific coverage at the variant loci, this program segments the genome into regions of homogeneous allele-specific copy number. It requires, as input, the read counts for each variant allele in a pair of case and control samples. For detection of somatic mutations, the case and control samples can be the tumor and normal sample from the same individual.
Helpers for parsing out the R functions and packages used in R scripts and notebooks.
This package provides a computationally efficient and statistically rigorous fast Kernel Machine method for multi-kernel analysis. The approach is based on a low-rank approximation to the nuisance effect kernel matrices. The algorithm is applicable to continuous, binary, and survival traits and is implemented using the existing single-kernel analysis software SKAT and coxKM'. coxKM can be obtained from <https://github.com/lin-lab/coxKM>.
Supports fMRI (functional magnetic resonance imaging) analysis tasks including reading in CIFTI', GIFTI and NIFTI data, temporal filtering, nuisance regression, and aCompCor (anatomical Components Correction) (Muschelli et al. (2014) <doi:10.1016/j.neuroimage.2014.03.028>).
The Futureverse is a set of packages for parallel and distributed processing with the future package at its core, cf. Bengtsson (2021) <doi:10.32614/RJ-2021-048>. This package is designed to make it easy to install common Futureverse packages in a single step. This package is intended for end-users, interactive use, and R scripts. Packages must not list it as a dependency - instead, explicitly declare each Futureverse package as a dependency as needed.
Parse and create Darwin Core (<http://rs.tdwg.org/dwc/>) Simple and Archives. Functionality includes reading and parsing all the files in a Darwin Core Archive, including the datasets and metadata; read and parse simple Darwin Core files; and validation of Darwin Core Archives.
The goal of this package is to provide an improved version of WA-PLS (Weighted Averaging Partial Least Squares) by including the tolerances of taxa and the frequency of the sampled climate variable. This package also provides a way of leave-out cross-validation that removes both the test site and sites that are both geographically close and climatically close for each cycle, to avoid the risk of pseudo-replication.
TrainFastImputation() uses training data to describe a multivariate normal distribution that the data approximates or can be transformed into approximating and stores this information as an object of class FastImputationPatterns'. FastImputation() function uses this FastImputationPatterns object to impute (make a good guess at) missing data in a single line or a whole data frame of data. This approximates the process used by Amelia <https://gking.harvard.edu/amelia> but is much faster when filling in values for a single line of data.
Plotting flood quantiles and their corresponding probabilities (return periods) on the probability papers. The details of relevant methods are available in Chow et al (1988, ISBN: 007070242X, 9780070702424), and Bobee and Ashkar (1991, ISBN: 0918334683, 9780918334688).
The futurize() function transpiles calls to sequential map-reduce functions such as base::lapply(), purrr::map(), foreach::foreach() %do% ... into concurrent alternatives, providing you with a simple, straightforward path to scalable parallel computing via the future ecosystem <doi:10.32614/RJ-2021-048>. By combining this function with R's native pipe operator, you have an convenient way for speeding up iterative computations with minimal refactoring, e.g. lapply(xs, fcn) |> futurize()', purrr::map(xs, fcn) |> futurize()', and foreach::foreach(x = xs) %do% fcn(x) |> futurize()'. Other map-reduce packages that be "futurized" are BiocParallel', plyr', crossmap packages. There is also support for growing set of domain-specific packages, including boot', glmnet', mgcv', lme4', and tm'.
This package provides methods for matrix factorization based on Wang and Stephens (2021) <https://jmlr.org/papers/v22/20-589.html>.
Data sets and utilities to accompany the second edition of "Foundations and Applications of Statistics: an Introduction using R" (R Pruim, published by AMS, 2017), a text covering topics from probability and mathematical statistics at an advanced undergraduate level. R is integrated throughout, and access to all the R code in the book is provided via the snippet() function.
This package provides a joint model for large-scale, competing risks time-to-event data with singular or multiple longitudinal biomarkers, implemented with the efficient algorithms developed by Li and colleagues (2022) <doi:10.1155/2022/1362913> and <doi:10.48550/arXiv.2506.12741>. The time-to-event data is modelled using a (cause-specific) Cox proportional hazards regression model with time-fixed covariates. The longitudinal biomarkers are modelled using a linear mixed effects model. The association between the longitudinal submodel and the survival submodel is captured through shared random effects. It allows researchers to analyze large-scale data to model biomarker trajectories, estimate their effects on event outcomes, and dynamically predict future events from patientsâ past histories. A function for simulating survival and longitudinal data for multiple biomarkers is also included alongside built-in datasets.