Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Allows you to conduct robust correlations on your non-normal data set. The robust correlations included in the package are median-absolute-deviation and median-based correlations. Li, J.C.H. (2022) <doi:10.5964/meth.8467>.
Uses the CMS application programming interface <https://dnav.cms.gov/api/healthdata> to provide users databases containing yearly Medicare reimbursement rates in the United States. Data can be acquired for the entire United States or only for specific localities. Currently, support is only provided for the Medicare Physician Fee Schedule, but support will be expanded for other CMS databases in future versions.
This package implements bound constrained optimal sample size allocation (BCOSSA) framework described in Bulus & Dong (2021) <doi:10.1080/00220973.2019.1636197> for power analysis of multilevel regression discontinuity designs (MRDDs) and multilevel randomized trials (MRTs) with continuous outcomes. Minimum detectable effect size (MDES) and power computations for MRDDs allow polynomial functional form specification for the score variable (with or without interaction with the treatment indicator). See Bulus (2021) <doi:10.1080/19345747.2021.1947425>.
Mainly used to build tables that are commonly presented for bio-medical/health research, such as basic characteristic tables or descriptive statistics.
This package implements the Centroid Decision Forest (CDF) as a single user-facing function CDF(). The method selects discriminative features via a multi-class class separability score (CSS), splits by nearest class centroid, and aggregates tree votes to produce predictions and class probabilities. Returns CSS-based feature importance as well. Amjad Ali, Saeed Aldahmani, Zardad Khan (2025) <doi:10.48550/arXiv.2503.19306>.
Utilize the shiny interface to generate Goodness of Fit (GOF) plots and tables for Non-Linear Mixed Effects (NLME / NONMEM) pharmacometric models. From the interface, users can customize model diagnostics and generate the underlying R code to reproduce the diagnostic plots and tables outside of the shiny session. Model diagnostics can be included in a rmarkdown document and rendered to desired output format.
Implementation of the CNAIM standard in R. Contains a series of algorithms which determine the probability of failure, consequences of failure and monetary risk associated with electricity distribution companies assets such as transformers and cables. Results are visualized in an easy-to-understand risk matrix.
Perform variable selection for Cox regression model with interval-censored data. Can deal with both low-dimensional and high-dimensional data. Case-cohort design can be incorporated. Two sets of covariates scenario can also be considered. The references are listed in the URL below.
The caroline R library contains dozens of functions useful for: database migration (dbWriteTable2), database style joins & aggregation (nerge, groupBy, & bestBy), data structure conversion (nv, tab2df), legend table making (sstable & leghead), automatic legend positioning for scatter and box plots (), plot annotation (labsegs & mvlabs), data visualization (pies, sparge, confound.grid & raPlot), character string manipulation (m & pad), file I/O (write.delim), batch scripting, data exploration, and more. The package's greatest contributions lie in the database style merge, aggregation and interface functions as well as in it's extensive use and propagation of row, column and vector names in most functions.
In statistical modeling, multiple models need to be compared based on certain criteria. The method described here uses eight metrics from AllMetrics package. â input_dfâ is the data frame (at least two columns for comparison) containing metrics values in different rows of a column (which denotes a particular modelâ s performance). First five metrics are expected to be minimum and last three metrics are expected to be maximum for a model to be considered good. Firstly, every metric value (among first five) is searched in every columns and minimum values are denoted as â MINâ and other values are denoted as â NAâ . Secondly, every metric (among last three) is searched in every columns and maximum values are denoted as â MAXâ and other values are denoted as â NAâ . â output_dfâ contains the similar number of rows (which is 8) and columns (which is number of models to be compared) as of â input_dfâ . Values in â output_dfâ are corresponding â NAâ , â MINâ or â MAXâ . Finally, the column containing minimum number of â NAâ values is denoted as the best column. â min_NA_colâ gives the name of the best column (model). â min_NA_valuesâ are the corresponding metrics values. âBestColumn_metricsâ is the data frame (dimension: 1*8) containing different metrics of the best column (model). â best_column_resultsâ is the final result (a list) containing all of these output elements. In special case, if two columns having equal NA', it will be checked among these two column which one is having least NA in first five rows and will be inferred as the best. More details about AllMetrics can be found in Garai (2023) <doi:10.13140/RG.2.2.18688.30723>.
The Cauchy Process can model pulsed continuous trait evolution on phylogenies. The likelihood is tractable, and is used for parameter inference and ancestral trait reconstruction. See Bastide and Didier (2023) <doi:10.1093/sysbio/syad053>.
This package provides a flexible, extendable representation of an ecological community and a range of functions for analysis and visualisation, focusing on food web, body mass and numerical abundance data. Allows inter-web comparisons such as examining changes in community structure over environmental, temporal or spatial gradients.
CHAP-GWAS (Chromosomal Haplotype-Integrated Genome-Wide Association Study) provides a dynamically adaptive framework for genome-wide association studies (GWAS) that integrates chromosome-scale haplotypes with single nucleotide polymorphism (SNP) analysis. The method identifies and extends haplotype variants based on their phenotypic associations rather than predefined linkage blocks, enabling high-resolution detection of quantitative trait loci (QTL). By leveraging long-range phased haplotype information, CHAP-GWAS improves statistical power and offers a more comprehensive view of the genetic architecture underlying complex traits.
This package provides a set of tools for evaluating clustering robustness using proportion of ambiguously clustered pairs (Senbabaoglu et al. (2014) <doi:10.1038/srep06207>), as well as similarity across methods and method stability using element-centric clustering comparison (Gates et al. (2019) <doi:10.1038/s41598-019-44892-y>). Additionally, this package enables stability-based parameter assessment for graph-based clustering pipelines typical in single-cell data analysis.
Fast categorization of items based on external code data identified by regular expressions. A typical use case considers patient with medically coded data, such as codes from the International Classification of Diseases ('ICD') or the Anatomic Therapeutic Chemical ('ATC') classification system. Functions of the package relies on a triad of objects: (1) case data with unit id:s and possible dates of interest; (2) external code data for corresponding units in (1) and with optional dates of interest and; (3) a classification scheme ('classcodes object) with regular expressions to identify and categorize relevant codes from (2). It is easy to introduce new classification schemes ('classcodes objects) or to use default schemes included in the package. Use cases includes patient categorization based on comorbidity indices such as Charlson', Elixhauser', RxRisk V', or the comorbidity-polypharmacy score (CPS), as well as adverse events after hip and knee replacement surgery.
This package provides an implementation of â Curricular Analyticsâ , a framework for analyzing and quantifying the complexity of academic curricula. Curricula are modelled as directed acyclic graphs and analytics are provided based on path lengths and edge density. This work directly comes from Heileman et al. (2018) <doi:10.48550/arXiv.1811.09676>.
Encrypts and decrypts strings using either the Caesar cipher or a pseudorandom number generation (using set.seed()) method.
In computationally demanding analysis projects, statisticians and data scientists asynchronously deploy long-running tasks to distributed systems, ranging from traditional clusters to cloud services. The crew.cluster package extends the mirai'-powered crew package with worker launcher plugins for traditional high-performance computing systems. Inspiration also comes from packages mirai by Gao (2023) <https://github.com/r-lib/mirai>, future by Bengtsson (2021) <doi:10.32614/RJ-2021-048>, rrq by FitzJohn and Ashton (2023) <https://github.com/mrc-ide/rrq>, clustermq by Schubert (2019) <doi:10.1093/bioinformatics/btz284>), and batchtools by Lang, Bischl, and Surmann (2017). <doi:10.21105/joss.00135>.
This package provides a collection of functions to generate a large variety of structures in high dimensions. These data structures are useful for testing, validating, and improving algorithms used in dimensionality reduction, clustering, machine learning, and visualization.
This package implements the three-step workflow for robust analysis of change in two repeated measurements of continuous outcomes, described in Ning et al. (in press), "Robust estimation of the effect of an exposure on the change in a continuous outcome", BMC Medical Research Methodology.
Process command line arguments, as part of a data analysis workflow. command makes it easier to construct a workflow consisting of lots of small, self-contained scripts, all run from a Makefile or shell script. The aim is a workflow that is modular, transparent, and reliable.
This package provides tools for connecting to CHILDES', an open repository for transcripts of parent-child interaction. For more information on the underlying data, see <https://langcog.github.io/childes-db-website/>.
Dissects a package environment or covr coverage object in order to cross reference tested code with the lines that are evaluated, as well as linking those evaluated lines to the documentation that they are described within. Connecting these three pieces of information provides a mechanism of linking tests to documented behaviors.
This package provides a lightweight data validation and testing toolkit for R. Its guiding philosophy is that adding code-based data checks to users existing workflow should be both quick and intuitive. The suite of functions included therefore mirror the common data checks many users already perform by hand or by eye. Additionally, the checkthat package is optimized to work within tidyverse data manipulation pipelines.