Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Variable and interaction selection are essential to classification in high-dimensional setting. In this package, we provide the implementation of SODA procedure, which is a forward-backward algorithm that selects both main and interaction effects under logistic regression and quadratic discriminant analysis. We also provide an extension, S-SODA, for dealing with the variable selection problem for semi-parametric models with continuous responses.
Bundles functions used to analyze the harmfulness of trial errors in criminal trials. Functions in the Scientific Analysis of Trial Errors ('sate') package help users estimate the probability that a jury will find a defendant guilty given jurors preferences for a guilty verdict and the uncertainty of that estimate. Users can also compare actual and hypothetical trial conditions to conduct harmful error analysis. The conceptual framework is discussed by Barry Edwards, A Scientific Framework for Analyzing the Harmfulness of Trial Errors, UCLA Criminal Justice Law Review (2024) <doi:10.5070/CJ88164341> and Barry Edwards, If The Jury Only Knew: The Effect Of Omitted Mitigation Evidence On The Probability Of A Death Sentence, Virginia Journal of Social Policy & the Law (2025) <https://vasocialpolicy.org/wp-content/uploads/2025/05/Edwards-If-The-Jury-Only-Knew.pdf>. The relationship between individual jurors verdict preferences and the probability that a jury returns a guilty verdict has been studied by Davis (1973) <doi:10.1037/h0033951>; MacCoun & Kerr (1988) <doi:10.1037/0022-3514.54.1.21>, and Devine et el. (2001) <doi:10.1037/1076-8971.7.3.622>, among others.
This package creates a contextual menu that can be triggered with keyboard shortcuts or programmatically. This can replace traditional sidebars or navigation bars, thereby enhancing the user experience with lighter user interfaces.
This package implements the Stratigraphic Plug Alignment (SPA) procedure for integrating sparsely sampled plug-based measurements (e.g., total organic carbon, porosity, mineralogy) with high-resolution X-ray fluorescence (XRF) geochemical data. SPA uses linear interpolation via the base approx() function with constrained extrapolation (rule = 1) to preserve stratigraphic order and avoid estimation beyond observed depths. The method aligns all datasets to a common depth grid, enabling high-resolution multivariate analysis and stratigraphic interpretation of core-based datasets such as those from the Utica and Point Pleasant formations. See R Core Team (2025) <https://stat.ethz.ch/R-manual/R-devel/library/stats/html/stats-package.html> and Omodolor (2025) <http://rave.ohiolink.edu/etdc/view?acc_num=case175262671767524> for methodological background and geological context.
Recently, regularized variable selection has emerged as a powerful tool to identify and dissect gene-environment interactions. Nevertheless, in longitudinal studies with high dimensional genetic factors, regularization methods for GÃ E interactions have not been systematically developed. In this package, we provide the implementation of sparse group variable selection, based on both the quadratic inference function (QIF) and generalized estimating equation (GEE), to accommodate the bi-level selection for longitudinal GÃ E studies with high dimensional genomic features. Alternative methods conducting only the group or individual level selection have also been included. The core modules of the package have been developed in C++.
The Hypothesis tests for the means of independent or paired groups. This package investigates the normality assumption automatically. Then, it tests the hypothesis tests for two independent or paired group means by using parametric or non-parametric tests. It uses the Shapiro-Wilk test to test the normality assumption. For independent two groups, If data comes from the normal distribution, the package uses the Z or t-test according to whether variances are known. For paired groups, it uses paired t-test under normal data sets. If data does not come from the normal distribution, the package uses the Wilcoxon test for independent and paired cases.
Set of tools aimed at wrapping some of the functionalities of the packages tools, utils and codetools into a nicer format so that an IDE can use them.
This package provides a set of functions used in teaching STATS 201/208 Data Analysis at the University of Auckland. The functions are designed to make parts of R more accessible to a large undergraduate population who are mostly not statistics majors.
This package provides a rich set of UI components for building Shiny applications, including inputs, containers, overlays, menus, and various utilities. All components from Fluent UI (the underlying JavaScript library) are available and have usage examples in R.
Computes confidence intervals for variance using the Chi-Square distribution, without requiring raw data. Wikipedia (2025) <https://en.wikipedia.org/wiki/Chi-squared_distribution>. All-in-One Chi Distribution CI provides functions to calculate confidence intervals for the population variance based on a chi-squared distribution, utilizing a sample variance and sample size. It offers only a simple all-in-one method for quick calculations to find the CI for Chi Distribution.
You can use the functions provided by the package to make various statistical tables, such as baseline data tables. Creates Table 1', i.e., a description of the baseline patient characteristics, which is essential in every medical research. Supports both continuous and categorical variables, as well as p-values and standardized mean differences. This method was described by Mary L McHugh (2013) <doi:10.11613/bm.2013.018>.
Uncertainty propagation analysis in spatial environmental modelling following methodology described in Heuvelink et al. (2007) <doi:10.1080/13658810601063951> and Brown and Heuvelink (2007) <doi:10.1016/j.cageo.2006.06.015>. The package provides functions for examining the uncertainty propagation starting from input data and model parameters, via the environmental model onto model outputs. The functions include uncertainty model specification, stochastic simulation and propagation of uncertainty using Monte Carlo (MC) techniques. Uncertain variables are described by probability distributions. Both numerical and categorical data types are handled. Spatial auto-correlation within an attribute and cross-correlation between attributes is accommodated for. The MC realizations may be used as input to the environmental models called from R, or externally.
This package provides a meta-package that loads the complete sitrep ecosystem for applied epidemiology analysis. This package provides report templates and automatically loads companion packages, including epitabulate (for epidemiological tables), epidict (for data dictionaries), epikit (for epidemiological utilities), and apyramid (for age-sex pyramids). Simply load sitrep to access all functions from the ecosystem.
Utilizes the Reliability-Adjusted Product Indicator (RAPI) method to estimate effects among latent variables, thus allowing for more precise definition and analysis of mediation and moderation models. Our simulation studies reveal that while silp may exhibit instability with smaller sample sizes and lower reliability scores (e.g., N = 100, omega = 0.7), implementing nearest positive definite matrix correction and bootstrap confidence interval estimation can significantly ameliorate this volatility. When these adjustments are applied, silp achieves estimations akin in quality to those derived from LMS. In conclusion, the silp package is a valuable tool for researchers seeking to explore complex relational structures between variables without resorting to commercial software. Cheung et al.(2021)<doi:10.1007/s10869-020-09717-0> Hsiao et al.(2018)<doi:10.1177/0013164416679877>.
This gadget allows you to use the recipes package belonging to tidymodels to carry out the data preprocessing tasks in an interactive way. Build your recipe by dragging the variables, visually analyze your data to decide which steps to use, add those steps and preprocess your data.
Extract glyph information from font data, and translate the outline curves to flattened paths or tessellated polygons. The converted data is returned as a data.frame in easy-to-plot format.
Implementation of prediction and inference procedures for Synthetic Control methods using least square, lasso, ridge, or simplex-type constraints. Uncertainty is quantified with prediction intervals as developed in Cattaneo, Feng, and Titiunik (2021) <doi:10.1080/01621459.2021.1979561> for a single treated unit and in Cattaneo, Feng, Palomba, and Titiunik (2025) <doi:10.1162/rest_a_01588> for multiple treated units and staggered adoption. More details about the software implementation can be found in Cattaneo, Feng, Palomba, and Titiunik (2025) <doi:10.18637/jss.v113.i01>.
The SALSO algorithm is an efficient randomized greedy search method to find a point estimate for a random partition based on a loss function and posterior Monte Carlo samples. The algorithm is implemented for many loss functions, including the Binder loss and a generalization of the variation of information loss, both of which allow for unequal weights on the two types of clustering mistakes. Efficient implementations are also provided for Monte Carlo estimation of the posterior expected loss of a given clustering estimate. See Dahl, Johnson, Müller (2022) <doi:10.1080/10618600.2022.2069779>.
This package provides a consistent interface to use various methods to calculate the periodogram and estimate the period of a rhythmic time-course. Methods include Lomb-Scargle, fast Fourier transform, and three versions of the chi-square periodogram. See Tackenberg and Hughey (2021) <doi:10.1371/journal.pcbi.1008567>.
Visualization and analysis of spatially resolved transcriptomics data. The spatialGE R package provides methods for visualizing and analyzing spatially resolved transcriptomics data, such as 10X Visium, CosMx, or csv/tsv gene expression matrices. It includes tools for spatial interpolation, autocorrelation analysis, tissue domain detection, gene set enrichment, and differential expression analysis using spatial mixed models.
This package provides tools to check variables contained in the user environment, and inspect the currently loaded package namespaces. The intended use is to allow user scripts to throw errors or warnings if unwanted variables exist or if unwanted packages are loaded.
This package creates and fits staged event tree probability models, which are probabilistic graphical models capable of representing asymmetric conditional independence statements for categorical variables. Includes functions to create, plot and fit staged event trees from data, as well as many efficient structure learning algorithms. References: Carli F, Leonelli M, Riccomagno E, Varando G (2022). <doi: 10.18637/jss.v102.i06>. Collazo R. A., Görgen C. and Smith J. Q. (2018, ISBN:9781498729604). Görgen C., Bigatti A., Riccomagno E. and Smith J. Q. (2018) <arXiv:1705.09457>. Thwaites P. A., Smith, J. Q. (2017) <arXiv:1510.00186>. Barclay L. M., Hutton J. L. and Smith J. Q. (2013) <doi:10.1016/j.ijar.2013.05.006>. Smith J. Q. and Anderson P. E. (2008) <doi:10.1016/j.artint.2007.05.004>.
Lightweight helpers for connecting to Microsoft SQL Server using DBI', odbc', and pool'. Provides simple wrappers for building connection arguments, establishing connections, and safely disconnecting.
Supporting materials for a course and book on data visualization. It contains utility functions for graphs and several sample data sets. See Healy (2019) <ISBN 978-0691181622>.