Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel search send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
This package provides a computational framework for analyzing mutations in immunoglobulin (Ig) sequences. Includes methods for Bayesian estimation of antigen-driven selection pressure, mutational load quantification, building of somatic hypermutation (SHM) models, and model-dependent distance calculations. Also includes empirically derived models of SHM for both mice and humans. Citations: Gupta and Vander Heiden, et al (2015) <doi:10.1093/bioinformatics/btv359>, Yaari, et al (2012) <doi:10.1093/nar/gks457>, Yaari, et al (2013) <doi:10.3389/fimmu.2013.00358>, Cui, et al (2016) <doi:10.4049/jimmunol.1502263>.
Given independent and identically distributed observations X(1), ..., X(n) from a Generalized Pareto distribution with shape parameter gamma in [-1,0], offers several estimates to compute estimates of gamma. The estimates are based on the principle of replacing the order statistics by quantiles of a distribution function based on a log--concave density function. This procedure is justified by the fact that the GPD density is log--concave for gamma in [-1,0].
This package implements the Shimazaki-Shinomoto method for optimizing the bin width of a histogram. This method minimizes the mean integrated squared error (MISE) and features a C++ backend for high performance and shift-averaging to remove edge-position bias. Ideally suits for time-dependent rate estimation and identifying intrinsic data structures. Supports both 1D and 2D data distributions. For more details see Shimazaki and Shinomoto (2007) "A Method for Selecting the Bin Size of a Time Histogram" <doi:10.1162/neco.2007.19.6.1503>.
Load and export SomaScan data via the SomaLogic Operating Co., Inc. structured text file called an ADAT ('*.adat'). For file format see <https://github.com/SomaLogic/SomaLogic-Data/blob/main/README.md>. The package also exports auxiliary functions for manipulating, wrangling, and extracting relevant information from an ADAT object once in memory.
Includes all the datasets of Sampling: Design and Analysis (3rd edition by Sharon Lohr) in R format and additional functions for analyzing and graphing probability samples.
Translates antibody levels measured in cross-sectional population samples into estimates of the frequency with which seroconversions (infections) occur in the sampled populations. Replaces the previous `seroincidence` package.
This package provides a client for running SPARQL queries directly from R. SPARQL (short for SPARQL Protocol and RDF Query Language) is a query language used to retrieve and manipulate data stored in RDF (Resource Description Framework) format.
Surveys to collect employment data so as to obtain data estimates on the number of employed people, the number of unemployed, and other employment indicators.
This package provides methods for the analysis of signed networks. This includes several measures for structural balance as introduced by Cartwright and Harary (1956) <doi:10.1037/h0046049>, blockmodeling algorithms from Doreian (2008) <doi:10.1016/j.socnet.2008.03.005>, various centrality indices, and projections of signed two-mode networks introduced by Schoch (2020) <doi:10.1080/0022250X.2019.1711376>.
This package provides a collection of sample datasets on various fields such as automotive performance and safety data to historical demographics and socioeconomic indicators, as well as recreational data. It serves as a resource for researchers and analysts seeking to perform analyses and derive insights from classic data sets in R.
Implementation of hybrid STL decomposition based time delay neural network model for univariate time series forecasting. For method details see Jha G K, Sinha, K (2014). <doi:10.1007/s00521-012-1264-z>, Xiong T, Li C, Bao Y (2018). <doi:10.1016/j.neucom.2017.11.053>.
This package implements SplitWise', a hybrid regression approach that transforms numeric variables into either single-split (0/1) dummy variables or retains them as continuous predictors. The transformation is followed by stepwise selection to identify the most relevant variables. The default iterative mode adaptively explores partial synergies among variables to enhance model performance, while an alternative univariate mode applies simpler transformations independently to each predictor. For details, see Kurbucz et al. (2025) <doi:10.48550/arXiv.2505.15423>.
Implementations of classical and machine learning models for survival analysis, including deep neural networks via keras and tensorflow'. Each model includes a separated fit and predict interface with consistent prediction types for predicting risk or survival probabilities. Models are either implemented from Python via reticulate <https://CRAN.R-project.org/package=reticulate>, from code in GitHub packages, or novel implementations using Rcpp <https://CRAN.R-project.org/package=Rcpp>. Neural networks are implemented from the Python package pycox <https://github.com/havakv/pycox>.
This sparklyr extension makes Flint time series library functionalities (<https://github.com/twosigma/flint>) easily accessible through R.
Consolidated data simulation, sample size calculation and analysis functions for several snSMART (small sample sequential, multiple assignment, randomized trial) designs under one library. See Wei, B., Braun, T.M., Tamura, R.N. and Kidwell, K.M. "A Bayesian analysis of small n sequential multiple assignment randomized trials (snSMARTs)." (2018) Statistics in medicine, 37(26), pp.3723-3732 <doi:10.1002/sim.7900>.
Datasets for the textbook Stat2: Modeling with Regression and ANOVA (second edition). The package also includes data for the first edition, Stat2: Building Models for a World of Data and a few functions for plotting diagnostics.
This package implements different kinds of bootstraps to estimate sampling variation from survey data with complex designs. Includes the rescaled bootstrap described in Rust and Rao (1996) <doi:10.1177/096228029600500305> and Rao and Wu (1988) <doi:10.1080/01621459.1988.10478591>.
Proposes application of spectral analysis and jack-knife resampling for multivariate sequence forecasting. The application allows for a fast random search in a compact space of hyper-parameters composed by Sequence Length and Jack-Knife Leave-N-Out.
This package provides functions that wrap HTML Bootstrap components code to enable the design and layout of informative landing home pages for Shiny applications. This can lead to a better user experience for the users and writing less HTML for the developer.
Set of functions that access information about deputies and votings in Polish diet from webpage <http://www.sejm.gov.pl>. The package was developed as a result of an internship in MI2 Group - <http://mi2.mini.pw.edu.pl>, Faculty of Mathematics and Information Science, Warsaw University of Technology.
In Shiny apps, it is sometimes useful to see a plot or a table in full screen. Using Shinyfullscreen', you can easily designate the HTML elements that can be displayed on fullscreen and use buttons to trigger the fullscreen view.
This package provides functions to calculate and visualise the Regional Consistency Probability (RCP) for single-arm multi-regional clinical trials (MRCTs) using the Effect Retention Approach (ERA). Six endpoint types are supported: continuous, binary, count (negative binomial), time-to-event via hazard ratio, milestone survival, and restricted mean survival time (RMST). For each endpoint, both a closed-form (or semi-analytical) solution and a Monte Carlo simulation approach are implemented. Two consistency evaluation methods are available: Method 1 (effect retention in Region 1 relative to the overall population) and Method 2 (simultaneous positive effect across all regions). Plotting functions generate faceted visualisations of RCP as a function of the regional allocation proportion, overlaying formula and simulation results for direct comparison. The methodology follows the Japanese MHLW guidelines for MRCTs. Abbreviations used: RCP (Regional Consistency Probability), MRCT (Multi-Regional Clinical Trial), RMST (Restricted Mean Survival Time), MHLW (Ministry of Health, Labour and Welfare).
This package provides useful UI components and input widgets for Shiny applications. The offered components allow to apply non-standard operations and view to your Shiny application, but also help to overcome common performance issues.
Seven different methods for multiple testing problems. The SGoF-type methods (see for example, Carvajal Rodrà guez et al., 2009 <doi:10.1186/1471-2105-10-209>; de Uña à lvarez, 2012 <doi:10.1515/1544-6115.1812>; Castro Conde et al., 2015 <doi:10.1177/0962280215597580>) and the BH and BY false discovery rate controlling procedures.