Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Allows search and visualisation of a collection of uniformly processed skeletal transcriptomic datasets. Includes methods to identify datasets where genes of interest are differentially expressed and find datasets with a similar gene expression pattern to a query dataset Soul J, Hardingham TE, Boot-Handford RP, Schwartz JM (2019) <doi:10.1093/bioinformatics/bty947>.
This package provides a toolbox that allows the user to implement accessibility related concepts.
This package provides methods for inference using stacked multiple imputations augmented with weights. The vignette provides example R code for implementation in general multiple imputation settings. For additional details about the estimation algorithm, we refer the reader to Beesley, Lauren J and Taylor, Jeremy M G (2020) â A stacked approach for chained equations multiple imputation incorporating the substantive modelâ <doi:10.1111/biom.13372>, and Beesley, Lauren J and Taylor, Jeremy M G (2021) â Accounting for not-at-random missingness through imputation stackingâ <arXiv:2101.07954>.
This package provides a search interface to look up terms on Google', Bing', DuckDuckGo', Startpage', Ecosia', rseek', Twitter', StackOverflow', RStudio Community', GitHub', and BitBucket'. Upon searching, a browser window will open with the aforementioned search results.
Extension to the spatstat family of packages, for analysing large datasets of spatial points on a network. The geometrically- corrected K function is computed using a memory-efficient tree-based algorithm described by Rakshit, Baddeley and Nair (2019).
To calculate the standard error of measurement (SEM) to assess the observer variability (inter- and intra-observer variation). The methods used in this package are referenced from Zoran B. PopoviÄ (2017) <doi:10.21037/cdt.2017.03.12>.
Building predictive models with stacking which is a type of ensemble learning. Learners can be specified from those implemented in caret'. For more information of the package, see Nukui and Onogi (2023) <doi:10.1101/2023.06.06.543970>.
Automates the creation of Dockerfiles for deploying Shiny applications. By integrating with renv for dependency management and leveraging Docker-based solutions, it simplifies the process of containerizing Shiny apps, ensuring reproducibility and consistency across different environments. Additionally, it facilitates the setup of CI/CD pipelines for building Docker images on both GitLab and GitHub.
Set of tools to fit a linear multiple or semi-parametric regression models with the possibility of non-informative random right or left censoring. Under this setup, the localization parameter of the response variable distribution is modeled by using linear multiple regression or semi-parametric functions, whose non-parametric components may be approximated by natural cubic spline or P-splines. The supported distribution for the model error is a generalized log-gamma distribution which includes the generalized extreme value and standard normal distributions as important special cases. Inference is based on likelihood, penalized likelihood and bootstrap methods. Lastly, some numerical and graphical devices for diagnostic of the fitted models are offered.
Search contributed R packages, sort by package.
This tool fits a non-parametric Bayesian model called a "hierarchically coupled mixture model with local dependence (HCMM-LD)" to the original microdata in order to generate synthetic microdata for privacy protection. The non-parametric feature of the adopted model is useful for capturing the joint distribution of the original input data in a highly flexible manner, leading to the generation of synthetic data whose distributional features are similar to that of the input data. The package allows the original input data to have missing values and impute them with the posterior predictive distribution, so no missing values exist in the synthetic data output. The method builds on the work of Murray and Reiter (2016) <doi:10.1080/01621459.2016.1174132>.
Implementation of the structural model for variances in order to detect differentially expressed genes from gene expression data.
Estimate the regression coefficients and the baseline hazard of proportional hazard Cox models with left, right or interval censored survival data using maximum penalised likelihood. A non-parametric smooth estimate of the baseline hazard function is provided.
Aggregates large single-cell data into metacell dataset by merging together gene expression of very similar cells. SuperCell uses velocyto.R <doi:10.1038/s41586-018-0414-6> <https://github.com/velocyto-team/velocyto.R> for RNA velocity and WeightedCluster <doi:10.12682/lives.2296-1658.2013.24> <https://mephisto.unige.ch/weightedcluster/> for weighted clustering on metacells. We also recommend installing scater Bioconductor package <doi:10.18129/B9.bioc.scater> <https://bioconductor.org/packages/release/bioc/html/scater.html>.
Import data from the STATcube REST API or from the open data portal of Statistics Austria. This package includes a client for API requests as well as parsing utilities for data which originates from STATcube'. Documentation about STATcubeR is provided by several vignettes included in the package as well as on the public pkgdown page at <https://statistikat.github.io/STATcubeR/>.
This package provides functions to compute standardized differences for numeric, binary, and categorical variables on Apache Spark DataFrames using sparklyr'. The implementation mirrors the methods used in the stddiff package but operates on distributed data. See Zhicheng Du, Yuantao Hao (2022) <doi:10.32614/CRAN.package.stddiff> for reference.
Extends the functionality of the package Synth as detailed in Abadie, Diamond, and Hainmueller (2011) <doi:10.18637/jss.v042.i13>. Includes generating and plotting placebos, post/pre-MSPE (Mean Squared Prediction Error) significance tests and plots, and calculating average treatment effects for multiple treated units.
Web front end for your R functions producing plots or tables. If you have a function or set of related functions, you can make them available over the internet through a web browser. This is the same motivation as the shiny package, but note that the development of shinylight is not in any way linked to that of shiny (beyond the use of the httpuv package). You might prefer shinylight to shiny if you want a lighter weight deployment with easier horizontal scaling, or if you want to develop your front end yourself in JavaScript and HTML just using a lightweight remote procedure call interface to your R code on the server.
Fit a regularized generalized linear model via penalized maximum likelihood. The model is fit for a path of values of the penalty parameter. Fits linear, logistic and Cox models.
Computes the Exposure-At-Default based on the standardized approach of CRR2 (SA-CCR). The simplified version of SA-CCR has been included, as well as the OEM methodology. Multiple trade types of all the five major asset classes are being supported including the Other Exposure and, given the inheritance- based structure of the application, the addition of further trade types is straightforward. The application returns a list of trees per Counterparty and CSA after automatically separating the trades based on the Counterparty, the CSAs, the hedging sets, the netting sets and the risk factors. The basis and volatility transactions are also identified and treated in specific hedging sets whereby the corresponding penalty factors are applied. All the examples appearing on the regulatory papers (both for the margined and the unmargined workflow) have been implemented including the latest CRR2 developments.
Set of functions to quantify and map the behaviour of winds generated by tropical storms and cyclones in space and time. It includes functions to compute and analyze fields such as the maximum sustained wind field, power dissipation index and duration of exposure to winds above a given threshold. It also includes functions to map the trajectories as well as characteristics of the storms.
This package provides a general framework for statistical simulation, which allows researchers to make use of a wide range of simulation designs with minimal programming effort. The package provides functionality for drawing samples from a distribution or a finite population, for adding outliers and missing values, as well as for visualization of the simulation results. It follows a clear object-oriented design and supports parallel computing to increase computational performance.
This package provides an S4 class for representing and interacting with sparse plus rank matrices. At the moment the implementation is quite spare, but the plan is eventually subclass Matrix objects.
In stability selection (N Meinshausen, P Bühlmann (2010) <doi:10.1111/j.1467-9868.2010.00740.x>) and consensus clustering (S Monti et al (2003) <doi:10.1023/A:1023949509487>), resampling techniques are used to enhance the reliability of the results. In this package (B Bodinier et al (2025) <doi:10.18637/jss.v112.i05>), hyper-parameters are calibrated by maximising model stability, which is measured under the null hypothesis that all selection (or co-membership) probabilities are identical (B Bodinier et al (2023a) <doi:10.1093/jrsssc/qlad058> and B Bodinier et al (2023b) <doi:10.1093/bioinformatics/btad635>). Functions are readily implemented for the use of LASSO regression, sparse PCA, sparse (group) PLS or graphical LASSO in stability selection, and hierarchical clustering, partitioning around medoids, K means or Gaussian mixture models in consensus clustering.