Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
This package provides functions for fitting a linear regression model with ARIMA errors using a filtered tau-estimate. The methodology is described in Maronna et al (2017, ISBN:9781119214687).
Fits the robust Bayesian Copas (RBC) selection model of Bai et al. (2020) <arXiv:2005.02930> for correcting and quantifying publication bias in univariate meta-analysis. Also fits standard random effects meta-analysis and the Copas-like selection model of Ning et al. (2017) <doi:10.1093/biostatistics/kxx004>.
Calculates robust Matthews Correlation Coefficient (MCC) and robust F-Beta Scores, as introduced by Holzmann and Klar (2024) <doi:10.48550/arXiv.2404.07661>. These performance metrics are designed for imbalanced classification problems. Plots the receiver operating characteristic curve (ROC curve) together with the recall / 1-precision curve.
Queries data from RDAP servers.
Analyses sentiment of a sentence in English and assigns score to it. It can classify sentences to the following categories of sentiments:- Positive, Negative, very Positive, very negative, Neutral. For a vector of sentences, it counts the number of sentences in each category of sentiment.In calculating the score, negation and various degrees of adjectives are taken into consideration. It deals only with English sentences.
Providing just one primary function, readit uses a set of reasonable heuristics to apply the appropriate reader function to the given file path. As long as the data file has an extension, and the data is (or can be coerced to be) rectangular, readit() can probably read it.
Rcpp Bindings for the C code of the Corpus Workbench ('CWB'), an indexing and query engine to efficiently analyze large corpora (<https://cwb.sourceforge.io>). RcppCWB is licensed under the GNU GPL-3, in line with the GPL-3 license of the CWB (<https://www.r-project.org/Licenses/GPL-3>). The CWB relies on pcre2 (BSD license, see <https://github.com/PCRE2Project/pcre2/blob/master/LICENCE.md>) and GLib (LGPL license, see <https://www.gnu.org/licenses/lgpl-3.0.en.html>). See the file LICENSE.note for further information. The package includes modified code of the rcqp package (GPL-2, see <https://cran.r-project.org/package=rcqp>). The original work of the authors of the rcqp package is acknowledged with great respect, and they are listed as authors of this package. To achieve cross-platform portability (including Windows), using Rcpp for wrapper code is the approach used by RcppCWB'.
This package provides tools to read various file types into one list of data structures, usually, but not limited to, data frames. Excel files are read sheet-wise, i.e., all or a selection of sheets can be read. Field delimiters and decimal separators are determined automatically.
Ports the Ripser <doi:10.48550/arXiv.1908.02518> and Cubical Ripser <doi:10.48550/arXiv.2005.12692> persistent homology calculation engines from C++. Can be used as a rapid calculation tool in topological data analysis pipelines.
This package provides interface to Google Fit REST API v1 (see <https://developers.google.com/fit/rest/v1/reference/>).
This package provides a portable Shiny tool to explore patient-level electronic health record data and perform chart review in a single integrated framework. This tool supports browsing clinical data in many different formats including multiple versions of the OMOP common data model as well as the MIMIC-III data model. In addition, chart review information is captured and stored securely via the Shiny interface in a REDCap (Research Electronic Data Capture) project using the REDCap API. See the ReviewR website for additional information, documentation, and examples.
It enables the identification of sequentialexperimentation orders for factorial designs that jointly reduce bias and the number of level changes. The method used is that presented by Conto et al. (2025), known as the Assignment-Expansion method, which consists of adapting the linear programming assignment problem to generate balanced experimentation orders. The properties identified are then generalized to designs with a larger number of factors and levels using the expansion method proposed by Correa et al. (2009) and later generalized by Bhowmik et al. (2017). For more details see Conto et al. (2025) <doi:10.1016/j.cie.2024.110844>, Correa et al. (2009) <doi:10.1080/02664760802499337> and Bhowmik et al. (2017) <doi:10.1080/03610926.2016.1152490>.
This package provides functions to identify Homozygous-by-Descent (HBD) segments associated with runs of homozygosity (ROH) and to estimate individual autozygosity (or inbreeding coefficient). HBD segments and autozygosity are assigned to multiple HBD classes with a model-based approach relying on a mixture of exponential distributions. The rate of the exponential distribution is distinct for each HBD class and defines the expected length of the HBD segments. These HBD classes are therefore related to the age of the segments (longer segments and smaller rates for recent autozygosity / recent common ancestor). The functions allow to estimate the parameters of the model (rates of the exponential distributions, mixing proportions), to estimate global and local autozygosity probabilities and to identify HBD segments with the Viterbi decoding. The method is fully described in Druet and Gautier (2017) <doi:10.1111/mec.14324> and Druet and Gautier (2022) <doi:10.1016/j.tpb.2022.03.001>.
This package performs exact rate ratio tests.
Graphics for statistics on a sphere, as applied to geological fault data, crystallography, earthquake focal mechanisms, radiation patterns, ternary plots and geographical/geological maps. Non-double couple plotting of focal spheres and source type maps are included for statistical analysis of moment tensors.
This package provides functions for generating k-record values and k-record times.
Cross validate large genetic data while specifying clinical variables that should always be in the model using the function cv(). An ROC plot from the cross validation data with AUC can be obtained using rocplot(), which also can be used to compare different models. Framework was built to handle genetic data, but works for any data.
This package provides a task-oriented R interface to the RDKit <https://www.rdkit.org> library through its Python API via reticulate'. The package offers high-level cheminformatics functionality, including molecule parsing, descriptor calculation, and fingerprint generation without replicating the native structure of RDKit'.
Robust estimators for the beta regression, useful for modeling bounded continuous data. Currently, four types of robust estimators are supported. They depend on a tuning constant which may be fixed or selected by a data-driven algorithm also implemented in the package. Diagnostic tools associated with the fitted model, such as the residuals and goodness-of-fit statistics, are implemented. Robust Wald-type tests are available. More details about robust beta regression are described in Maluf et al. (2025) <doi:10.1007/s00184-024-00949-1>.
Semi-Automated Marketing Mix Modeling (MMM) aiming to reduce human bias by means of ridge regression and evolutionary algorithms, enables actionable decision making providing a budget allocation and diminishing returns curves and allows ground-truth calibration to account for causation.
We provide several avenues to predict and account for user-based mortality and tag loss during mark-recapture studies. When planning a study on a target species, the retentionmort_generation() function can be used to produce multiple synthetic mark-recapture datasets to anticipate the error associated with a planned field study to guide method development to reduce error. Similarly, if field data was already collected, the retentionmort() function can be used to predict the error from already generated data to adjust for user-based mortality and tag loss. The test_dataset_retentionmort() function will provide an example dataset of how data should be inputted into the function to run properly. Lastly, the retentionmort_figure() function can be used on any dataset generated from either model function to produce an rmarkdown printout of preliminary analysis associated with the model, including summary statistics and figures. Methods and results pertaining to the formation of this package can be found in McCutcheon et al. (in review, "Predicting tagging-related mortality and tag loss during mark-recapture studies").
The evaluation criteria of rangeland health, condition and landscape function analysis based on species diversity and functional diversity of rangeland plant communities.
Predict fish year-class strength by calibration regression analysis of multiple recruitment index series.
The rearrangement operator (Hardy, Littlewood, and Polya 1952) for univariate, bivariate, and trivariate point estimates of monotonic functions. The package additionally provides a function that creates simultaneous confidence intervals for univariate functions and applies the rearrangement operator to these confidence intervals.