Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Multi-environment genomic prediction for training and test environments using penalized factorial regression. Predictions are made using genotype-specific environmental sensitivities as in Millet et al. (2019) <doi:10.1038/s41588-019-0414-y>.
This package provides a compositional statistical framework for absolute proportion estimation between fractions in RNA sequencing data. FracFixR addresses the fundamental challenge in fractionated RNA-seq experiments where library preparation and sequencing depth obscure the original proportions of RNA fractions. It reconstructs original fraction proportions using non-negative linear regression, estimates the "lost" unrecoverable fraction, corrects individual transcript frequencies, and performs differential proportion testing between conditions. Supports any RNA fractionation protocol including polysome profiling, sub-cellular localization, and RNA-protein complex isolation.
Designed to streamline the process of analyzing genotyping data from Fluidigm machines, this package offers a suite of tools for data handling and analysis. It includes functions for converting Fluidigm data to format used by PLINK', estimating errors, calculating pairwise similarities, determining pairwise similarity loci, and generating a similarity matrix.
Computes likelihood ratios based on pigmentation traits. Also, it allows computing conditional probabilities for unidentified individuals based on missing person characteristics. A set of tailored plots are incorporated to analyze likelihood ratio distributions.
Extends the capabilities for flexible partitioning and model-based clustering available in the packages flexclust and flexmix to handle ordinal and mixed-with-ordinal data types via new distance, centroid and driver functions that make various assumptions regarding ordinality. Using them within the flex-scheme allows for easy comparisons across methods.
An interface to the fastText library <https://github.com/facebookresearch/fastText>. The package can be used for text classification and to learn word vectors. An example how to use fastTextR can be found in the README file.
Basic analysis of all penalties taken in the German men's Bundesliga between the start of its inaugural season and May 2017. The main functions are suitable printing and plotting functions. Flexible selection of a player is supported via grep. Missed penalties can easily be included or excluded, depending on the user's wishes.
This package implements the method of Hofmeyr, D.P. (2021) <DOI:10.1109/TPAMI.2019.2930501> for fast evaluation of univariate kernel smoothers based on recursive computations. Applications to the basic problems of density and regression function estimation are provided, as well as some projection pursuit methods for which the objective is based on non-parametric functionals of the projected density, or conditional density of a response given projected covariates. The package is accompanied by an instructive paper in the Journal of Statistical Software <doi:10.18637/jss.v101.i03>.
This package creates a HTML widget which displays the results of searching for a pattern in files in a given folder. The results can be viewed in the RStudio viewer pane, included in a R Markdown document or in a Shiny application. Also provides a Shiny application allowing to run this widget and to navigate in the files found by the search. Instead of creating a HTML widget, it is also possible to get the results of the search in a tibble'. The search is performed by the grep command-line utility.
This package creates dynamic grid layouts of images that can be included in Shiny applications and R markdown documents.
This package provides a set of tools for data wrangling, spatial data analysis, statistical modeling (including direct, model-assisted, photo-based, and small area tools), and USDA Forest Service data base tools. These tools are aimed to help Foresters, Analysts, and Scientists extract and perform analyses on USDA Forest Service data.
An interactive shiny'-based tool for exploration and quality assurance and quality control (QA/QC) of eddy covariance flux tower data processing. It generates data-point removal code via user-directed selection from a scatterplot, and can export a cleaned .csv with removed points set to NA plus an R script for reproducibility. Reference: Key (2025) <DOI:10.5281/zenodo.15597159>.
This package provides methods for computing and visualizing wildfire ignition exposure and directional vulnerability that are published in a series of scientific publications are automated by the functions in this package. See Beverly et al. (2010) <doi:10.1071/WF09071>, Beverly et al. (2021) <doi:10.1007/s10980-020-01173-8>, and Beverly and Forbes (2023) <doi:10.1007/s11069-023-05885-3> for background and methodology.
Fit data to an ellipse, hyperbola, or parabola. Bootstrapping is available when needed. The conic curve can be rotated through an arbitrary angle and the fit will still succeed. Helper functions are provided to convert generator coefficients from one style to another, generate test data sets, rotate conic section parameters, and so on. References include Nikolai Chernov (2014) "Fitting ellipses, circles, and lines by least squares" <https://people.cas.uab.edu/~mosya/cl/>; A. W. Fitzgibbon, M. Pilu, R. B. Fisher (1999) "Direct Least Squares Fitting of Ellipses" IEEE Trans. PAMI, Vol. 21, pages 476-48; N. Chernov, Q. Huang, and H. Ma (2014) "Fitting quadratic curves to data points", British Journal of Mathematics & Computer Science, 4, 33-60; N. Chernov and H. Ma (2011) "Least squares fitting of quadratic curves and surfaces", Computer Vision, Editor S. R. Yoshida, Nova Science Publishers, pp. 285-302.
Reads and writes ARFF files. ARFF (Attribute-Relation File Format) files are like CSV files, with a little bit of added meta information in a header and standardized NA values. They are quite often used for machine learning data sets and were introduced for the WEKA machine learning Java toolbox. See <https://waikato.github.io/weka-wiki/formats_and_processing/arff_stable/> for further info on ARFF and for <http://www.cs.waikato.ac.nz/ml/weka/> for more info on WEKA'. farff gets rid of the Java dependency that RWeka enforces, and it is at least a faster reader (for bigger files). It uses readr as parser back-end for the data section of the ARFF file. Consistency with RWeka is tested on Github and Travis CI with hundreds of ARFF files from OpenML'.
Automatically process Fluorescence Recovery After Photobleaching (FRAP) data and generate consistent, publishable figures. Note: this package does not replace ImageJ (or its equivalence) in raw image quantification. Some references about the methods: Sprague, Brian L. (2004) <doi:10.1529/biophysj.103.026765>; Day, Charles A. (2012) <doi:10.1002/0471142956.cy0219s62>.
An implementation of the methodologies described in Xi Liu, Afshin A. Divani, and Alexander Petersen (2022) <doi:10.1016/j.csda.2022.107421>, including truncated functional linear and truncated functional logistic regression models.
The purpose of this package is to tests whether a given moment of the distribution of a given sample is finite or not. For heavy-tailed distributions with tail exponent b, only moments of order smaller than b are finite. Tail exponent and heavy- tailedness are notoriously difficult to ascertain. But the finiteness of moments (including fractional moments) can be tested directly. This package does that following the test suggested by Trapani (2016) <doi:10.1016/j.jeconom.2015.08.006>.
Processes data from The Social Networks and Fertility Survey, downloaded from <https://dataarchive.lissdata.nl>, including correcting respondent errors and transforming network data into network objects to facilitate analyses and visualisation.
This package provides a friendly interface for modifying data frames with a sequence of piped commands built upon the tidyverse Wickham et al., (2019) <doi:10.21105/joss.01686> . The majority of commands wrap dplyr mutate statements in a convenient way to concisely solve common issues that arise when tidying small to medium data sets. Includes smart defaults and allows flexible selection of columns via tidyselect'.
This package provides a faster implementation of Bayesian Causal Forests (BCF; Hahn et al. (2020) <doi:10.1214/19-BA1195>), which uses regression tree ensembles to estimate the conditional average treatment effect of a binary treatment on a scalar output as a function of many covariates. This implementation avoids many redundant computations and memory allocations present in the original BCF implementation, allowing the model to be fit to larger datasets. The implementation was originally developed for the 2022 American Causal Inference Conference's Data Challenge. See Kokandakar et al. (2023) <doi:10.1353/obs.2023.0024> for more details.
Input has to be in the form of vectors of lower class limits and upper class limits and frequencies; the output will give a cumulative frequency distribution table with cumulative frequency plot.
Integrate Item Response Theory (IRT) and Federated Learning to estimate traditional IRT models, including the 2-Parameter Logistic (2PL) and the Graded Response Models, with enhanced privacy. It allows for the estimation in a distributed manner without compromising accuracy. A user-friendly shiny application is included.
Construction and smart selection of Gaussian process models for analysis of computer experiments with emphasis on treatment of functional inputs that are regularly sampled. This package offers: (i) flexible modeling of functional-input regression problems through the fairly general Gaussian process model; (ii) built-in dimension reduction for functional inputs; (iii) heuristic optimization of the structural parameters of the model (e.g., active inputs, kernel function, type of distance). An in-depth tutorial in the use of funGp is provided in Betancourt et al. (2024) <doi:10.18637/jss.v109.i05> and Metamodeling background is provided in Betancourt et al. (2020) <doi:10.1016/j.ress.2020.106870>. The algorithm for structural parameter optimization is described in <https://hal.science/hal-02532713>.