Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Interface for loading data from ActiveCampaign API v3 <https://developers.activecampaign.com/reference>. Provide functions for getting data by deals, contacts, accounts, campaigns and messages.
Provide function for get data from YouTube Data API <https://developers.google.com/youtube/v3/docs/>, YouTube Analytics API <https://developers.google.com/youtube/analytics/reference/> and YouTube Reporting API <https://developers.google.com/youtube/reporting/v1/reports>.
Distance-sampling (<doi:10.1007/978-3-319-19219-2>) is a field survey and analytical method that estimates density and abundance of survey targets (e.g., animals) when detection probability declines with observation distance. Distance-sampling is popular in ecology, especially when survey targets are observed from aerial platforms (e.g., airplane or drone), surface vessels (e.g., boat or truck), or along walking transects. Analysis involves fitting smooth (parametric) curves to histograms of observation distances and using those functions to adjust density estimates for missed targets. Routines included here fit curves to observation distance histograms, estimate effective sampling area, density of targets in surveyed areas, and the abundance of targets in a surrounding study area. Confidence interval estimation uses built-in bootstrap resampling. Help files are extensive and have been vetted by multiple authors. Many tutorials are available on the package's website (URL below).
This package provides an interface to the Vamp audio analysis plugin system <https://www.vamp-plugins.org/> developed by Queen Mary University of London's Centre for Digital Music. Enables loading and running Vamp plugins for various audio analysis tasks including tempo detection, onset detection, spectral analysis, and audio feature extraction. Supports mono and stereo audio with automatic channel adaptation and domain conversion.
Constrained clustering, transfer functions, and other methods for analysing Quaternary science data.
This is a wrapper function for image(), which makes reasonable raster plots with nice axis and other useful features.
Loads Blackrock <https://blackrockneurotech.com> neural signal data files into the memory, provides utility tools to extract the data into common formats such as plain-text tsv and HDF5'.
This package provides tools for qPCR data analysis using Delta Ct and Delta Delta Ct methods, including t-test, Wilcoxon-test, ANOVA models, and publication-ready visualizations. The package supports multiple target, and multiple reference genes, and uses a calculation framework adopted from Ganger et al. (2017) <doi:10.1186/s12859-017-1949-5> and Taylor et al. (2019) <doi:10.1016/j.tibtech.2018.12.002>, covering both the Livak and Pfaffl methods.
This package provides a single method implementing multiple approaches to generate pseudo-random vectors whose components sum up to one (see, e.g., Maziero (2015) <doi:10.1007/s13538-015-0337-8>). The components of such vectors can for example be used for weighting objectives when reducing multi-objective optimisation problems to a single-objective problem in the socalled weighted sum scalarisation approach.
R implementation of SIDES-based subgroup search algorithms (Lipkovich et al. (2017) <doi:10.1002/sim.7064>).
Rcpp bindings to the native C++ implementation of MS Numpress, that provides two compression schemes for numeric data from mass spectrometers. The library provides implementations of 3 different algorithms, 1 designed to compress first order smooth data like retention time or M/Z arrays, and 2 for compressing non smooth data with lower requirements on precision like ion count arrays. Refer to the publication (Teleman et al., (2014) <doi:10.1074/mcp.O114.037879>) for more details.
Enhances the R Optimization Infrastructure ('ROI') package by registering the quadprog solver. It allows for solving quadratic programming (QP) problems.
Diagnostics and data preparation for random effects within estimator, random effects within-idiosyncratic estimator, between-within-idiosyncratic model, and cross-classified between model. Mundlak, Yair (1978) <doi:10.2307/1913646>. Hausman, Jeffrey (1978) <doi:10.2307/1913827>. Allison, Paul (2009) <doi:10.4135/9781412993869>. Neuhaus, J.M., and J. D. Kalbfleisch (1998) <doi:10.2307/3109770>.
This package provides a GUI front-end for ggplot2 supports Kaplan-Meier plot, histogram, Q-Q plot, box plot, errorbar plot, scatter plot, line chart, pie chart, bar chart, contour plot, and distribution plot.
This package implements the pseudo-R2D2 prior for ordinal regression from the paper "Pseudo-R2D2 prior for high-dimensional ordinal regression" by Yanchenko (2025) <doi:10.1007/s11222-025-10667-x>. In particular, it provides code to evaluate the probability distribution function for the cut-points, compute the log-likelihood, calculate the hyper-parameters for the global variance parameter, find the distribution of McFadden's coefficient-of-determination, and fit the model in rstan'. Please cite the paper if you use these codes.
Read and write labelled sparse matrices in text format as used by software such as SVMLight', LibSVM', ThunderSVM', LibFM', xLearn', XGBoost', LightGBM', and others. Supports labelled data for regression, classification (binary, multi-class, multi-label), and ranking (with qid field), and can handle header metadata and comments in files.
Generates a project and repo for easy initialization of a GitHub repo for R workshops. The repo includes a README with instructions to ensure that all users have the needed packages, an RStudio project with the right directories and the proper data. The repo can then be used for hosting code taught during the workshop.
Since the early 1970s eyewitness testimony researchers have recognised the importance of estimating properties such as lineup bias (is the lineup biased against the suspect, leading to a rate of choosing higher than one would expect by chance?), and lineup size (how many reasonable choices are in fact available to the witness? A lineup is supposed to consist of a suspect and a number of additional members, or foils, whom a poor-quality witness might mistake for the perpetrator). Lineup measures are descriptive, in the first instance, but since the earliest articles in the literature researchers have recognised the importance of reasoning inferentially about them. This package contains functions to compute various properties of laboratory or police lineups, and is intended for use by researchers in forensic psychology and/or eyewitness testimony research. Among others, the r4lineups package includes functions for calculating lineup proportion, functional size, various estimates of effective size, diagnosticity ratio, homogeneity of the diagnosticity ratio, ROC curves for confidence x accuracy data and the degree of similarity of faces in a lineup.
Defines functions that can be used to collect provenance as an R script executes or during a console session. The output is a text file in PROV-JSON format.
Makes documents containing plots and tables from a table of R codes. Can make "HTML", "pdf('LaTex')", "docx('MS Word')" and "pptx('MS Powerpoint')" documents with or without R code. In the package, modularized shiny app codes are provided. These modules are intended for reuse across applications.
This package provides a comprehensive suite of utilities for univariate continuous probability distributions and reliability models. Includes functions to compute the probability density, cumulative distribution, quantile, reliability, and hazard functions, along with random variate generation. Also offers diagnostic and model assessment tools such as Quantile-Quantile (Q-Q) and Probability-Probability (P-P) plots, the Kolmogorov-Smirnov goodness-of-fit test, and model selection criteria including the Akaike Information Criterion (AIC) and Bayesian Information Criterion (BIC). Currently implements the following distributions: Burr X, Chen, Exponential Extension, Exponentiated Logistic, Exponentiated Weibull, Exponential Power, Flexible Weibull, Generalized Exponential, Gompertz, Generalized Power Weibull, Gumbel, Inverse Generalized Exponential, Linear Failure Rate, Log-Gamma, Logistic-Exponential, Logistic-Rayleigh, Log-log, Marshall-Olkin Extended Exponential, Marshall-Olkin Extended Weibull, and Weibull Extension distributions. Serves as a valuable resource for teaching and research in probability theory, reliability analysis, and applied statistical modeling.
Estimates of standard errors of popular risk and performance measures for asset or portfolio returns using methods as described in Chen and Martin (2021) <doi:10.21314/JOR.2020.446>.
Simple, native RethinkDB client.
This package provides a collection of implementations of semi-supervised classifiers and methods to evaluate their performance. The package includes implementations of, among others, Implicitly Constrained Learning, Moment Constrained Learning, the Transductive SVM, Manifold regularization, Maximum Contrastive Pessimistic Likelihood estimation, S4VM and WellSVM.