Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel search send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
It enables the identification of sequentialexperimentation orders for factorial designs that jointly reduce bias and the number of level changes. The method used is that presented by Conto et al. (2025), known as the Assignment-Expansion method, which consists of adapting the linear programming assignment problem to generate balanced experimentation orders. The properties identified are then generalized to designs with a larger number of factors and levels using the expansion method proposed by Correa et al. (2009) and later generalized by Bhowmik et al. (2017). For more details see Conto et al. (2025) <doi:10.1016/j.cie.2024.110844>, Correa et al. (2009) <doi:10.1080/02664760802499337> and Bhowmik et al. (2017) <doi:10.1080/03610926.2016.1152490>.
This package contains basic tools for visualizing, interpreting, and building regression models. It has been designed for use with the book Introduction to Regression and Modeling with R by Adam Petrie, Cognella Publishers, ISBN: 978-1-63189-250-9.
Defines colour palettes and themes for Royal Statistical Society (RSS) publications, including Significance magazine. Palettes and themes are supported in both base R and ggplot2 graphics, and are intended to be used by authors submitting to RSS publications.
Adaptation of the Matlab tsEVA toolbox developed by Lorenzo Mentaschi available here: <https://github.com/menta78/tsEva>. It contains an implementation of the Transformed-Stationary (TS) methodology for non-stationary extreme value Analysis (EVA) as described in Mentaschi et al. (2016) <doi:10.5194/hess-20-3527-2016>. In synthesis this approach consists in: (i) transforming a non-stationary time series into a stationary one to which the stationary extreme value theory can be applied; and (ii) reverse-transforming the result into a non-stationary extreme value distribution. RtsEva offers several options for trend estimation (mean, extremes, seasonal) and contains multiple plotting functions displaying different aspects of the non-stationarity of extremes.
Robust pairwise correlations based on estimates of scale, particularly on "FastQn" one-step M-estimate.
We provide a number of algorithms to estimate fundamental statistics including Fréchet mean and geometric median for manifold-valued data. Also, C++ header files are contained that implement elementary operations on manifolds such as Sphere, Grassmann, and others. See Bhattacharya and Bhattacharya (2012) <doi:10.1017/CBO9781139094764> if you are interested in statistics on manifolds, and Absil et al (2007, ISBN:9780691132983) on computational aspects of optimization on matrix manifolds.
Routines to interact with the Numerai Machine Learning Tournament API <https://numer.ai>. The functionality includes the ability to automatically download the current tournament data, submit predictions, and to get information for your user.
Random univariate and multivariate finite mixture model generation, estimation, clustering, latent class analysis and classification. Variables can be continuous, discrete, independent or dependent and may follow normal, lognormal, Weibull, gamma, Gumbel, binomial, Poisson, Dirac, uniform or circular von Mises parametric families.
An implementation of a stochastic heuristic method for performing multidimensional function optimization. The method is inspired in the Cross-Entropy Method. It does not relies on derivatives, neither imposes particularly strong requirements into the function to be optimized. Additionally, it takes profit from multi-core processing to enable optimization of time-consuming functions.
Get your data (forms, structures, answers) from Coletum <https://coletum.com> to handle and analyse.
This package provides functions to correct biased transition and fertility estimates in population projection matrices caused by small sample sizes. Small or short-term studies frequently produce structural zeros (biologically possible transitions never observed) and structural ones (transitions estimated at 100% survival, stasis, or mortality that are biologically implausible). Both distort matrix structure and bias estimates of population growth. Implements a multinomial-Dirichlet Bayesian prior for transition probabilities and a Gamma-Poisson prior for reproduction, allowing analysts to incorporate prior biological knowledge and regularise estimates from rare or unobserved events. Includes functions to compute marginal posterior credible intervals for all transition probabilities (transition_CrI()), visualise those intervals as point-range plots (plot_transition_CrI()), and display the full posterior beta density for each matrix entry (plot_transition_density()). Methods are described in Tremblay et al. (2021) <doi:10.1016/j.ecolmodel.2021.109526>.
Convert README.md to vignettes when installing packages without vignettes.
This package provides a native R implementation for encoding and decoding sixel graphics (<https://vt100.net/docs/vt3xx-gp/chapter14.html>), and a dedicated sixel graphics device that allows plots to be rendered directly within compatible terminal emulators.
An integrated set of tools to allow data users to conduct meteorological normalisation and counterfactual modelling for air quality data. The meteorological normalisation technique uses predictive random forest models to remove variation of pollutant concentrations so trends and interventions can be explored in a robust way. For examples, see Grange et al. (2018) <doi:10.5194/acp-18-6223-2018> and Grange and Carslaw (2019) <doi:10.1016/j.scitotenv.2018.10.344>. The random forest models can also be used for counterfactual or business as usual (BAU) modelling by using the models to predict, from the model's perspective, the future. For an example, see Grange et al. (2021) <doi:10.5194/acp-2020-1171>.
This package provides a set of tools to reconstruct ordered ontogenic trajectories from single cell RNAseq data.
This package provides tools for diagnosing the reproducibility of statistical model outputs under data perturbations. Implements bootstrap, subsampling, and noise-based perturbation schemes and computes coefficient stability, p-value stability, selection stability, prediction stability, and a composite reproducibility index on a 0 to 100 scale. Includes cross-validation ranking stability for model comparison and visualization utilities. Optional backends support robust M-estimation ('MASS') and penalized regression ('glmnet'). Bootstrap perturbation follows Efron and Tibshirani (1993, ISBN:9780412042317); selection stability follows Meinshausen and Buhlmann (2010) <doi:10.1111/j.1467-9868.2010.00740.x>; reproducibility framework follows Peng (2011) <doi:10.1126/science.1213847>.
Queries data from RDAP servers.
R^2 measure of explained variation under the semiparametric additive hazards model is estimated. The measure can be used as a measure of predictive capability and therefore it can be adopted in model selection process. Rava, D. and Xu, R. (2020) <arXiv:2003.09460>.
Generates tile maps for the East Caucasian language family, inspired by the Typological Atlas of the Languages of Daghestan (TALD, <https://lingconlab.ru/tald/>). It leverages ggplot2 to create visually informative maps, displaying rectangles for each language and allowing for color-coding based on linguistic features. The package includes a built-in dataset of 56 languages and the template for their distribution and provides flexibility to customize the tile map's appearance. The default template can be modified via the ability to hide or rename languages. It's designed to be used with external data tables containing language information and features, offering a tool for visualizing the geographic distribution and linguistic characteristics of East Caucasian languages.
Enhances the R Optimization Infrastructure ('ROI') package with the NLopt solver for solving nonlinear optimization problems.
Estimates of standard errors of popular risk and performance measures for asset or portfolio returns using methods as described in Chen and Martin (2021) <doi:10.21314/JOR.2020.446>.
Adds the MIxing-Data Sampling (MIDAS, Ghysels et al. (2007) <doi:10.1080/07474930600972467>) components to a variety of GARCH and MEM (Engle (2002) <doi:10.1002/jae.683>, Engle and Gallo (2006) <doi:10.1016/j.jeconom.2005.01.018>, and Amendola et al. (2024) <doi:10.1016/j.seps.2023.101764>) models, with the aim of predicting the volatility with additional low-frequency (that is, MIDAS) terms. The estimation takes place through simple functions, which provide in-sample and (if present) and out-of-sample evaluations. rumidas also offers a summary tool, which synthesizes the main information of the estimated model. There is also the possibility of generating one-step-ahead and multi-step-ahead forecasts.
Offers bathymetric interpolation using Inverse Distance Weighted and Ordinary Kriging via the gstat and terra packages. Other functions focus on quantifying physical aquatic habitats (e.g., littoral, epliminion, metalimnion, hypolimnion) from interpolated digital elevation models (DEMs). Functions were designed to calculate these metrics across water levels for use in reservoirs but can be applied to any DEM and will provide values for fixed conditions. Parameters like Secchi disk depth or estimated photic zone, thermocline depth, and water level fluctuation depth are included in most functions.
Extension to REddyProc that allows reading data from netCDF files.