Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Create, Plot and Compare Replication Timing Profiles. The method is described in Muller et al., (2014) <doi: 10.1093/nar/gkt878>.
Defines functions that can be used to collect provenance as an R script executes or during a console session. The output is a text file in PROV-JSON format.
This package provides a collection of functions to simulate dice rolls and the like. In particular, experiments and exercises can be performed looking at combinations and permutations of values in dice rolls and coin flips, together with the corresponding frequencies of occurrences. When applying each function, the user has to input the number of times (rolls, flips) to toss the dice. Needless to say, the more the tosses, the more the frequencies approximate the actual probabilities. Moreover, the package provides functions to generate non-transitive sets of dice (like Efron's) and to check whether a given set of dice is non-transitive with given probability.
Collection of functions designed to compute risk-based portfolios as described in Ardia et al. (2017) <doi:10.1007/s10479-017-2474-7> and Ardia et al. (2017) <doi:10.21105/joss.00171>.
This package provides a collection of functions for numerical construction of optimal discriminating designs. At the current moment T-optimal designs (which maximize the lower bound for the power of F-test for regression model discrimination), KL-optimal designs (for lognormal errors) and their robust analogues can be calculated with the package.
It helps you to read (.dim) images with CRS directly into R programming. One can import both Sentinel 1 and 2 images or any processed data with this software.
Generate a table of cumulative water influx into hydrocarbon reservoirs over time using un-steady and pseudo-steady state models. Van Everdingen, A. F. and Hurst, W. (1949) <doi:10.2118/949305-G>. Fetkovich, M. J. (1971) <doi:10.2118/2603-PA>. Yildiz, T. and Khosravi, A. (2007) <doi:10.2118/103283-PA>.
Predicting regulatory DNA elements based on epigenomic signatures. This package is more of a set of building blocks than a direct solution. REPTILE regulatory prediction pipeline is built on this R package. See <https://github.com/yupenghe/REPTILE> for more information.
Utilities for sparse signal recovery suitable for compressed sensing. L1, L2 and TV penalties, DFT basis matrix, simple sparse signal generator, mutual cumulative coherence between two matrices and examples, Lp complex norm, scaling back regression coefficients.
This package provides functions for studying realized genetic relatedness between people. Users will be able to simulate inheritance patterns given pedigree structures, generate SNP marker data given inheritance patterns, and estimate realized relatedness between pairs of individuals using SNP marker data. See Wang (2017) <doi:10.1534/genetics.116.197004>. This work was supported by National Institutes of Health grants R37 GM-046255.
Fits a multivariate value-added model (VAM), see Broatch, Green, and Karl (2018) <doi:10.32614/RJ-2018-033> and Broatch and Lohr (2012) <doi:10.3102/1076998610396900>, with normally distributed test scores and a binary outcome indicator. A pseudo-likelihood approach, Wolfinger (1993) <doi:10.1080/00949659308811554>, is used for the estimation of this joint generalized linear mixed model. The inner loop of the pseudo-likelihood routine (estimation of a linear mixed model) occurs in the framework of the EM algorithm presented by Karl, Yang, and Lohr (2013) <DOI:10.1016/j.csda.2012.10.004>. This material is based upon work supported by the National Science Foundation under grants DRL-1336027 and DRL-1336265.
Efficient framework for ridge redundancy analysis (rrda), tailored for high-dimensional omics datasets where the number of predictors exceeds the number of samples. The method leverages Singular Value Decomposition (SVD) to avoid direct inversion of the covariance matrix, enhancing scalability and performance. It also introduces a memory-efficient storage strategy for coefficient matrices, enabling practical use in large-scale applications. The package supports cross-validation for selecting regularization parameters and reduced-rank dimensions, making it a robust and flexible tool for multivariate analysis in omics research. Please refer to our article (Yoshioka et al., 2025) for more details.
This package implements methods described by the paper Robins and Tsiatis (1991) <DOI:10.1080/03610929108830654>. These use g-estimation to estimate the causal effect of a treatment in a two-armed randomised control trial where non-compliance exists and is measured, under an assumption of an accelerated failure time model and no unmeasured confounders.
This package provides a generic implementation of the RStudio connection contract to make it easier for database connections, and other type of connections, opened via R packages integrate with the connections pane inside the RStudio interactive development environment (IDE).
Doubly ranked tests are nonparametric tests for grouped functional and multivariate data. The testing procedure first ranks a matrix (or three dimensional array) of data by column (if a matrix) or by cell (across the third dimension if an array). By default, it calculates a sufficient statistic for the subject's order within the sample using the observed ranks, taken over the columns or cells. Depending on the number of groups, G, the summarized ranks are then analyzed using either a Wilcoxon Rank Sum test (G = 2) or a Kruskal-Wallis (G greater than 2).
This package contains three functions that access environmental data from any ERDDAPâ ¢ data web service. The rxtracto() function extracts data along a trajectory for a given "radius" around the point. The rxtracto_3D() function extracts data in a box. The rxtractogon() function extracts data in a polygon. All of those three function use the rerddap package to extract the data, and should work with any ERDDAPâ ¢ server. There are also two functions, plotBBox() and plotTrack() that use the plotdap package to simplify the creation of maps of the data.
This package provides access to the Ravelry API <https://www.ravelry.com/groups/ravelry-api>. An R wrapper for pulling data from Ravelry.com', an organizational tool for crocheters, knitters, spinners, and weavers. You can retrieve pattern, yarn, author, and shop information by search or by a given id.
Pattern matching, extraction, replacement and other string processing operations using Google's RE2 <https://github.com/google/re2> regular-expression engine. Consistent interface (similar to stringr'). RE2 uses finite-automata based techniques, and offers a fast and safe alternative to backtracking regular-expression engines like those used in stringr', stringi and other PCRE implementations.
Load data from vk.com api about your communiti users and views, ads performance, post on user wall and etc. For more information see API Documentation <https://vk.com/dev/first_guide>.
Modeling and plotting functions for Reliability Growth Analysis (RGA). Models include the Duane (1962) <doi:10.1109/TA.1964.4319640>, Non-Homogeneous Poisson Process (NHPP) by Crow (1975) <https://apps.dtic.mil/sti/citations/ADA020296>, Piecewise Weibull NHPP by Guo et al. (2010) <doi:10.1109/RAMS.2010.5448029>, and Piecewise Weibull NHPP with Change Point Detection based on the segmented package by Muggeo (2024) <https://cran.r-project.org/package=segmented>.
Build regular expressions piece by piece using human readable code. This package contains number-related functionality, and is primarily intended to be used by package developers.
Facilitates the design and generation of optimal color (or symbol) codes that can be used to mark and identify individual animals. These codes are made such that the IDs are robust to partial erasure: even if sections of the code are lost, the entire identity of the animal can be reconstructed. Thus, animal subjects are not confused and no ambiguity is introduced.
Various functions to fit models for non-normal repeated measurements, such as Binary Random Effects Models with Two Levels of Nesting, Bivariate Beta-binomial Regression Models, Marginal Bivariate Binomial Regression Models, Cormack capture-recapture models, Continuous-time Hidden Markov Chain Models, Discrete-time Hidden Markov Chain Models, Changepoint Location Models using a Continuous-time Two-state Hidden Markov Chain, generalized nonlinear autoregression models, multivariate Gaussian copula models, generalized non-linear mixed models with one random effect, generalized non-linear mixed models using h-likelihood for one random effect, Repeated Measurements Models for Counts with Frailty or Serial Dependence, Repeated Measurements Models for Continuous Variables with Frailty or Serial Dependence, Ordinal Random Effects Models with Dropouts, marginal homogeneity models for square contingency tables, correlated negative binomial models with Kalman update. References include Lindsey's text books, JK Lindsey (2001) <isbn:10-0198508123> and JK Lindsey (1999) <isbn:10-0198505590>.
This package provides tools to read, write, visualize Protein Data Bank (PDB) files and perform some structural manipulations.