Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
This package provides useful tools which supplement the use of Simulx software and R connectors ('Monolix Suite'). Simulx is an easy, efficient and flexible application for clinical trial simulations. You need Simulx software to be installed in order to use RsSimulx package. Among others tasks, RsSimulx provides the same functions as package mlxR does with a compatibility with Simulx software.
Imports real-time thermo cycler (qPCR) data from Real-time PCR Data Markup Language (RDML) and transforms to the appropriate formats of the qpcR and chipPCR packages, as described in Rodiger et al. (2017) <doi:10.1093/bioinformatics/btx528>. Contains a dendrogram visualization for the structure of RDML object and GUI for RDML editing.
Partitions the phenotypic variance of a plastic trait, studied through its reaction norm. The variance partition distinguishes between the variance arising from the average shape of the reaction norms (V_Plas) and the (additive) genetic variance . The latter is itself separated into an environment-blind component (V_G/V_A) and the component arising from plasticity (V_GxE/V_AxE). The package also provides a way to further partition V_Plas into aspects (slope/curvature) of the shape of the average reaction norm (pi-decomposition) and partition V_Add (gamma-decomposition) and V_AxE (iota-decomposition) into the impact of genetic variation in the reaction norm parameters. Reference: de Villemereuil & Chevin (2025) <doi:10.32942/X2NC8B>.
An implementation of the Heroicons icon library for shiny applications and other R web-based projects. You can search, render, and customize icons without CSS or JavaScript dependencies.
Personalized assignment to one of many treatment arms via regularized and clustered joint assignment forests as described in Ladhania, Spiess, Ungar, and Wu (2023) <doi:10.48550/arXiv.2311.00577>. The algorithm pools information across treatment arms: it considers a regularized forest-based assignment algorithm based on greedy recursive partitioning that shrinks effect estimates across arms; and it incorporates a clustering scheme that combines treatment arms with consistently similar outcomes.
Really Poor Man's Graphical User Interface, used to create interactive R analysis sessions with simple R commands.
Apply sensitivity analysis for offline policy evaluation, as implemented in Jung et al. (2017) <arXiv:1702.04690> based on Rosenbaum and Rubin (1983) <http://www.jstor.org/stable/2345524>.
Defines colour palettes and themes for Royal Statistical Society (RSS) publications, including Significance magazine. Palettes and themes are supported in both base R and ggplot2 graphics, and are intended to be used by authors submitting to RSS publications.
This package provides a comprehensive collection of practical and easy-to-use tools for regression analysis of recurrent events, with or without the presence of a (possibly) informative terminal event described in Chiou et al. (2023) <doi:10.18637/jss.v105.i05>. The modeling framework is based on a joint frailty scale-change model, that includes models described in Wang et al. (2001) <doi:10.1198/016214501753209031>, Huang and Wang (2004) <doi:10.1198/016214504000001033>, Xu et al. (2017) <doi:10.1080/01621459.2016.1173557>, and Xu et al. (2019) <doi:10.5705/SS.202018.0224> as special cases. The implemented estimating procedure does not require any parametric assumption on the frailty distribution. The package also allows the users to specify different model forms for both the recurrent event process and the terminal event.
Estimation, forecasting, simulation, and portfolio construction for regime-switching models with exogenous variables as in Pelletier (2006) <doi:10.1016/j.jeconom.2005.01.013>.
Download and open manifest files provided by the Copernicus Global Land Service data <https://land.copernicus.eu/global/>. The manifest files are available at: <https://land.copernicus.vgt.vito.be/manifest/>. Also see: <https://land.copernicus.eu/global/access/>. Before you can download the data, you will first need to register to create a username and password.
This package provides a comprehensive suite of utilities for univariate continuous probability distributions and reliability models. Includes functions to compute the probability density, cumulative distribution, quantile, reliability, and hazard functions, along with random variate generation. Also offers diagnostic and model assessment tools such as Quantile-Quantile (Q-Q) and Probability-Probability (P-P) plots, the Kolmogorov-Smirnov goodness-of-fit test, and model selection criteria including the Akaike Information Criterion (AIC) and Bayesian Information Criterion (BIC). Currently implements the following distributions: Burr X, Chen, Exponential Extension, Exponentiated Logistic, Exponentiated Weibull, Exponential Power, Flexible Weibull, Generalized Exponential, Gompertz, Generalized Power Weibull, Gumbel, Inverse Generalized Exponential, Linear Failure Rate, Log-Gamma, Logistic-Exponential, Logistic-Rayleigh, Log-log, Marshall-Olkin Extended Exponential, Marshall-Olkin Extended Weibull, and Weibull Extension distributions. Serves as a valuable resource for teaching and research in probability theory, reliability analysis, and applied statistical modeling.
Empirical best linear unbiased prediction (EBLUP) and robust prediction of the area-level means under the basic unit-level model. The model can be fitted by maximum likelihood or a (robust) M-estimator. Mean square prediction error is computed by a parametric bootstrap.
Point and interval estimation of linear parameters with data obtained from complex surveys (including stratified and clustered samples) when randomization techniques are used. The randomized response technique was developed to obtain estimates that are more valid when studying sensitive topics. Estimators and variances for 14 randomized response methods for qualitative variables and 7 randomized response methods for quantitative variables are also implemented. In addition, some data sets from surveys with these randomization methods are included in the package.
Toolbox for chemometrics analysis of bidimensional gas chromatography data. This package import data for common scientific data format (NetCDF) and fold it to 2D chromatogram. Then, it can perform preprocessing and multivariate analysis. In the preprocessing algorithms, baseline correction, smoothing, and peak alignment are available. While in multivariate analysis, multiway principal component analysis is incorporated.
Suppose we have a data matrix, which is the superposition of a low-rank component and a sparse component. Candes, E. J., Li, X., Ma, Y., & Wright, J. (2011). Robust principal component analysis?. Journal of the ACM (JACM), 58(3), 11. prove that we can recover each component individually under some suitable assumptions. It is possible to recover both the low-rank and the sparse components exactly by solving a very convenient convex program called Principal Component Pursuit; among all feasible decompositions, simply minimize a weighted combination of the nuclear norm and of the L1 norm. This package implements this decomposition algorithm resulting with Robust PCA approach.
Enables researchers to sample redistricting plans from a pre-specified target distribution using Sequential Monte Carlo and Markov Chain Monte Carlo algorithms. The package allows for the implementation of various constraints in the redistricting process such as geographic compactness and population parity requirements. Tools for analysis such as computation of various summary statistics and plotting functionality are also included. The package implements the SMC algorithm of McCartan and Imai (2023) <doi:10.1214/23-AOAS1763>, the enumeration algorithm of Fifield, Imai, Kawahara, and Kenny (2020) <doi:10.1080/2330443X.2020.1791773>, the Flip MCMC algorithm of Fifield, Higgins, Imai and Tarr (2020) <doi:10.1080/10618600.2020.1739532>, the Merge-split/Recombination algorithms of Carter et al. (2019) <doi:10.48550/arXiv.1911.01503> and DeFord et al. (2021) <doi:10.1162/99608f92.eb30390f>, and the Short-burst optimization algorithm of Cannon et al. (2020) <doi:10.48550/arXiv.2011.02288>.
This package contains various tools to estimate, analyze, and visualize Response Item Networks. ResIN dummy-codes ordered and qualitative response choices from (survey) data, calculates pairwise associations and maps the location of each item response as a node in a force-directed network. Please refer to <https://www.resinmethod.net/> for more details.
This package provides a toolbox created by members of the International Union for Conservation of Nature (IUCN) Red List of Ecosystems Committee for Scientific Standards. Primarily, it is a set of tools suitable for calculating the metrics required for making assessments of species and ecosystems against the IUCN Red List of Threatened Species and the IUCN Red List of Ecosystems categories and criteria. See the IUCN website for detailed guidelines, the criteria, publications and other information.
The Stuttgart Neural Network Simulator (SNNS) is a library containing many standard implementations of neural networks. This package wraps the SNNS functionality to make it available from within R. Using the RSNNS low-level interface, all of the algorithmic functionality and flexibility of SNNS can be accessed. Furthermore, the package contains a convenient high-level interface, so that the most common neural network topologies and learning algorithms integrate seamlessly into R.
Bayes estimation of probit choice models in cross-sectional and panel settings. The package can analyze binary, multivariate, ordered, and ranked choices, as well as heterogeneity of choice behavior among deciders. The main functionality includes model fitting via Gibbs sampling, tools for convergence diagnostic, choice data simulation, in-sample and out-of-sample choice prediction, and model selection using information criteria and Bayes factors. The latent class model extension facilitates preference-based decider classification, where the number of latent classes can be inferred via the Dirichlet process or a weight-based updating heuristic. This allows for flexible modeling of choice behavior without the need to impose structural constraints. For a reference on the method, see Oelschlaeger and Bauer (2021) <https://trid.trb.org/view/1759753>.
The cnpy library written by Carl Rogers provides read and write facilities for files created with (or for) the NumPy extension for Python'. Vectors and matrices of numeric types can be read or written to and from files as well as compressed files. Support for integer files is available if the package has been built with as C++11 which should be the default on all platforms since the release of R 3.3.0.
HTML formats and templates for rmarkdown documents, with some extra features such as automatic table of contents, lightboxed figures, dynamic crosstab helper.
This package provides functions and methods for manipulating SNOMED CT concepts. The package contains functions for loading the SNOMED CT release into a convenient R environment, selecting SNOMED CT concepts using regular expressions, and navigating the SNOMED CT ontology. It provides the SNOMEDconcept S3 class for a vector of SNOMED CT concepts (stored as 64-bit integers) and the SNOMEDcodelist S3 class for a table of concepts IDs with descriptions. The package can be used to construct sets of SNOMED CT concepts for research (<doi:10.1093/jamia/ocac158>). For more information about SNOMED CT visit <https://www.snomed.org/>.