Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Construct and plot objective hierarchies and associated value and utility functions. Evaluate the values and utilities and visualize the results as colored objective hierarchies or tables. Visualize uncertainty by plotting median and quantile intervals within the nodes of objective hierarchies. Get numerical results of the evaluations in standard R data types for further processing.
This package provides a time series of the national grid demand (high-voltage electric power transmission network) in the UK since 2011.
This package provides methods for managing under- and over-enrollment in Simon's Two-Stage Design are offered by providing adaptive threshold adjustments and sample size recalibration. It also includes post-inference analysis tools to support clinical trial design and evaluation. The package is designed to enhance flexibility and accuracy in trial design, ensuring better outcomes in oncology and other clinical studies. Yunhe Liu, Haitao Pan (2024). Submitted.
Fits hierarchical models of animal abundance and occurrence to data collected using survey methods such as point counts, site occupancy sampling, distance sampling, removal sampling, and double observer sampling. Parameters governing the state and observation processes can be modeled as functions of covariates. References: Kellner et al. (2023) <doi:10.1111/2041-210X.14123>, Fiske and Chandler (2011) <doi:10.18637/jss.v043.i10>.
Assess essential unidimensionality using external validity information using the procedure proposed by Ferrando & Lorenzo-Seva (2019) <doi:10.1177/0013164418824755>. Provides two indices for assessing differential and incremental validity, both based on a second-order modelling schema for the general factor.
Downloads data from the UK Police public data API, the full docs of which are available at <https://data.police.uk/docs/>. Includes data on police forces and police force areas, crime reports, and the use of stop-and-search powers.
Fast flattening of hierarchical data structures (e.g. JSON, XML) into data.frames with a flexible spec language.
Offers tools for parsing and analyzing URL datasets, extracting key components and identifying common patterns. It aids in examining website architecture and identifying SEO issues, helping users optimize web presence and content strategy.
This package provides a tool to define the rare biosphere. ulrb solves the problem of the definition of rarity by replacing arbitrary thresholds with an unsupervised machine learning algorithm (partitioning around medoids, or k-medoids). This algorithm works for any type of microbiome data, provided there is an abundance table. This method also works for non-microbiome data.
Changes the column names of the inputted dataset to the correct names from the Uniform Crime Report codebook for the "Offenses Known and Clearance by Arrest" datasets from 1998-2014.
This package provides a generic reference Bayesian analysis of unidimensional mixture distributions obtained by a location-scale parameterisation of the model is implemented. The including functions simulate and summarize posterior samples for location-scale mixture models using a weakly informative prior. There is no need to define priors for scale-location parameters except two hyperparameters in which are associated with a Dirichlet prior for weights and a simplex.
Pseudo-random number generation of 17 univariate distributions proposed by Demirtas. (2005) <DOI:10.22237/jmasm/1114907220>.
We propose a new procedure, called model uncertainty variance, which can quantify the uncertainty of model selection on Autoregressive Moving Average models. The model uncertainty variance not pay attention to the accuracy of prediction, but focus on model selection uncertainty and providing more information of the model selection results. And to estimate the model measures, we propose an simplify and faster algorithm based on bootstrap method, which is proven to be effective and feasible by Monte-Carlo simulation. At the same time, we also made some optimizations and adjustments to the Model Confidence Bounds algorithm, so that it can be applied to the time series model selection method. The consistency of the algorithm result is also verified by Monte-Carlo simulation. We propose a new procedure, called model uncertainty variance, which can quantify the uncertainty of model selection on Autoregressive Moving Average models. The model uncertainty variance focuses on model selection uncertainty and providing more information of the model selection results. To estimate the model uncertainty variance, we propose an simplified and faster algorithm based on bootstrap method, which is proven to be effective and feasible by Monte-Carlo simulation. At the same time, we also made some optimizations and adjustments to the Model Confidence Bounds algorithm, so that it can be applied to the time series model selection method. The consistency of the algorithm result is also verified by Monte-Carlo simulation. Please see Li,Y., Luo,Y., Ferrari,D., Hu,X. and Qin,Y. (2019) Model Confidence Bounds for Variable Selection. Biometrics, 75:392-403.<DOI:10.1111/biom.13024> for more information.
Verb-like functions to work with messy data, often derived from spreadsheets or parsed PDF tables. Includes functions for unwrapping values broken up across rows, relocating embedded grouping values, and to annotate meaningful formatting in spreadsheet files.
If a procedure consists of several stages and there are several models that can be selected for each stage, uncertainty of the procedure can be decomposed by stages or models. This package includes the ANOVA-based method, the cumulative uncertainty-based method, and the balanced decomposition method. Yongdai Kim et al. (2019) <doi:10.1016/j.hydroa.2019.100024> is a related paper which is accessible via the URL below.
Enables the user to calculate Value at Risk (VaR) and Expected Shortfall (ES) by means of various parametric and semiparametric GARCH-type models. For the latter the estimation of the nonparametric scale function is carried out by means of a data-driven smoothing approach. Model quality, in terms of forecasting VaR and ES, can be assessed by means of various backtesting methods such as the traffic light test for VaR and a newly developed traffic light test for ES. The approaches implemented in this package are described in e.g. Feng Y., Beran J., Letmathe S. and Ghosh S. (2020) <https://ideas.repec.org/p/pdn/ciepap/137.html> as well as Letmathe S., Feng Y. and Uhde A. (2021) <https://ideas.repec.org/p/pdn/ciepap/141.html>.
For each string in a set of strings, determine a unique tag that is a substring of fixed size k unique to that string, if it has one. If no such unique substring exists, the least frequent substring is used. If multiple unique substrings exist, the lexicographically smallest substring is used. This lexicographically smallest substring of size k is called the "UniqTag" of that string.
Top-down and bottom-up algorithms for nonparametric function estimation in Gaussian noise using Unbalanced Haar wavelets.
UpSet.js is a re-implementation of UpSetR to create interactive set visualizations for more than three sets. This is a htmlwidget wrapper around the JavaScript library UpSet.js'.
This package provides a variational mapping approach that reveals and expands future temporal dynamics from folded high-dimensional geometric distance spaces, unfold turns a set of time series into a 4D block of pairwise distances between reframed windows, learns a variational mapper that maps those distances to the next reframed window, and produces horizon-wise predictive functions for each input series. In short: it unfolds the future path of each series from a folded geometric distance representation.
Two Phase I designs are implemented in the package: the classical 3+3 and the Continual Reassessment Method (<doi:10.2307/2531628>). Simulations tools are also available to estimate the operating characteristics of the methods with several user-dependent options.
This package provides a tool for checking how much information is disclosed when reporting summary statistics.
Demographic data on the United States at the county and state levels spanning multiple years.
The udder quarter infection data set contains infection times of individual cow udder quarters with Corynebacterium bovis (Laevens et al. 1997 <DOI:10.3168/jds.S0022-0302(97)76295-7>). Obviously, the four udder quarters are clustered within a cow, and udder quarters are sampled only approximately monthly, generating interval-censored data. The data set contains both covariates that change within a cow (e.g., front and rear udder quarters) and covariates that change between cows (e.g., parity [the number of previous calvings]). The correlation between udder infection times within a cow also is of interest, because this is a measure of the infectivity of the agent causing the disease. Various models have been applied to address the problem of interdependence for right-censored event times. These models, as applied to this data set, can be found back in the publications found in the reference list.