Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Package to select best model among several linear and nonlinear models. The main function uses the gnls() function from the nlme package to fit the data to nine regression models, named: "linear", "quadratic", "cubic", "logistic", "exponential", "power", "monod", "haldane", "logit".
This shows how NONMEM(R) software works. NONMEM's classical estimation methods like First Order(FO) approximation', First Order Conditional Estimation(FOCE)', and Laplacian approximation are explained.
In many binary classification applications, such as disease diagnosis and spam detection, practitioners commonly face the need to limit type I error (i.e., the conditional probability of misclassifying a class 0 observation as class 1) so that it remains below a desired threshold. To address this need, the Neyman-Pearson (NP) classification paradigm is a natural choice; it minimizes type II error (i.e., the conditional probability of misclassifying a class 1 observation as class 0) while enforcing an upper bound, alpha, on the type I error. Although the NP paradigm has a century-long history in hypothesis testing, it has not been well recognized and implemented in classification schemes. Common practices that directly limit the empirical type I error to no more than alpha do not satisfy the type I error control objective because the resulting classifiers are still likely to have type I errors much larger than alpha. As a result, the NP paradigm has not been properly implemented for many classification scenarios in practice. In this work, we develop the first umbrella algorithm that implements the NP paradigm for all scoring-type classification methods, including popular methods such as logistic regression, support vector machines and random forests. Powered by this umbrella algorithm, we propose a novel graphical tool for NP classification methods: NP receiver operating characteristic (NP-ROC) bands, motivated by the popular receiver operating characteristic (ROC) curves. NP-ROC bands will help choose in a data adaptive way and compare different NP classifiers.
Indices, heuristics, simulations and strategies to help determine the number of factors/components to retain in exploratory factor analysis and principal component analysis.
Nonparametric test of independence between a pair of spatial objects (random fields, point processes) based on random shifts with torus or variance correction. See MrkviÄ ka et al. (2021) <doi:10.1016/j.spasta.2020.100430>, DvoŠák et al. (2022) <doi:10.1111/insr.12503>, DvoŠák and MrkviÄ ka (2024) <doi:10.1080/10618600.2024.2357626>.
This package provides a collection of network analytic (convenience) functions which are missing in other standard packages. This includes triad census with attributes <doi:10.1016/j.socnet.2019.04.003>, core-periphery models <doi:10.1016/S0378-8733(99)00019-2>, and several graph generators. Most functions are build upon igraph'.
Based on Natural Earth <https://www.naturalearthdata.com/>, a subset of countries can easily be selected with their administrative boundaries, joined with an external data frame and plotted as a thematic map.
This package provides tools for modelling, ML estimation, validation analysis and simulation of non homogeneous Poisson processes in time.
Fit multinomial logistic regression with a penalty on the nuclear norm of the estimated regression coefficient matrix, using proximal gradient descent.
Annotated neuroblastoma copy number profiles, a benchmark data set for change-point detection algorithms, as described by Hocking et al. <doi:10.1186/1471-2105-14-164>.
Nonparametric maximum likelihood estimation or Gaussian quadrature for overdispersed generalized linear models and variance component models.
Density, distribution function, quantile function and random generation for the 3D Navarro, Frenk & White (NFW) profile. For details see Robotham & Howlett (2018) <arXiv:1805.09550>.
Accompanies the book "Nonparametric Statistical Methods Using R, 2nd Edition" by Kloke and McKean (2024, ISBN:9780367651350). Includes methods, datasets, and random number generation useful for the study of robust and/or nonparametric statistics. Emphasizes classical nonparametric methods for a variety of designs --- especially one-sample and two-sample problems. Includes methods for general scores, including estimation and testing for the two-sample location problem as well as Hogg's adaptive method.
Computes the probability density function, the cumulative density function, quantile function, random numbers and measures of inference for the following families exponentiated generalized gull alpha power family, exponentiated gull alpha powerfamily, gull alpha power family.
This package provides tools for estimating Receiver Operating Characteristic (ROC) curves, building confidence bands, comparing several curves both for dependent and independent data, estimating the cumulative-dynamic ROC curve in presence of censored data, and performing meta-analysis studies, among others.
Collapse, partition, combine, fill gaps in and expand date/time ranges.
This package provides a reproducible workflow for binning and visualizing NMR (nuclear magnetic resonance) spectra from environmental samples. The nmrrr package is intended for post-processing of NMR data, including importing, merging and, cleaning data from multiple files, visualizing NMR spectra, performing binning/integrations for compound classes, and relative abundance calculations. This package can be easily inserted into existing analysis workflows by users to help with analyzing and interpreting NMR data.
This package provides a number series generator that creates number series items based on cognitive models.
Automated population pharmacokinetic modeling framework for data-driven initialisation, model evaluation, and metaheuristic optimization. Supports genetic algorithms, ant colony optimization, tabu search, and stepwise procedures for automated model selection and parameter estimation within the nlmixr2 ecosystem.
With this package, it is possible to compute nonparametric simultaneous confidence intervals for relative contrast effects in the unbalanced one way layout. Moreover, it computes simultaneous p-values. The simultaneous confidence intervals can be computed using multivariate normal distribution, multivariate t-distribution with a Satterthwaite Approximation of the degree of freedom or using multivariate range preserving transformations with Logit or Probit as transformation function. 2 sample comparisons can be performed with the same methods described above. There is no assumption on the underlying distribution function, only that the data have to be at least ordinal numbers. See Konietschke et al. (2015) <doi:10.18637/jss.v064.i09> for details.
This package provides a simple function for easier package loading and auto-installation.
Creation and selection of N-way Partial Least Squares (NPLS) models. Selection of the optimal number of components can be done using ncrossreg(). NPLS was originally described by Rasmus Bro, see <doi:10.1002/%28SICI%291099-128X%28199601%2910%3A1%3C47%3A%3AAID-CEM400%3E3.0.CO%3B2-C>.
Simulate demand and attributes for ready to launch new products during their life cycle, or during their introduction and growth phases. You provide the number of products, attributes, time periods and/or other parameters and npdsim can simulate for you the demand for each product during the considered time periods, and the attributes of each product. The simulation for the demand is based on the idea that each product has a shape and a level, where the level is the cumulative demand over the considered time periods, and the shape is the normalized demand across those time periods.
Estimation of relatively complex nonlinear mixed-effects models, including the Sigmoidal Mixed Model and the Piecewise Linear Mixed Model with abrupt or smooth transition, through a single intuitive line of code and with automated generation of starting values.