Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
The fusion learning method uses a model selection algorithm to learn from multiple data sets across different experimental platforms through group penalization. The responses of interest may include a mix of discrete and continuous variables. The responses may share the same set of predictors, however, the models and parameters differ across different platforms. Integrating information from different data sets can enhance the power of model selection. Package is based on Xin Gao, Raymond J. Carroll (2017) <arXiv:1610.00667v1>.
This package implements fast, scalable optimization algorithms for fitting generalized principal components analysis (GLM-PCA) models, as described in "A Generalization of Principal Components Analysis to the Exponential Family" Collins M, Dasgupta S, Schapire RE (2002, ISBN:9780262271738), and subsequently "Feature Selection and Dimension Reduction for Single-Cell RNA-Seq Based on a Multinomial Model" Townes FW, Hicks SC, Aryee MJ, Irizarry RA (2019) <doi:10.1186/s13059-019-1861-6>.
This package provides a collection of commonly used univariate and multivariate time series forecasting models including automatically selected exponential smoothing (ETS) and autoregressive integrated moving average (ARIMA) models. These models work within the fable framework provided by the fabletools package, which provides the tools to evaluate, visualise, and combine models in a workflow consistent with the tidyverse.
For ordinal rating data, consider the accelerated EM algorithm to estimate and test models within the family of CUB models (where CUB stands for Combination of a discrete Uniform and a shifted Binomial distributions). The procedure is built upon Louis identity for the observed information matrix. Best-subset variable selection is then implemented since it becomes more feasible from the computational point of view.
Interactive data visualization for data practitioners. flourishcharts allows users to visualize their data using Flourish graphs that are grounded in data storytelling principles. Users can create racing bar & line charts, as well as other interactive elements commonly found in D3 graphics, easily in R and Python'. The package relies on an enterprise API provided by Flourish', a data visualization platform <https://developers.flourish.studio/api/introduction/>.
This package provides a collection of functions to optimize portfolios and to analyze them from different points of view.
Aids in analysing data from a food frequency questionnaire known as the Harvard Service Food Frequency Questionnaire (HSFFQ). Functions from this package use answers from the HSFFQ to generate estimates of daily consumed micronutrients, calories, macronutrients on an individual level. The package also calculates food quotients on individual and group levels. Foodquotient calculation is an often tedious step in the calculation of total human energy expenditure (TEE) using the doubly labeled water method, which is the gold standard for measuring TEE.
This package provides a simple way to unload none-base packages and remove all global variables.
To help you access, transform, analyze, and visualize ForestGEO data, we developed a collection of R packages (<https://forestgeo.github.io/fgeo/>). This package, in particular, helps you to easily import, filter, and modify ForestGEO data. To learn more about ForestGEO visit <https://forestgeo.si.edu/>.
This package provides functions for plotting probability density functions, distribution functions, survival functions, hazard functions and computing distribution moments. The implementation is inspired by Delignette-Muller and Dutang (2015) <doi:10.18637/jss.v064.i04>.
Aim: Supports the most frequently used methods to combine forecasts. Among others: Simple average, Ordinary Least Squares, Least Absolute Deviation, Constrained Least Squares, Variance-based, Best Individual model, Complete subset regressions and Information-theoretic (information criteria based).
This package provides an interface to the Flickr API <https://www.flickr.com/services/api/> and allows R users to download data on Flickr.
This package provides a handy tool to calculate carbon footprints from air travel based on three-letter International Air Transport Association (IATA) airport codes or latitude and longitude. footprint first calculates the great-circle distance between departure and arrival destinations. It then uses the Department of Environment, Food & Rural Affairs (DEFRA) greenhouse gas conversion factors for business air travel to estimate the carbon footprint. These conversion factors consider trip length, flight class (e.g. economy, business), and emissions metric (e.g. carbon dioxide equivalent, methane).
Easily use Font Awesome icons as shiny favicons (the icons that appear on browser tabs). Font Awesome (<https://fontawesome.com/>) is a popular set of icons that can be used in web pages. favawesome provides a simple way to use these icons as favicons in shiny applications and other HTML pages.
This package implements fractional differencing with Autoregressive Moving Average models to analyse long-memory time series data. Traditional ARIMA models typically use integer values for differencing, which are suitable for time series with short memory or anti-persistent behaviour. In contrast, the Fractional ARIMA model allows fractional differencing, enabling it to effectively capture long memory characteristics in time series data. The âfracARMAâ package is user-friendly and allows users to manually input the fractional differencing parameter, which can be obtained using various estimators such as the GPH estimator, Sperio method, or Wavelet method and many. Additionally, the package enables users to directly feed the time series data, AR order, MA order, fractional differencing parameter, and the proportion of training data as a split ratio, all in a single command. The package is based on the reference from the paper of Irshad and others (2024, <doi:10.22271/maths.2024.v9.i6b.1906>).
The futurize() function transpiles calls to sequential map-reduce functions such as base::lapply(), purrr::map(), foreach::foreach() %do% ... into concurrent alternatives, providing you with a simple, straightforward path to scalable parallel computing via the future ecosystem <doi:10.32614/RJ-2021-048>. By combining this function with R's native pipe operator, you have an convenient way for speeding up iterative computations with minimal refactoring, e.g. lapply(xs, fcn) |> futurize()', purrr::map(xs, fcn) |> futurize()', and foreach::foreach(x = xs) %do% fcn(x) |> futurize()'. Other map-reduce packages that be "futurized" are BiocParallel', plyr', crossmap packages. There is also support for growing set of domain-specific packages, including boot', glmnet', mgcv', lme4', and tm'.
This package provides a collection of R games and other funny stuff, such as the classic Mine sweeper and sliding puzzles.
This package produces forest plots using ggplot2 from models produced by functions such as stats::lm(), stats::glm() and survival::coxph().
With the functions in this package you can check the validity of the following financial instrument identifiers: FIGI (Financial Instrument Global Identifier <https://www.openfigi.com/about/figi>), CUSIP (Committee on Uniform Security Identification Procedures <https://www.cusip.com/identifiers.html#/CUSIP>), ISIN (International Securities Identification Number <https://www.cusip.com/identifiers.html#/ISIN>), SEDOL (Stock Exchange Daily Official List <https://www2.lseg.com/SEDOL-masterfile-service-tech-guide-v8.6>). You can also calculate the FIGI checksum of 11-character strings, which can be useful if you want to create your own FIGI identifiers.
This package provides a fast Rcpp'-based implementation of polynomially-computable voting theory methods for committee ranking and scoring. The package includes methods such as Approval Voting (AV), Satisfaction Approval Voting (SAV), sequential Proportional Approval Voting (PAV), and sequential Phragmen's Rule. Weighted variants of these methods are also provided, allowing for differential voter influence.
Social Relations Analysis with roles ("Family SRM") are computed, using a structural equation modeling approach. Groups ranging from three members up to an unlimited number of members are supported and the mean structure can be computed. Means and variances can be compared between different groups of families and between roles.
In the Cramérâ Lundberg risk process perturbed by a Wiener process, this packages provides approximations to the probability of ruin within a finite time horizon. Currently, there are three methods implemented: The first one uses saddlepoint approximation (two variants are provided), the second one uses importance sampling and the third one is based on the simulation of a dual process. This last method is not very accurate and only given here for completeness.
Compare variables of interest between (potentially large numbers of) spatial interactions and meta-variables. Spatial variables are summarized using K, or other, functions, and projected for use in a modified random forest model. The model allows comparison of functional and non-functional variables to each other and to noise, giving statistical significance to the results. Included are preparation, modeling, and interpreting tools along with example datasets, as described in VanderDoes et al., (2023) <doi:10.1101/2023.07.18.549619>.
Allows user to obtain subsets of columns of data or vectors within a list. These subsets will match the original data in terms of average and variation, but have a consistent length of data per column. It is intended for use on automated data generation which may not always output the same N per replicate or sample.