Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Fetching data from Amazon Kinesis Streams using the Java-based MultiLangDaemon interacting with Amazon Web Services ('AWS') for easy stream processing from R. For more information on Kinesis', see <https://aws.amazon.com/kinesis>.
This package provides WHO Child Growth Standards (z-scores) with confidence intervals and standard errors around the prevalence estimates, taking into account complex sample designs. More information on the methods is available online: <https://www.who.int/tools/child-growth-standards>.
This package provides tools for downloading hourly averages, daily maximums and minimums from each of the pollution, wind, and temperature measuring stations or geographic zones in the Mexico City metro area. The package also includes the locations of each of the stations and zones. See <http://aire.cdmx.gob.mx/> for more information.
Interface to Altair <https://altair-viz.github.io>, which itself is a Python interface to Vega-Lite <https://vega.github.io/vega-lite/>. This package uses the Reticulate framework <https://rstudio.github.io/reticulate/> to manage the interface between R and Python'.
This package implements Bayesian estimation and inference for alpha-mixture survival models, including Weibull and Exponential based components, with tools for simulation and posterior summaries. The methods target applications in reliability and biomedical survival analysis. The package implements Bayesian estimation for the alpha-mixture methodology introduced in Asadi et al. (2019) <doi:10.1017/jpr.2019.72>.
Computationally efficient procedures for regularized estimation with the semiparametric additive hazards regression model.
Analysis of task-related functional magnetic resonance imaging (fMRI) activity at the level of individual participants is commonly based on general linear modelling (GLM) that allows us to estimate to what extent the blood oxygenation level dependent (BOLD) signal can be explained by task response predictors specified in the GLM model. The predictors are constructed by convolving the hypothesised timecourse of neural activity with an assumed hemodynamic response function (HRF). To get valid and precise estimates of task response, it is important to construct a model of neural activity that best matches actual neuronal activity. The construction of models is most often driven by predefined assumptions on the components of brain activity and their duration based on the task design and specific aims of the study. However, our assumptions about the onset and duration of component processes might be wrong and can also differ across brain regions. This can result in inappropriate or suboptimal models, bad fitting of the model to the actual data and invalid estimations of brain activity. Here we present an approach in which theoretically driven models of task response are used to define constraints based on which the final model is derived computationally using the actual data. Specifically, we developed autohrf â a package for the R programming language that allows for data-driven estimation of HRF models. The package uses genetic algorithms to efficiently search for models that fit the underlying data well. The package uses automated parameter search to find the onset and duration of task predictors which result in the highest fitness of the resulting GLM based on the fMRI signal under predefined restrictions. We evaluate the usefulness of the autohrf package on publicly available datasets of task-related fMRI activity. Our results suggest that by using autohrf users can find better task related brain activity models in a quick and efficient manner.
In panel data settings, specifies set of candidate models, fits them to data from pre-treatment validation periods, and selects model as average over candidate models, weighting each by posterior probability of being most robust given its differential average prediction errors in pre-treatment validation periods. Subsequent estimation and inference of causal effect's bounds accounts for both model and sampling uncertainty, and calculates the robustness changepoint value at which bounds go from excluding to including 0. The package also includes a range of diagnostic plots, such as those illustrating models differential average prediction errors and the posterior distribution of which model is most robust.
This package contains data from an observational study concerning possible effects of light daily alcohol consumption on survival and on HDL cholesterol. It also replicates various simple analyses in Rosenbaum (2025a) <doi:10.1080/09332480.2025.2473291>. Finally, it includes new R code in wgtRankCef() that implements and replicates a new method for constructing evidence factors in observational block designs.
Obtain network structures from animal GPS telemetry observations and statistically analyse them to assess their adequacy for social network analysis. Methods include pre-network data permutations, bootstrapping techniques to obtain confidence intervals for global and node-level network metrics, and correlation and regression analysis of the local network metrics.
Allows users to stem Arabic texts for text analysis.
This package performs the two-sample Ansariâ Bradley test (Ansari & Bradley, 1960 <https://www.jstor.org/stable/2237814>) for univariate, distinct data in the presence of missing values, as described in Zeng et al. (2025) <doi:10.48550/arXiv.2509.20332>. This method does not make any assumptions about the missingness mechanisms and controls the Type I error regardless of the missing values by taking all possible missing values into account.
This package provides a Tool for Semi-Automating the Statistical Disclosure Control of Research Outputs.
Supports the analysis of oceanographic data recorded by Argo autonomous drifting profiling floats. Functions are provided to (a) download and cache data files, (b) subset data in various ways, (c) handle quality-control flags and (d) plot the results according to oceanographic conventions. A shiny app is provided for easy exploration of datasets. The package is designed to work well with the oce package, providing a wide range of processing capabilities that are particular to oceanographic analysis. See Kelley, Harbin, and Richards (2021) <doi:10.3389/fmars.2021.635922> for more on the scientific context and applications.
Nonparametric data-driven approach to discovering heterogeneous subgroups in a selection-on-observables framework. aggTrees allows researchers to assess whether there exists relevant heterogeneity in treatment effects by generating a sequence of optimal groupings, one for each level of granularity. For each grouping, we obtain point estimation and inference about the group average treatment effects. Please reference the use as Di Francesco (2022) <doi:10.2139/ssrn.4304256>.
Routines for re-scaling isotope maps using known-origin tissue isotope data, assigning origin of unknown samples, and summarizing and assessing assignment results. Methods are adapted from Wunder (2010, in ISBN:9789048133536) and Vander Zanden, H. B. et al. (2014) <doi:10.1111/2041-210X.12229> as described in Ma, C. et al. (2020) <doi:10.1111/2041-210X.13426>.
The Ata method (Yapar et al. (2019) <doi:10.15672/hujms.461032>), an alternative to exponential smoothing (described in Yapar (2016) <doi:10.15672/HJMS.201614320580>, Yapar et al. (2017) <doi:10.15672/HJMS.2017.493>), is a new univariate time series forecasting method which provides innovative solutions to issues faced during the initialization and optimization stages of existing forecasting methods. Forecasting performance of the Ata method is superior to existing methods both in terms of easy implementation and accurate forecasting. It can be applied to non-seasonal or seasonal time series which can be decomposed into four components (remainder, level, trend and seasonal). This methodology performed well on the M3 and M4-competition data. This package was written based on Ali Sabri Taylanâ s PhD dissertation.
Response surface designs (RSDs) are widely used for Response Surface Methodology (RSM) based optimization studies, which aid in exploring the relationship between a group of explanatory variables and one or more response variable(s) (G.E.P. Box and K.B. Wilson (1951), "On the experimental attainment of optimum conditions" ; M. Hemavathi, Shashi Shekhar, Eldho Varghese, Seema Jaggi, Bikas Sinha & Nripes Kumar Mandal (2022) <DOI: 10.1080/03610926.2021.1944213>."Theoretical developments in response surface designs: an informative review and further thoughts".). Second order rotatable designs are the most prominent and popular class of designs used for process and product optimization trials but it is suitable for situations when all the number of levels for each factor is the same. In many practical situations, RSDs with asymmetric levels (J.S. Mehta and M.N. Das (1968). "Asymmetric rotatable designs and orthogonal transformations" ; M. Hemavathi, Eldho Varghese, Shashi Shekhar & Seema Jaggi (2020) <DOI: 10.1080/02664763.2020.1864817>. "Sequential asymmetric third order rotatable designs (SATORDs)" .) are more suitable as these designs explore more regions in the design space.This package contains functions named Asords() ,CCD_coded(), CCD_original(), SORD_coded() and SORD_original() for generating asymmetric/symmetric RSDs along with the randomized layout. It also contains another function named Pred.var() for generating the variance of predicted response as well as the moment matrix based on a second order model.
Imports Azure Application Insights for web pages into Shiny apps via Microsoft's JavaScript snippet. Allows app developers to submit page tracking and submit events.
This package provides functions that facilitate the use of accepted taxonomic nomenclature, collection of functional trait data, and assignment of functional group classifications to phytoplankton species. Possible classifications include Morpho-functional group (MFG; Salmaso et al. 2015 <doi:10.1111/fwb.12520>) and CSR (Reynolds 1988; Functional morphology and the adaptive strategies of phytoplankton. In C.D. Sandgren (ed). Growth and reproductive strategies of freshwater phytoplankton, 388-433. Cambridge University Press, New York). Versions 2.0.0 and later includes new functions for querying the algaebase online taxonomic database (www.algaebase.org), however these functions require a valid API key that must be acquired from the algaebase administrators. Note that none of the algaeClassify authors are affiliated with algaebase in any way. Taxonomic names can also be checked against a variety of taxonomic databases using the Global Names Resolver service via its API (<https://resolver.globalnames.org/api>). In addition, currently accepted and outdated synonyms, and higher taxonomy, can be extracted for lists of species from the ITIS database using wrapper functions for the ritis package. The algaeClassify package is a product of the GEISHA (Global Evaluation of the Impacts of Storms on freshwater Habitat and Structure of phytoplankton Assemblages), funded by CESAB (Centre for Synthesis and Analysis of Biodiversity) and the U.S. Geological Survey John Wesley Powell Center for Synthesis and Analysis, with data and other support provided by members of GLEON (Global Lake Ecology Observation Network). DISCLAIMER: This software has been approved for release by the U.S. Geological Survey (USGS). Although the software has been subjected to rigorous review, the USGS reserves the right to update the software as needed pursuant to further analysis and review. No warranty, expressed or implied, is made by the USGS or the U.S. Government as to the functionality of the software and related material nor shall the fact of release constitute any such warranty. Furthermore, the software is released on condition that neither the USGS nor the U.S. Government shall be held liable for any damages resulting from its authorized or unauthorized use.
An interface to Azure Queue Storage'. This is a cloud service for storing large numbers of messages, for example from automated sensors, that can be accessed remotely via authenticated calls using HTTP or HTTPS. Queue storage is often used to create a backlog of work to process asynchronously. Part of the AzureR family of packages.
The AIPW package implements the augmented inverse probability weighting, a doubly robust estimator, for average causal effect estimation with user-defined stacked machine learning algorithms. To cite the AIPW package, please use: "Yongqi Zhong, Edward H. Kennedy, Lisa M. Bodnar, Ashley I. Naimi (2021). AIPW: An R Package for Augmented Inverse Probability Weighted Estimation of Average Causal Effects. American Journal of Epidemiology. <doi:10.1093/aje/kwab207>". Visit: <https://yqzhong7.github.io/AIPW/> for more information.
This package provides non-invasive annotation of package load calls such as \codelibrary(), \codep_load(), and \coderequire() so that we can have an idea of what the packages we are loading are meant for.
This is a simple and powerful package to create, render, preview, and deploy documentation websites for R packages. It is a lightweight and flexible alternative to pkgdown', with support for many documentation generators, including Quarto', Docute', Docsify', and MkDocs'.