Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Wraps the Ollama <https://ollama.com> API, which can be used to communicate with generative large language models locally.
The minimal rrapply'-package contains a single function rrapply(), providing an extended implementation of R'-base rapply() by allowing to recursively apply a function to elements of a nested list based on a general condition function and including the possibility to prune or aggregate nested list elements from the result. In addition, special arguments can be supplied to access the name, location, parents and siblings in the nested list of the element under evaluation. The rrapply() function builds upon rapply()'s native C implementation and requires no other package dependencies.
Bindings for additional models for use with the parsnip package. Models include prediction rule ensembles (Friedman and Popescu, 2008) <doi:10.1214/07-AOAS148>, C5.0 rules (Quinlan, 1992 ISBN: 1558602380), and Cubist (Kuhn and Johnson, 2013) <doi:10.1007/978-1-4614-6849-3>.
This package provides utility functions that extend the capabilities of the reference-based multiple imputation package rbmi'. It supports clinical trial analysis workflows with functions for managing imputed datasets, applying analysis methods across imputations, and tidying results for reporting.
This package provides tools to fit and simulate realizations from relational event models.
Helps to fit thermal performance curves (TPCs). rTPC contains 26 model formulations previously used to fit TPCs and has helper functions to set sensible start parameters, upper and lower parameter limits and estimate parameters useful in downstream analyses, such as cardinal temperatures, maximum rate and optimum temperature. See Padfield et al. (2021) <doi:10.1111/2041-210X.13585>.
Use the <https://api.nbp.pl/> API through R. Retrieve currency exchange rates and gold prices data published by the National Bank of Poland in form of convenient R objects.
Perform risk-adjusted regression and sensitivity analysis as developed in "Mitigating Omitted- and Included-Variable Bias in Estimates of Disparate Impact" Jung et al. (2024) <arXiv:1809.05651>.
This package provides a single method implementing multiple approaches to generate pseudo-random vectors whose components sum up to one (see, e.g., Maziero (2015) <doi:10.1007/s13538-015-0337-8>). The components of such vectors can for example be used for weighting objectives when reducing multi-objective optimisation problems to a single-objective problem in the socalled weighted sum scalarisation approach.
Allows for production of Czekanowski's Diagrams with clusters. See K. Bartoszek, A. Vasterlund (2020) <doi:10.2478/bile-2020-0008> and K. Bartoszek, Y. Luo (2023) <doi:10.14708/ma.v51i2.7259>. The suggested FuzzyDBScan package (which allows for fuzzy clustering) can be obtained from <https://github.com/henrifnk/FuzzyDBScan/> (or from CRAN's Archive <https://cran.r-project.org/src/contrib/Archive/FuzzyDBScan/>).
This package implements the regularized exponentially tilted empirical likelihood method. Details of the method are given in Kim, MacEachern, and Peruggia (2023) <doi:10.48550/arXiv.2312.17015>. This work was supported by the U.S. National Science Foundation under Grants No. SES-1921523 and DMS-2015552.
Test for effects of both individual factors and their interaction on replicated spatial patterns in a two factorial design, as explained in Ramon et al. (2016) <doi:10.1111/ecog.01848>.
Rasterize images using a 3D software renderer. 3D scenes are created either by importing external files, building scenes out of the included objects, or by constructing meshes manually. Supports point and directional lights, anti-aliased lines, shadow mapping, transparent objects, translucent objects, multiple materials types, reflection, refraction, environment maps, multicore rendering, bloom, tone-mapping, and screen-space ambient occlusion.
Automatically apply different strategies to optimize R code. rco functions take R code as input, and returns R code as output.
Utility functions to download data from the RESOURCECODE hindcast database of sea-states, time series of sea-state parameters and time series of 1D and 2D wave spectra. See <https://resourcecode.ifremer.fr> for more details about the available data. Also provides facilities to plot and analyse downloaded data, such as computing the sea-state parameters from both the 1D and 2D surface elevation variance spectral density.
Load data by campaigns, ads, ad sets and insights, ad account and business manager from Facebook Marketing API into R. For more details see official documents by Facebook Marketing API <https://developers.facebook.com/docs/marketing-api>.
Encode network data as strings of printable ASCII characters. Implemented functions include encoding and decoding adjacency matrices, edgelists, igraph, and network objects to/from formats graph6', sparse6', and digraph6'. The formats and methods are described in McKay, B.D. and Piperno, A (2014) <doi:10.1016/j.jsc.2013.09.003>.
Sends texts to the <https://www.receptiviti.com> API to be scored, and facilitates the creation of custom norms and local results databases.
Fits an Ising model to a binary dataset using L1 regularized logistic regression and extended BIC. Also includes a fast lasso logistic regression function for high-dimensional problems. Uses the libLBFGS optimization library by Naoaki Okazaki.
Perform the complete processing of a set of proton nuclear magnetic resonance spectra from the free induction decay (raw data) and based on a processing sequence (macro-command file). An additional file specifies all the spectra to be considered by associating their sample code as well as the levels of experimental factors to which they belong. More detail can be found in Jacob et al. (2017) <doi:10.1007/s11306-017-1178-y>.
Stan implementation of the Theory of Visual Attention (TVA; Bundesen, 1990; <doi:10.1037/0033-295X.97.4.523>) and numerous convenience functions for generating, compiling, fitting, and analyzing TVA models.
Suite of utilities for accessing and manipulating data from the KoboToolbox API. KoboToolbox is a robust platform designed for field data collection in various disciplines. This package aims to simplify the process of fetching and handling data from the API. Detailed documentation for the KoboToolbox API can be found at <https://support.kobotoolbox.org/api.html>.
The key function get_vintage_data() returns a dataframe and is the window into the Census Bureau API requiring just a dataset name, vintage(year), and vector of variable names for survey estimates/percentages. Other functions assist in searching for available datasets, geographies, group/variable concepts of interest. Also provided are functions to access and layer (via standard piping) displayable geometries for the US, states, counties, blocks/tracts, roads, landmarks, places, and bodies of water. Joining survey data with many of the geometry functions is built-in to produce choropleth maps.
The Radiant Design menu includes interfaces for design of experiments, sampling, and sample size calculation. The application extends the functionality in radiant.data'.