Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Retrieve, map and summarize data from the VertNet.org archives (<https://vertnet.org/>). Functions allow searching by many parameters, including taxonomic names, places, and dates. In addition, there is an interface for conducting spatially delimited searches, and another for requesting large datasets via email.
The Kolmogorov-Smirnov (K-S) statistic is a standard method to measure the model strength for credit risk scoring models. This package calculates the Kâ S statistic and plots the true-positive rate and false-positive rate to measure the model strength. This package was written with the credit marketer, who uses risk models in conjunction with his campaigns. The users could read more details from Thrasher (1992) <doi:10.1002/dir.4000060408> and pyks <https://pypi.org/project/pyks/>.
Describes a new procedure of reducing items in a rating scale called Rating Scale Reduction (RSR). The new stop criterion in RSR procedure is added (stop global max). The function order is replaced by sort.list.
Computes word, character, and non-whitespace character counts in R Markdown documents and Jupyter notebooks, with or without code chunks. Returns results as a data frame.
Read Statistical Data and Metadata Exchange (SDMX) XML data. This the main transmission format used in official statistics. Data can be imported from local SDMX-ML files or a SDMX web-service and will be read in as is into a dataframe object. The RapidXML C++ library <https://rapidxml.sourceforge.net/> is used to parse the XML data.
Empirical best linear unbiased prediction (EBLUP) and robust prediction of the area-level means under the basic unit-level model. The model can be fitted by maximum likelihood or a (robust) M-estimator. Mean square prediction error is computed by a parametric bootstrap.
An extension for roxygen2 to embed Shinylive applications in the package documentation.
Various tools for handling fuzzy measures, calculating Shapley value and interaction index, Choquet and Sugeno integrals, as well as fitting fuzzy measures to empirical data are provided. Construction of fuzzy measures from empirical data is done by solving a linear programming problem by using lpsolve package, whose source in C adapted to the R environment is included. The description of the basic theory of fuzzy measures is in the manual in the Doc folder in this package. Please refer to the following: [1] <https://personal-sites.deakin.edu.au/~gleb/fmtools.html> [2] G. Beliakov, H. Bustince, T. Calvo, A Practical Guide to Averaging', Springer, (2016, ISBN: 978-3-319-24753-3). [3] G. Beliakov, S. James, J-Z. Wu, Discrete Fuzzy Measures', Springer, (2020, ISBN: 978-3-030-15305-2).
This is a companion package of the book "R Programming: Zero to Pro" <https://r02pro.github.io/>. It contains the datasets used in the book and provides interactive exercises corresponding to the book. It covers a wide range of topics including visualization, data transformation, tidying data, data input and output.
Extract text or metadata from over a thousand file types, using Apache Tika <https://tika.apache.org/>. Get either plain text or structured XHTML content.
Suite of tools for using D3', a library for producing dynamic, interactive data visualizations. Supports translating objects into D3 friendly data structures, rendering D3 scripts, publishing D3 visualizations, incorporating D3 in R Markdown, creating interactive D3 applications with Shiny, and distributing D3 based htmlwidgets in R packages.
An implementation of algorithms for estimation of the graphical lasso regularization parameter described in Pedro Cisneros-Velarde, Alexander Petersen and Sang-Yun Oh (2020) <http://proceedings.mlr.press/v108/cisneros20a.html>.
Visualize the objects in orbits in 2D and 3D. The packages is under developing to plot the orbits of objects in polar coordinate system. See the examples in demo.
This package performs Wavelet Lifting Transforms focusing on signal denoising and functional data analysis (FDA). Implements a hybrid architecture with a zero-allocation C++ core for high-performance processing. Features include unified offline (batch) denoising, causal (real-time) filtering using a ring buffer engine, and adaptive recursive thresholding.
This package implements standard and reference based multiple imputation methods for continuous longitudinal endpoints (Gower-Page et al. (2022) <doi:10.21105/joss.04251>). In particular, this package supports deterministic conditional mean imputation and jackknifing as described in Wolbers et al. (2022) <doi:10.1002/pst.2234>, Bayesian multiple imputation as described in Carpenter et al. (2013) <doi:10.1080/10543406.2013.834911>, and bootstrapped maximum likelihood imputation as described in von Hippel and Bartlett (2021) <doi: 10.1214/20-STS793>.
This package provides interface to the Bioinfo-C (internal name: BIOS') library and utilities. ribiosUtils is a Swiss-knife for computational biology in drug discovery, providing functions and utilities with minimal external dependency and maximal efficiency.
An interface to the Integrated Taxonomic Information System ('ITIS') (<https://www.itis.gov>). Includes functions to work with the ITIS REST API methods (<https://www.itis.gov/ws_description.html>), as well as the Solr web service (<https://www.itis.gov/solr_documentation.html>).
Collect your data on digital marketing campaigns from reddit Ads using the Windsor.ai API <https://windsor.ai/api-fields/>.
This package provides a tool for multiply imputing missing data using MIDAS', a deep learning method based on denoising autoencoder neural networks (see Lall and Robinson, 2022; <doi:10.1017/pan.2020.49>). This algorithm offers significant accuracy and efficiency advantages over other multiple imputation strategies, particularly when applied to large datasets with complex features. Alongside interfacing with Python to run the core algorithm, this package contains functions for processing data before and after model training, running imputation model diagnostics, generating multiple completed datasets, and estimating regression models on these datasets. For more information see Lall and Robinson (2023) <doi:10.18637/jss.v107.i09>.
This is a collection of functions designed for simulating, estimating and forecasting seasonal functional autoregressive time series of order one. These methods are addressed in the manuscript: <https://www.monash.edu/business/ebs/research/publications/ebs/wp16-2019.pdf>.
This package provides a reliable and validated tool that captures detailed risk metrics such as R CMD check, test coverage, traceability matrix, documentation, dependencies, reverse dependencies, suggested dependency analysis, repository data, and enhanced reporting for R packages that are local or stored on remote repositories such as GitHub, CRAN, and Bioconductor.
This package provides a method to download Department of Education College Scorecard data using the public API <https://collegescorecard.ed.gov/data/data-documentation/>. It is based on the dplyr model of piped commands to select and filter data in a single chained function call. An API key from the U.S. Department of Education is required.
This package provides tools for creating data validation pipelines and tidy reports. This package offers a framework for exploring and validating data frame like objects using dplyr grammar of data manipulation.
Adds menu items for discrete choice experiments (DCEs) to the R Commander. DCE is a question-based survey method that designs various combinations (profiles) of attribute levels using the experimental designs, asks respondents to select the most preferred profile in each choice set, and then measures preferences for the attribute levels by analyzing the responses. For details on DCEs, refer to Louviere et al. (2000) <doi:10.1017/CBO9780511753831>.