Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Interface to SWI'-'Prolog', <https://www.swi-prolog.org/>. This package is normally not loaded directly, please refer to package rolog instead. The purpose of this package is to provide the Prolog runtime on systems that do not have a software installation of SWI'-'Prolog'.
Nuclear Decay Data for Dosimetric Calculations from the International Commission on Radiological Protection from ICRP Publication 107. Ann. ICRP 38 (3). Eckerman, Keith and Endo, Akira 2008 <doi:10.1016/j.icrp.2008.10.004> <https://www.icrp.org/publication.asp?id=ICRP%20Publication%20107>. This is a database of the physical data needed in calculations of radionuclide-specific protection and operational quantities. The data is prescribed by the ICRP, the international authority on radiation dose standards, for estimating dose from the intake of or exposure to radionuclides in the workplace and the environment. The database contains information on the half-lives, decay chains, and yields and energies of radiations emitted in nuclear transformations of 1252 radionuclides of 97 elements.
Create an R Journal Rmarkdown template article, that will generate html and pdf versions of your paper. Check that the paper folder has all the required components needed for submission. Examples of R Journal publications can be found at <https://journal.r-project.org>.
An R interface for processing concentration-response datasets using Curvep, a response noise filtering algorithm. The algorithm was described in the publications (Sedykh A et al. (2011) <doi:10.1289/ehp.1002476> and Sedykh A (2016) <doi:10.1007/978-1-4939-6346-1_14>). Other parametric fitting approaches (e.g., Hill equation) are also adopted for ease of comparison. 3-parameter Hill equation from tcpl package (Filer D et al., <doi:10.1093/bioinformatics/btw680>) and 4-parameter Hill equation from Curve Class2 approach (Wang Y et al., <doi:10.2174/1875397301004010057>) are available. Also, methods for calculating the confidence interval around the activity metrics are also provided. The methods are based on the bootstrap approach to simulate the datasets (Hsieh J-H et al. <doi:10.1093/toxsci/kfy258>). The simulated datasets can be used to derive the baseline noise threshold in an assay endpoint. This threshold is critical in the toxicological studies to derive the point-of-departure (POD).
Extract the implied risk neutral density from options using various methods.
This package provides a data structure and toolkit for documenting and recoding categorical data that can be shared in other statistical software.
This is a Google Forms and Google Classroom API Wrapper for R for managing Google Classrooms from R. The documentation for these APIs is here <https://developers.google.com/forms/api/guides> .
This package provides a simple R -> Stata interface allowing the user to execute Stata commands (both inline and from a .do file) from R.
Recursive partitioning for least absolute deviation regression trees. Another algorithm from the 1984 book by Breiman, Friedman, Olshen and Stone in addition to the rpart package (Breiman, Friedman, Olshen, Stone (1984, ISBN:9780412048418).
Computation of (direct and indirect) revealed preferences, fast non-parametric tests of rationality axioms (WARP, SARP, GARP), simulation of axiom-consistent data, and detection of axiom-consistent subpopulations. Rationality tests follow Varian (1982) <doi:10.2307/1912771>, axiom-consistent subpopulations follow Crawford and Pendakur (2012) <doi:10.1111/j.1468-0297.2012.02545.x>.
Fit statistical models based on the Dawid-Skene model - Dawid and Skene (1979) <doi:10.2307/2346806> - to repeated categorical rating data. Full Bayesian inference for these models is supported through the Stan modelling language. rater also allows the user to extract and plot key parameters of these models.
This package provides functions to calculate Sample Number and Average Sample Number for Repetitive Group Sampling Plan Based on Cpk as given in Aslam et al. (2013) (<DOI:10.1080/00949655.2012.663374>).
Facilities for working with Atlantis box-geometry model (BGM) files. Atlantis is a deterministic, biogeochemical, whole-of-ecosystem model. Functions are provided to read from BGM files directly, preserving their internal topology, as well as helper functions to generate spatial data from these mesh forms. This functionality aims to simplify the creation and modification of box and geometry as well as the ability to integrate with other data sources.
Build powerful pivot tables (aka Pivot Grid, Pivot Chart, Cross-Tab) and dynamically slice & dice / drag n drop your data. rpivotTable is a wrapper of pivottable', a powerful open-source Pivot Table library implemented in JavaScript by Nicolas Kruchten. Aligned to pivottable v2.19.0.
Hybrid Mortality Modelling (HMM) provides a framework in which mortality around "the accident hump" and at very old ages can be modelled under a single model. The graphics codes necessary for visualization of the models output are included here. Specifically, the graphics are based on the assumption that, the mortality rates can be expressed as a function of the area under the curve between the crude mortality rates plots and the tangential transform of the force of mortality.
Create interactive radar charts using the Chart.js JavaScript library and the htmlwidgets package. Chart.js <http://www.chartjs.org/> is a lightweight library that supports several types of simple chart using the HTML5 canvas element. This package provides an R interface specifically to the radar chart, sometimes called a spider chart, for visualising multivariate data.
Implementation of the relative placement algorithm widely used in the scoring of Lindy Hop and West Coast Swing dance contests.
Processes standard recommendation datasets (e.g., a user-item rating matrix) as input and generates rating predictions and lists of recommended items. Standard algorithm implementations which are included in this package are the following: Global/Item/User-Average baselines, Weighted Slope One, Item-Based KNN, User-Based KNN, FunkSVD, BPR and weighted ALS. They can be assessed according to the standard offline evaluation methodology (Shani, et al. (2011) <doi:10.1007/978-0-387-85820-3_8>) for recommender systems using measures such as MAE, RMSE, Precision, Recall, F1, AUC, NDCG, RankScore and coverage measures. The package (Coba, et al.(2017) <doi: 10.1007/978-3-319-60042-0_36>) is intended for rapid prototyping of recommendation algorithms and education purposes.
Reliable and flexible tools for scoring redistricting plans using common measures and metrics. These functions provide key direct access to tools useful for non-simulation analyses of redistricting plans, such as for measuring compactness or partisan fairness. Tools are designed to work with the redist package seamlessly.
This package provides a Bayesian-weighted estimator and two unweighted estimators are developed to estimate the number of newly found rare species in additional ecological samples. Among these methods, the Bayesian-weighted estimator and an unweighted (Chao-derived) estimator are of high accuracy and recommended for practical applications. Technical details of the proposed estimators have been well described in the following paper: Shen TJ, Chen YH (2018) A Bayesian weighted approach to predicting the number of newly discovered rare species. Conservation Biology, In press.
Compute an exact CI for the population mean under a random effects model. The routines implement the algorithm described in Michael, Thronton, Xie, and Tian (2017) <https://haben-michael.github.io/research/Exact_Inference_Meta.pdf>.
An Rcpp interface for Eunjeon project <http://eunjeon.blogspot.com/>. The mecab-ko and mecab-ko-dic is based on a C++ library, and part-of-speech tagging with them is useful when the spacing of source Korean text is not correct. This package provides part-of-speech tagging and tokenization function for Korean text.
This package provides a method to download Department of Education College Scorecard data using the public API <https://collegescorecard.ed.gov/data/data-documentation/>. It is based on the dplyr model of piped commands to select and filter data in a single chained function call. An API key from the U.S. Department of Education is required.
Toolbox with shiny applications for widely used psychometric methods. Those methods include following analysis: Item analysis, item response theory calibration, principal component analysis, confirmatory factor analysis - structural equation modeling, generating simulated data. References: Chalmers (2012, <doi:10.18637/jss.v048.i06>); Revelle (2022, <https://CRAN.R-project.org/package=psych Version = 2.2.9.>); Rosseel (2012, <doi:10.18637/jss.v048.i02>); Magis & Raiche (2012, <doi:10.18637/jss.v048.i08>); Magis & Barrada (2017, <doi:10.18637/jss.v076.c01>).