Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
An implementation of R's DBI interface using ODBC package as a back-end. This allows R to connect to any DBMS that has a ODBC driver.
Designed for the import, analysis, and visualization of dosimetric and volumetric data in Radiation Oncology, the tools herein enable import of dose-volume histogram information from multiple treatment planning system platforms and 3D structural representations and dosimetric information from DICOM-RT files. These tools also enable subsequent visualization and statistical analysis of these data.
Evaluates R source codes by variable and/or functions names. Similar source codes should deliver similarity coefficients near one. Since neither the frequency nor the order of the used names is considered, a manual inspection of the R source code is required to check for similarity. Possible use cases include detection of code clones for improving software quality and of plagiarism amongst students assignments.
Provide estimation and data generation tools for the quantile generalized beta regression model. For details, see Bourguignon, Gallardo and Saulo <arXiv:2110.04428> The package also provides tools to perform covariates selection.
Extend Rasch and Item Response Theory (IRT) analyses by providing tools for post-processing the output from five major IRT packages (i.e., eRm', psychotools', ltm', mirt', and TAM'). The current version provides the plotPIccc() function, which extracts from the return object of the originating package all information required to draw an extended Person-Item-Map (PIccc), showing any combination of * category characteristic curves (CCCs), * threshold characteristic curves (TCCs), * item characteristic curves (ICCs), * category information functions (CIFs), * item information functions (IIFs), * test information function (TIF), and the * standard error curve (S.E.). for uni- and multidimensional models (as far as supported by each package). It allows for selecting dimensions, items, and categories to plot and offers numerous options to adapt the output. The return object contains all calculated values for further processing.
In order to facilitate R instruction for actuaries, we have organized several sets of publicly available data of interest to non-life actuaries. In addition, we suggest a set of packages, which most practicing actuaries will use routinely. Finally, there is an R markdown skeleton for basic reserve analysis.
Create and combine HTML and PDF reports from within R. Possibility to design tables and listings for reporting and also include R plots.
Traditional noise filtering methods aim at removing noisy samples from a classification dataset. This package adapts classic and recent filtering techniques for use in regression problems, and it also incorporates methods specifically designed for regression data. In order to do this, it uses approaches proposed in the specialized literature, such as Martin et al. (2021) [<doi:10.1109/ACCESS.2021.3123151>] and Arnaiz-Gonzalez et al. (2016) [<doi:10.1016/j.eswa.2015.12.046>]. Thus, the goal of the implemented noise filters is to eliminate samples with noise in regression datasets.
Estimates flexible epidemiological effect measures including both differences and ratios using the parametric G-formula developed as an alternative to inverse probability weighting. It is useful for estimating the impact of interventions in the presence of treatment-confounder-feedback. G-computation was originally described by Robbins (1986) <doi:10.1016/0270-0255(86)90088-6> and has been described in detail by Ahern, Hubbard, and Galea (2009) <doi:10.1093/aje/kwp015>; Snowden, Rose, and Mortimer (2011) <doi:10.1093/aje/kwq472>; and Westreich et al. (2012) <doi:10.1002/sim.5316>.
This package provides a method generate() is implemented in this package for the random generation of vector time series according to models obtained by RMAWGEN', vars or other packages. This package was created to generalize the algorithms of the RMAWGEN package for the analysis and generation of any environmental vector time series.
Connector to the REST API of a Rock R server, to perform operations on a remote R server session, or administration tasks. See Rock documentation at <https://rockdoc.obiba.org/>.
Visualize networks using the javascript library roughjs'. This allows to draw sketchy, hand-drawn-like networks.
This package implements a null model analysis to quantify concurrent temporal niche overlap (i.e., activity or phenology) among biological identities (e.g., individuals, populations, species) using the Rosario randomization algorithm Castro-Arellano et al. (2010) <doi:10.1111/j.2041-210X.2010.00031.x>.
This package provides functions for generating k-record values and k-record times.
This package provides a RESTful API wrapper for accessing the GENESIS database of the German Federal Statistical Office (Destatis) as well as its Census Database and the database of Germany's regional statistics. Supports data search functions, credential management, result caching, and handling remote background jobs for large datasets.
Enhances the R Optimization Infrastructure ('ROI') package with the optimx package.
This package provides datasets related to the Star Trek fictional universe and functions for working with the data. The package also provides access to real world datasets based on the televised series and other related licensed media productions. It interfaces with the Star Trek API (STAPI) (<http://stapi.co/>), Memory Alpha (<https://memory-alpha.fandom.com/wiki/Portal:Main>), and Memory Beta (<https://memory-beta.fandom.com/wiki/Main_Page>) to retrieve data, metadata and other information relating to Star Trek. It also contains several local datasets covering a variety of topics. The package also provides functions for working with data from other Star Trek-related R data packages containing larger datasets not stored in rtrek'.
Praat <https://www.fon.hum.uva.nl/praat/> is a widely used tool for manipulating, annotating and analyzing speech and acoustic data. It stores annotation data in a format called a TextGrid'. This package provides a way to read these files into R.
Allows work with Management API for load counters, segments, filters, user permissions and goals list from Yandex Metrica, Reporting API allows you to get information about the statistics of site visits and other data without using the web interface, Logs API allows to receive non-aggregated data and Compatible with Google Analytics Core Reporting API v3 allows receive information about site traffic and other data using field names from Google Analytics Core API. For more information see official documents <https://yandex.ru/dev/metrika/doc/api2/concept/about-docpage>.
Reads river network shape files and computes network distances. Also included are a variety of computation and graphical tools designed for fisheries telemetry research, such as minimum home range, kernel density estimation, and clustering analysis using empirical k-functions with a bootstrap envelope. Tools are also provided for editing the river networks, meaning there is no reliance on external software.
Create custom keyboard shortcuts to examine code selected in the Rstudio editor. F3 can for example yield str(selection) and F7 open the source code of CRAN and base package functions on github'.
This package provides an interface to Mapzen'-based APIs (including geocode.earth, Nextzen, and NYC GeoSearch) for geographic search and geocoding, isochrone calculation, and vector data to draw map tiles. See <https://www.mapzen.com/documentation/> for more information. The original Mapzen has gone out of business, but rmapzen can be set up to work with any provider who implements the Mapzen API.
Helps users in quickly visualizing risk-of-bias assessments performed as part of a systematic review. It allows users to create weighted bar-plots of the distribution of risk-of-bias judgments within each bias domain, in addition to traffic-light plots of the specific domain-level judgments for each study. The resulting figures are of publication quality and are formatted according the risk-of-bias assessment tool use to perform the assessments. Currently, the supported tools are ROB2.0 (for randomized controlled trials; Sterne et al (2019) <doi:10.1136/bmj.l4898>), ROBINS-I (for non-randomised studies of interventions; Sterne et al (2016) <doi:10.1136/bmj.i4919>), and QUADAS-2 (for diagnostic accuracy studies; Whiting et al (2011) <doi:10.7326/0003-4819-155-8-201110180-00009>).
We provide a toolbox to fit and simulate a univariate or multivariate damped random walk process that is also known as an Ornstein-Uhlenbeck process or a continuous-time autoregressive model of the first order, i.e., CAR(1) or CARMA(1, 0). This process is suitable for analyzing univariate or multivariate time series data with irregularly-spaced observation times and heteroscedastic measurement errors. When it comes to the multivariate case, the number of data points (measurements/observations) available at each observation time does not need to be the same, and the length of each time series can vary. The number of time series data sets that can be modeled simultaneously is limited to ten in this version of the package. We use Kalman-filtering to evaluate the resulting likelihood function, which leads to a scalable and efficient computation in finding maximum likelihood estimates of the model parameters or in drawing their posterior samples. Please pay attention to loading the data if this package is used for astronomical data analyses; see the details in the manual. Also see Hu and Tak (2020) <arXiv:2005.08049>.