Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Network Common Data Form (netCDF) files are widely used for scientific data. Library-level access in R is provided through packages RNetCDF and ncdf4. The package ncdfCF is built on top of RNetCDF and makes the data and its attributes available as a set of R6 classes that are informed by the Climate and Forecasting Metadata Conventions. Access to the data uses standard R subsetting operators and common function forms.
Writing interfaces to command line software is cumbersome. The cmdfun package provides a framework for building function calls to seamlessly interface with shell commands by allowing lazy evaluation of command line arguments. It also provides methods for handling user-specific paths to tool installs or secrets like API keys. Its focus is to equally serve package builders who wish to wrap command line software, and to help analysts stay inside R when they might usually leave to execute non-R software.
Artificial Bee Colony (ABC) is one of the most recently defined algorithms by Dervis Karaboga in 2005, motivated by the intelligent behavior of honey bees. It is as simple as Particle Swarm Optimization (PSO) and Differential Evolution (DE) algorithms, and uses only common control parameters such as colony size and maximum cycle number. The r-abcoptim implements the Artificial bee colony optimization algorithm http://mf.erciyes.edu.tr/abc/pub/tr06_2005.pdf. This version is a work-in-progress and is written in R code.
ICGE is a package that helps to estimate the number of real clusters in data as well as to identify atypical units. The underlying methods are based on distances rather than on unit x variables.
ActiLife generates activity counts from data collected by Actigraph accelerometers. Actigraph is one of the most common research-grade accelerometers. There is considerable research validating and developing algorithms for human activity using ActiLife counts. Unfortunately, ActiLife counts are proprietary and difficult to implement if researchers use different accelerometer brands. The code creates ActiLife counts from raw acceleration data for different accelerometer brands.
This package provides tools for determining estimability of linear functions of regression coefficients, and epredict methods that handle non-estimable cases correctly.
The Predictive Model Markup Language (PMML) is an XML-based language which provides a way for applications to define machine learning, statistical and data mining models and to share models between PMML compliant applications. More information about the PMML industry standard and the Data Mining Group can be found at http://dmg.org/. The generated PMML can be imported into any PMML consuming application, such as Zementis Predictive Analytics products.
Tidyft is an extension of data.table. It uses modifification by reference whenever possible. This toolkit is designed for big data analysis in high-performance desktop or laptop computers. The syntax of the package is similar or identical to tidyverse.
The CommonMark specification defines a rationalized version of markdown syntax. This package uses the cmark reference implementation for converting markdown text into various formats including HTML, LaTeX and groff man. In addition, it exposes the markdown parse tree in XML format. The latest version of this package also adds support for Github extensions including tables, autolinks and strikethrough text.
This is a package for estimation of one-dimensional probability distributions including kernel density estimation, weighted empirical cumulative distribution functions, Kaplan-Meier and reduced-sample estimators for right-censored data, heat kernels, kernel properties, quantiles and integration.
This package implements nested cross-validation applied to the glmnet and caret packages. With glmnet this includes cross-validation of elastic net alpha parameter. A number of feature selection filter functions (t-test, Wilcoxon test, ANOVA, Pearson/Spearman correlation, random forest, ReliefF) for feature selection are provided and can be embedded within the outer loop of the nested CV. Nested CV can be also be performed with the caret package giving access to the large number of prediction methods available in caret.
This package provides an R interface to the QuickJS portable JavaScript engine. The engine is bundled entirely within the package, requiring no external system dependencies beyond a C compiler.
This package provides a minimal R and C++ API for parsing well-known binary and well-known text representation of geometries to and from R-native formats. Well-known binary is compact and fast to parse; well-known text is human-readable and is useful for writing tests. These formats are only useful in R if the information they contain can be accessed in R, for which high-performance functions are provided here.
Create and manage unique directories for each TensorFlow training run. This package provides a unique, time stamped directory for each run along with functions to retrieve the directory of the latest run or latest several runs.
Set of tools for reading and processing spatial data. The aim is to supply the workflow to create thematic maps. This package also facilitates tmap, the package for visualizing thematic maps.
This package provides tools to export R data as LaTeX and HTML tables.
This package provides a collection of functions to explore and to investigate basic properties of financial returns and related quantities. The covered fields include techniques of explorative data analysis and the investigation of distributional properties, including parameter estimation and hypothesis testing. Even more, there are several utility functions for data handling and management.
This package provides functions and datasets for the book "Modern Applied Statistics with S" (4th edition, 2002) by Venables and Ripley.
This package uses both ridge and lasso penalties (and extensions) to penalize specific parameters in structural equation models. The package offers additional cost functions, cross validation, and other extensions beyond traditional structural equation models. It also contains a function to perform exploratory mediation (XMed).
This package provides tools to infer the code style (which style rules are followed and which ones are not) from one package and use it to check another. This makes it easier to find and correct the most important problems first.
This package provides an R Markdown format for converting an R Markdown document to a grid-oriented dashboard. The dashboard flexibly adapts the size of its components to the containing web page.
This is a package for fast and user-friendly estimation of econometric models with multiple fixed-effects. It includes ordinary least squares (OLS), generalized linear models (GLM) and the negative binomial. The core of the package is based on optimized parallel C++ code, scaling especially well for large data sets. The method to obtain the fixed-effects coefficients is based on Berge (2018). It further provides tools to export and view the results of several estimations with intuitive design to cluster the standard-errors.
This package implements many algorithms for statistical learning on sparse matrices: matrix factorizations, matrix completion, elastic net regressions, factorization machines. The rsparse package also enhances the Matrix package by providing methods for multithreaded <sparse, dense> matrix products and native slicing of the sparse matrices in Compressed Sparse Row (CSR) format.
mlr3 enables efficient, object-oriented programming on the building blocks of machine learning. It provides R6 objects for tasks, learners, resamplings, and measures. The package is geared towards scalability and larger datasets by supporting parallelization and out-of-memory data-backends like databases. While mlr3 focuses on the core computational operations, add-on packages provide additional functionality.