Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel search send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Offers a graphical user interface for the calculation of the mean measure of divergence, with facilities for trait selection and graphical representations <doi:10.1002/ajpa.23336>.
This package provides a collection of tools for the analysis of habitat selection.
Fit Generalized Additive Models (GAM) using mgcv with parsnip'/'tidymodels via additive <doi:10.5281/zenodo.4784245>. tidymodels is a collection of packages for machine learning; see Kuhn and Wickham (2020) <https://www.tidymodels.org>). The technical details of mgcv are described in Wood (2017) <doi:10.1201/9781315370279>.
This package provides the functions for planning and conducting a clinical trial with adaptive sample size determination. Maximal statistical efficiency will be exploited even when dramatic or multiple adaptations are made. Such a trial consists of adaptive determination of sample size at an interim analysis and implementation of frequentist statistical test at the interim and final analysis with a prefixed significance level. The required assumptions for the stage-wise test statistics are independent and stationary increments and normality. Predetermination of adaptation rule is not required.
The tools in this package are intended to help researchers assess multiple treatment-covariate interactions with data from a parallel-group randomized controlled clinical trial. The methods implemented in the package were proposed in Kovalchik, Varadhan and Weiss (2013) <doi: 10.1002/sim.5881>.
An interactive document on the topic of one-way and two-way analysis of variance using rmarkdown and shiny packages. Runtime examples are provided in the package function as well as at <https://kartikeyab.shinyapps.io/ANOVAShiny/>.
This package implements the differential equations associated to different versions of Allometric Trophic Models (ATN) to estimate the temporal dynamics of species biomasses in food webs. It offers several features to generate synthetic food webs and to parametrise models as well as a wrapper to the ODE solver deSolve.
Collect your data on digital marketing campaigns from Apple Search Ads using the Windsor.ai API <https://windsor.ai/api-fields/>.
This package provides a function that implements the acceptance-rejection method in an optimized manner to generate pseudo-random observations for discrete or continuous random variables. Proposed by von Neumann J. (1951), <https://mcnp.lanl.gov/pdf_files/>, the function is optimized to work in parallel on Unix-based operating systems and performs well on Windows systems. The acceptance-rejection method implemented optimizes the probability of generating observations from the desired random variable, by simply providing the probability function or probability density function, in the discrete and continuous cases, respectively. Implementation is based on references CASELLA, George at al. (2004) <https://www.jstor.org/stable/4356322>, NEAL, Radford M. (2003) <https://www.jstor.org/stable/3448413> and Bishop, Christopher M. (2006, ISBN: 978-0387310732).
This package provides a toolkit for archaeological time series and time intervals. This package provides a system of classes and methods to represent and work with archaeological time series and time intervals. Dates are represented as "rata die" and can be converted to (virtually) any calendar defined by Reingold and Dershowitz (2018) <doi:10.1017/9781107415058>. This packages offers a simple API that can be used by other specialized packages.
Automatic normalisation of a data frame to third normal form, with the intention of easing the process of data cleaning. (Usage to design your actual database for you is not advised.) Originally inspired by the AutoNormalize library for Python by Alteryx (<https://github.com/alteryx/autonormalize>), with various changes and improvements. Automatic discovery of functional or approximate dependencies, normalisation based on those, and plotting of the resulting "database" via Graphviz', with options to exclude some attributes at discovery time, or remove discovered dependencies at normalisation time.
Simple animated versions of basic R plots, using the animation package. Includes animated versions of plot, barplot, persp, contour, filled.contour, hist, curve, points, lines, text, symbols, segments, and arrows.
This package creates interactive Venn diagrams using the amCharts5 library for JavaScript'. They can be used directly from the R console, from RStudio', in shiny applications, and in rmarkdown documents.
This package provides a modeling package compiling applicability domain methods in R. It combines different methods to measure the amount of extrapolation new samples can have from the training set. See <doi:10.4018/IJQSPR.2016010102> for an overview of applicability domains.
Allows access to selected services that are part of the Google Adwords API <https://developers.google.com/adwords/api/docs/guides/start>. Google Adwords is an online advertising service by Google', that delivers Ads to users. This package offers a authentication process using OAUTH2'. Currently, there are two methods of data of accessing the API, depending on the type of request. One method uses SOAP requests which require building an XML structure and then sent to the API. These are used for the ManagedCustomerService and the TargetingIdeaService'. The second method is by building AWQL queries for the reporting side of the Google Adwords API.
The AHP method (Analytic Hierarchy Process) is a multi-criteria decision-making method addressing choice and outranking problems. The method enables to perform the analysis of alternatives in each type of criterion and then provides a global performance of each alternative in the decision context. The main difference of this package is the possibility of evaluating the alternatives using quantitative data, by numerical representation, and qualitative data, using the Saaty scale, providing preference relation between variables by a pairwise evaluation.
Perform parallel factor analysis (PARAFAC: Hitchcock, 1927) <doi:10.1002/sapm192761164> on fluorescence excitation-emission matrices: handle scattering signal and inner filter effect, scale the dataset, fit the model; perform split-half validation or jack-knifing. Modified approaches such as Whittaker interpolation, randomised split-half, and fluorescence and scattering model estimation are also available. The package has a low dependency footprint and has been tested on a wide range of R versions.
This package provides a systematic framework for neural networkâ based model selection and forecasting using single hidden layer feed-forward networks. It evaluates all possible combinations of predictor variables and hidden layer configurations, selecting the optimal model based on predictive accuracy criteria such as root mean squared error (RMSE) and mean absolute percentage error (MAPE). Predictors are automatically standardized, and model performance is assessed using out-of-sample validation. The package is designed for empirical modelling and forecasting in economics, agriculture, trade, climate, and related applied research domains where nonlinear relationships and robust predictive performance are of primary interest.
This package implements techniques to estimate the unknown quantities related to two-component admixture models, where the two components can belong to any distribution (note that in the case of multinomial mixtures, the two components must belong to the same family). Estimation methods depend on the assumptions made on the unknown component density; see Bordes and Vandekerkhove (2010) <doi:10.3103/S1066530710010023>, Patra and Sen (2016) <doi:10.1111/rssb.12148>, and Milhaud, Pommeret, Salhi, Vandekerkhove (2024) <doi:10.3150/23-BEJ1593>. In practice, one can estimate both the mixture weight and the unknown component density in a wide variety of frameworks. On top of that, hypothesis tests can be performed in one and two-sample contexts to test the unknown component density (see Milhaud, Pommeret, Salhi and Vandekerkhove (2022) <doi:10.1016/j.jspi.2021.05.010>, and Milhaud, Pommeret, Salhi, Vandekerkhove (2024) <doi:10.3150/23-BEJ1593>). Finally, clustering of unknown mixture components is also feasible in a K-sample setting (see Milhaud, Pommeret, Salhi, Vandekerkhove (2024) <https://jmlr.org/papers/v25/23-0914.html>).
Simulation and estimation tools for various types of ambit processes, including trawl processes and weighted trawl processes.
Fast processing of ArcGIS FeatureCollection protocol buffers in R. It is designed to work seamlessly with httr2 and integrates with sf'.
Helper functions for working with Regional Ocean Modeling System ROMS output. See <https://www.myroms.org/> for more information about ROMS'.
Interactive R tutorials written using learnr for Field (2016), "An Adventure in Statistics", <ISBN:9781446210451>. Topics include general workflow in R and Rstudio', the R environment and tidyverse', summarizing data, model fitting, central tendency, visualising data using ggplot2', inferential statistics and robust estimation, hypothesis testing, the general linear model, comparing means, repeated measures designs, factorial designs, multilevel models, growth models, and generalized linear models (logistic regression).
Automatic fixed rank kriging for (irregularly located) spatial data using a class of basis functions with multi-resolution features and ordered in terms of their resolutions. The model parameters are estimated by maximum likelihood (ML) and the number of basis functions is determined by Akaike's information criterion (AIC). For spatial data with either one realization or independent replicates, the ML estimates and AIC are efficiently computed using their closed-form expressions when no missing value occurs. Details regarding the basis function construction, parameter estimation, and AIC calculation can be found in Tzeng and Huang (2018) <doi:10.1080/00401706.2017.1345701>. For data with missing values, the ML estimates are obtained using the expectation- maximization algorithm. Apart from the number of basis functions, there are no other tuning parameters, making the method fully automatic. Users can also include a stationary structure in the spatial covariance, which utilizes LatticeKrig package.