Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Extends the test-based Bayes factor (TBF) methodology to multinomial regression models and discrete time-to-event models with competing risks. The TBF methodology has been well developed and implemented for the generalised linear model [Held et al. (2015) <doi:10.1214/14-STS510>] and for the Cox model [Held et al. (2016) <doi:10.1002/sim.7089>].
Interface to the API for TreeBASE <http://treebase.org> from R. TreeBASE is a repository of user-submitted phylogenetic trees (of species, population, or genes) and the data used to create them.
Suite of tropical geometric tools for use in machine learning applications. These methods may be summarized in the following references: Yoshida, et al. (2022) <doi:10.2140/astat.2023.14.37>, Barnhill et al. (2023) <doi:10.48550/arXiv.2303.02539>, Barnhill and Yoshida (2023) <doi:10.3390/math11153433>, Aliatimis et al. (2023) <doi:10.1007/s11538-024-01327-8>, Yoshida et al. (2022) <doi:10.1109/TCBB.2024.3420815>, and Yoshida et al. (2019) <doi:10.1007/s11538-018-0493-4>.
This package implements two tests for same-source of toolmarks. The chumbley_non_random() test follows the paper "An Improved Version of a Tool Mark Comparison Algorithm" by Hadler and Morris (2017) <doi:10.1111/1556-4029.13640>. This is an extension of the Chumbley score as previously described in "Validation of Tool Mark Comparisons Obtained Using a Quantitative, Comparative, Statistical Algorithm" by Chumbley et al (2010) <doi:10.1111/j.1556-4029.2010.01424.x>. fixed_width_no_modeling() is based on correlation measures in a diamond shaped area of the toolmark as described in Hadler (2017).
Forecasting competitions are of increasing importance as a mean to learn best practices and gain knowledge. Data leakage is one of the most common issues that can often be found in competitions. Data leaks can happen when the training data contains information about the test data. For example: randomly chosen blocks of time series are concatenated to form a new time series, scale-shifts, repeating patterns in time series, white noise is added in the original time series to form a new time series, etc. tsdataleaks package can be used to detect data leakages in a collection of time series.
Demonstration functions that can be used in a classroom to demonstrate statistical concepts, or on your own to better understand the concepts or the programming.
Interface to TensorFlow IO', Datasets and filesystem extensions maintained by `TensorFlow SIG-IO` <https://github.com/tensorflow/community/blob/master/sigs/io/CHARTER.md>.
An implementation of the time-series Susceptible-Infected-Recovered (TSIR) model using a number of different fitting options for infectious disease time series data. The manuscript based on this package can be found here <doi:10.1371/journal.pone.0185528>. The method implemented here is described by Finkenstadt and Grenfell (2000) <doi:10.1111/1467-9876.00187>.
Useful functions to connect to TM1 <https://www.ibm.com/uk-en/products/planning-and-analytics> instance from R via REST API. With the functions in the package, data can be imported from TM1 via mdx view or native view, data can be sent to TM1', processes and chores can be executed, and cube and dimension metadata information can be taken.
Simple toolkit for working with TOML text. Based on tomledit which allows for modifying TOML while preserving order, comments,and whitespace.
These functions generate data frames on troop deployments and military basing using U.S. Department of Defense data on overseas military deployments. This package provides functions for pulling country-year troop deployment and basing data. Subsequent versions will hopefully include cross-national data on deploying countries.
Helper functions for MASCOTNUM / RT-UQ <https://uq.math.cnrs.fr/> algorithm template, for design of numerical experiments practice: algorithm template parser to support MASCOTNUM specification <https://github.com/MASCOTNUM/algorithms>, ask & tell decoupling injection (inspired by <https://search.r-project.org/CRAN/refmans/sensitivity/html/decoupling.html>) to use "crimped" algorithms (like uniroot(), optim(), ...) from outside R, basic template examples: Brent algorithm for 1 dim root finding and L-BFGS-B from base optim().
Interface to TensorFlow Datasets, a high-level library for building complex input pipelines from simple, re-usable pieces. See <https://www.tensorflow.org/guide> for additional details.
Pacote para a analise de experimentos com um ou dois fatores com testemunhas adicionais conduzidos no delineamento inteiramente casualizado ou em blocos casualizados. "Package for the analysis of one or two-way experiments with additional controls conducted in a completely randomized design or in a randomized block design".
This package performs Three-Mode Principal Components Analysis, which carries out Tucker Models.
This package provides tools to perform multiple comparison analyses, based on the well-known Tukey's "Honestly Significant Difference" (HSD) test. In models involving interactions, TukeyC stands out from other R packages by implementing intuitive and easy-to-use functions. In addition to accommodating traditional R methods such as lm() and aov(), it has also been extended to objects of the lmer() class, that is, mixed models with fixed effects. For more details see Tukey (1949) <doi:10.2307/3001913>.
For when your colors absolutely should not be excluded from the narrative.
Allows users to quickly load multiple patients electrocardiographic (ECG) data at once and conduct relevant time analysis of heart rate variability (HRV) without manual edits from a physician or data cleaning specialist. The package provides the unique ability to iteratively filter, plot, and store time analysis results in a data frame while writing plots to a predefined folder. This streamlines the workflow for HRV analysis across multiple datasets. Methods are based on Rodrà guez-Liñares et al. (2011) <doi:10.1016/j.cmpb.2010.05.012>. Examples of applications using this package include Kwon et al. (2022) <doi:10.1007/s10286-022-00865-2> and Lawrence et al. (2023) <doi:10.1016/j.autneu.2022.103056>.
Compile snippets of LaTeX directly into images from the R console to view in the RStudio viewer pane, Shiny apps and RMarkdown documents.
Programs for Martinussen and Scheike (2006), `Dynamic Regression Models for Survival Data', Springer Verlag. Plus more recent developments. Additive survival model, semiparametric proportional odds model, fast cumulative residuals, excess risk models and more. Flexible competing risks regression including GOF-tests. Two-stage frailty modelling. PLS for the additive risk model. Lasso in the ahaz package.
This package provides tools for constructing and analyzing two-phase experimental designs under correlated error structures. Version 1.1.1 includes improved efficiency factor classification with tolerance control, updated plot visualizations, and improved clarity of the results. The conceptual framework and the term two-phase were introduced by McIntyre (1955) <doi:10.2307/3001770>).
This package provides a tm Source to create corpora from articles exported from the LexisNexis content provider as HTML files. It is able to read both text content and meta-data information (including source, date, title, author and pages). Note that the file format is highly unstable: there is no warranty that this package will work for your corpus, and you may have to adjust the code to adapt it to your particular format.
Create interactive tables, calendars, charts and markdown WYSIWYG editor with TOAST UI <https://ui.toast.com/> libraries to integrate in shiny applications or rmarkdown HTML documents.
Multinomial (inverse) regression inference for text documents and associated attributes. For details see: Taddy (2013 JASA) Multinomial Inverse Regression for Text Analysis <arXiv:1012.2098> and Taddy (2015, AoAS), Distributed Multinomial Regression, <arXiv:1311.6139>. A minimalist partial least squares routine is also included. Note that the topic modeling capability of earlier textir is now a separate package, maptpx'.