Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
This package provides functions for analyzing citizens bicycle usage pattern and predicting rental amount on specific conditions. Functions on this package interacts with data on tashudata package, a drat repository. tashudata package contains rental/return history on public bicycle system('Tashu'), weather for 3 years and bicycle station information. To install this data package, see the instructions at <https://github.com/zeee1/Tashu_Rpackage>. top10_stations(), top10_paths() function visualizes image showing the most used top 10 stations and paths. daily_bike_rental() and monthly_bike_rental() shows daily, monthly amount of bicycle rental. create_train_dataset(), create_test_dataset() is data processing function for prediction. Bicycle rental history from 2013 to 2014 is used to create training dataset and that on 2015 is for test dataset. Users can make random-forest prediction model by using create_train_model() and predict amount of bicycle rental in 2015 by using predict_bike_rental().
This package provides the means to convert multiqc_data.json files, produced by the wonderful MultiQC tool, into tidy data frames for downstream analysis in R. This analysis might involve cohort analysis, quality control visualisation, change-point detection, statistical process control, clustering, or any other type of quality analysis.
Displays processing time in a clear and structured way. One function supports iterative workflows by predicting and showing the total time required, while another reports the time taken for individual steps within a process.
Fitting models for, and simulation of, trend locally stationary wavelet (TLSW) time series models, which take account of time-varying trend and dependence structure in a univariate time series. The TLSW model, and its estimation, is described in McGonigle, Killick and Nunes (2022a) <doi:10.1111/jtsa.12643>, (2022b) <doi:10.1214/22-EJS2044>. Further information regarding the use of the package, along with detailed examples, can be found in McGonigle, Killick and Nunes (2025) <doi:10.18637/jss.v115.i10>. New users will likely want to start with the TLSW function.
This package provides a Tcl/Tk Graphical User Interface (GUI) to display images than can be zoomed and panned using the mouse and keyboard shortcuts. tkImgR read and write different image formats (PPM/PGM, PNG and GIF) using the standard Tcl/Tk distribution (>=8.6), but other formats (JPEG, TIFF, CR2) can be handled using the tkImg package for Tcl/Tk'.
This package provides a bootstrap test which decides whether two dose response curves can be assumed as equal concerning their maximum absolute deviation. A plenty of choices for the model types are available, which can be found in the DoseFinding package, which is used for the fitting of the models. See <doi:10.1080/01621459.2017.1281813> for details.
Cooperative game theory models decision-making situations in which a group of agents, called players, may achieve certain benefits by cooperating to reach an optimal outcome. It has great potential in different fields, since it offers a scenario to analyze and solve problems in which cooperation is essential to achieve a common goal. The TUGLab (Transferable Utility Games Laboratory) R package contains a set of scripts that could serve as a helpful complement to the books and other materials used in courses on cooperative game theory, and also as a practical tool for researchers working in this field. The TUGLab project was born in 2006 trying to highlight the geometrical aspects of the theory of cooperative games for 3 and 4 players. TUGlabWeb is an online platform on which the basic functions of TUGLab are implemented, and it is being used all over the world as a resource in degree, master's and doctoral programs. This package is an extension of the first versions and enables users to work with games in general (computational restrictions aside). The user can check properties of games, compute well-known games and calculate several set-valued and single-valued solutions such as the core, the Shapley value, the nucleolus or the core-center. The package also illustrates how the Shapley value flexibly adapts to various cooperative game settings, including weighted players and coalitions, a priori unions, and restricted communication structures. In keeping with the original philosophy of the first versions, special emphasis is placed on the graphical representation of the solution concepts for 3 and 4 players.
Implementation of Time to Target plot based on the work of Ribeiro and Rosseti (2015) <DOI:10.1007/s11590-014-0760-8>, that describe a numerical method that gives the probability of an algorithm A finds a solution at least as good as a given target value in smaller computation time than algorithm B.
We provide a tidy grammar of population genetics, facilitating the manipulation and analysis of data on biallelic single nucleotide polymorphisms (SNPs). tidypopgen scales to very large genetic datasets by storing genotypes on disk, and performing operations on them in chunks, without ever loading all data in memory. The full functionalities of the package are described in Carter et al. (2025) <doi:10.1111/2041-210x.70204>.
This package implements the truncated harmonic mean estimator (THAMES) of the reciprocal marginal likelihood using posterior samples and unnormalized log posterior values via reciprocal importance sampling. Metodiev, Perrot-Dockès, Ouadah, Irons, Latouche, & Raftery (2024). Bayesian Analysis. <doi:10.1214/24-BA1422>.
Use SQL SELECT statements to query R data frames.
The textrank algorithm is an extension of the Pagerank algorithm for text. The algorithm allows to summarize text by calculating how sentences are related to one another. This is done by looking at overlapping terminology used in sentences in order to set up links between sentences. The resulting sentence network is next plugged into the Pagerank algorithm which identifies the most important sentences in your text and ranks them. In a similar way textrank can also be used to extract keywords. A word network is constructed by looking if words are following one another. On top of that network the Pagerank algorithm is applied to extract relevant words after which relevant words which are following one another are combined to get keywords. More information can be found in the paper from Mihalcea, Rada & Tarau, Paul (2004) <https://www.aclweb.org/anthology/W04-3252/>.
Provide functions to estimate the coefficients in high-dimensional linear regressions via a tuning-free and robust approach. The method was published in Wang, L., Peng, B., Bradic, J., Li, R. and Wu, Y. (2020), "A Tuning-free Robust and Efficient Approach to High-dimensional Regression", Journal of the American Statistical Association, 115:532, 1700-1714(JASAâ s discussion paper), <doi:10.1080/01621459.2020.1840989>. See also Wang, L., Peng, B., Bradic, J., Li, R. and Wu, Y. (2020), "Rejoinder to â A tuning-free robust and efficient approach to high-dimensional regression". Journal of the American Statistical Association, 115, 1726-1729, <doi:10.1080/01621459.2020.1843865>; Peng, B. and Wang, L. (2015), "An Iterative Coordinate Descent Algorithm for High-Dimensional Nonconvex Penalized Quantile Regression", Journal of Computational and Graphical Statistics, 24:3, 676-694, <doi:10.1080/10618600.2014.913516>; Clémençon, S., Colin, I., and Bellet, A. (2016), "Scaling-up empirical risk minimization: optimization of incomplete u-statistics", The Journal of Machine Learning Research, 17(1):2682â 2717; Fan, J. and Li, R. (2001), "Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties", Journal of the American Statistical Association, 96:456, 1348-1360, <doi:10.1198/016214501753382273>.
Utilities to retrieve and tidy U.S. macroeconomic data series from public government data providers. Functions streamline access to series from the Federal Reserve Bank of St. Louis Federal Reserve Economic Data (FRED), the Bureau of Labor Statistics flat files, and the Bureau of Economic Analysis National Income and Product Accounts tables, then return consistent, tidy data frames ready for modeling and graphics. The package includes helpers for date alignment, log-linear projections, and common macro diagnostics, along with convenience plot builders for quick publication-quality charts.
Fit a threshold regression model for Interval Censored Data based on the first-hitting-time of a boundary by the sample path of a Wiener diffusion process. The threshold regression methodology is well suited to applications involving survival and time-to-event data.
Likelihood ratio and maximum likelihood statistics are provided that can be used as alternatives to p-values Colquhoun (2017) <doi:10.1098/rsos.171085>. Arguments can be either p-values or t-statistics. together with degrees of freedom. For the function tTOlr', the argument twoSided has the default twoSided = TRUE'.
Models the direction of the maximum horizontal stress using relative plate motion parameters. Statistical algorithms to evaluate the modeling results compared with the observed data. Provides plots to visualize the results. Methods described in Stephan et al. (2023) <doi:10.1038/s41598-023-42433-2> and Wdowinski (1998) <doi:10.1016/S0079-1946(98)00091-3>.
Fast calculation of the Subtree Prune and Regraft (SPR), Tree Bisection and Reconnection (TBR) and Replug distances between unrooted trees, using the algorithms of Whidden and Matsen (2017) <doi:10.48550/arXiv.1511.07529>.
Factor and autoregressive models for matrix and tensor valued time series. We provide functions for estimation, simulation and prediction. The models are discussed in Li et al (2021) <doi:10.48550/arXiv.2110.00928>, Chen et al (2020) <DOI:10.1080/01621459.2021.1912757>, Chen et al (2020) <DOI:10.1016/j.jeconom.2020.07.015>, and Xiao et al (2020) <doi:10.48550/arXiv.2006.02611>.
Runs tests using the testthat package but allows for multiple attempts for a single test. This is useful for noisy or flaky tests that generally pass but can fail due to occasional random errors, such as numeric instability or using random data.
This package provides tools for computing various vector summaries of persistence diagrams studied in Topological Data Analysis. For improved computational efficiency, all code for the vector summaries is written in C++ using the Rcpp and RcppArmadillo packages.
This package provides several confidence interval and testing procedures, based on either semiparametric (using event-specific win ratios) or nonparametric measures, including the ratio of integrated cumulative hazard (RICH) and the ratio of integrated transformed cumulative hazard (RITCH), for treatment effect inference with terminal and non-terminal events under competing risks. The semiparametric results were developed in Yang et al. (2022 <doi:10.1002/sim.9266>), and the nonparametric results were developed in Yang (2025 <doi:10.1002/sim.70205>). For comparison, results for the win ratio (Finkelstein and Schoenfeld 1999 <doi:10.1002/(SICI)1097-0258(19990615)18:11%3C1341::AID-SIM129%3E3.0.CO;2-7>), Pocock et al. 2012 <doi:10.1093/eurheartj/ehr352>, and Bebu and Lachin 2016 <doi:10.1093/biostatistics/kxv032>) are included. The package also supports univariate survival analysis with a single event. In this package, effect size estimates and confidence intervals are obtained for each event type, and several testing procedures are implemented for the global null hypothesis of no treatment effect on either terminal or non-terminal events. Furthermore, a test of proportional hazards assumptions, under which the event-specific win ratios converge to hazard ratios, and a test of equal hazard ratios, are provided. For summarizing the treatment effect across all events, confidence intervals for linear combinations of the event-specific win ratios, RICH, or RITCH are available using pre-determined or data-driven weights. Asymptotic properties of these inference procedures are discussed in Yang et al. (2022 <doi:10.1002/sim.9266>) and Yang (2025 <doi:10.1002/sim.70205>).
Accompanies the texts Time Series for Data Science with R by Woodward, Sadler and Robertson & Applied Time Series Analysis with R, 2nd edition by Woodward, Gray, and Elliott. It is helpful for data analysis and for time series instruction.
Facilities to work with vector and raster data in efficient repeatable and systematic work flow. Missing functionality in existing packages is included here to allow extraction from raster data with simple features and Spatial types and to make extraction consistent and straightforward. Extract cell numbers from raster data and return the cells as a data frame rather than as lists of matrices or vectors. The functions here allow spatial data to be used without special handling for the format currently in use.