Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
This package provides data sets and functions for exploration of Pakistan Population Census 2023 (<https://www.pbs.gov.pk/>).
Early generation breeding trials are to be conducted in multiple environments where it may not be possible to replicate all the lines in each environment due to scarcity of resources. For such situations, partially replicated (p-Rep) designs have wide application potential as only a proportion of the test lines are replicated at each environment. A collection of several utility functions related to p-Rep designs have been developed. Here, the package contains six functions for a complete stepwise analytical study of these designs. Five functions pRep1(), pRep2(), pRep3(), pRep4() and pRep5(), are used to generate five new series of p-Rep designs and also compute average variance factors and canonical efficiency factors of generated designs. A fourth function NCEV() is used to generate incidence matrix (N), information matrix (C), canonical efficiency factor (E) and average variance factor (V). This function is general in nature and can be used for studying the characterization properties of any block design. A construction procedure for p-Rep designs was given by Williams et al.(2011) <doi:10.1002/bimj.201000102> which was tedious and time consuming. Here, in this package, five different methods have been given to generate p-Rep designs easily.
Visualize the partitions of simple decision trees, involving one or two predictors, on the scale of the original data. Provides an intuitive alternative to traditional tree diagrams, by visualizing how a decision tree divides the predictor space in a simple 2D plot alongside the original data. The parttree package supports both classification and regression trees from rpart and partykit', as well as trees produced by popular frontend systems like tidymodels and mlr3'. Visualization methods are provided for both base R graphics and ggplot2'.
This package provides a comprehensive collection of tools for creating, manipulating and visualising pedigrees and genetic marker data. Pedigrees can be read from text files or created on the fly with built-in functions. A range of utilities enable modifications like adding or removing individuals, breaking loops, and merging pedigrees. An online tool for creating pedigrees interactively, based on pedtools', is available at <https://magnusdv.shinyapps.io/quickped>. pedtools is the hub of the pedsuite', a collection of packages for pedigree analysis. A detailed presentation of the pedsuite is given in the book Pedigree Analysis in R (Vigeland, 2021, ISBN:9780128244302).
This package provides tools to print a compact, readable directory tree for a folder or project. The package can automatically detect common project roots (e.g., RStudio .Rproj files) and formats output for quick inspection of code and data organization. It supports typical tree customizations such as limiting depth, excluding files using ignore patterns, and producing clean, aligned text output suitable for console use, reports, and reproducible documentation. A snapshot helper can also render the tree output to a PNG image for sharing in issues, teaching material, or project documentation.
Reconstruct pedigrees from genotype data, by optimising the likelihood over all possible pedigrees subject to given restrictions. Tailor-made plots facilitate evaluation of the output. This package is part of the pedsuite ecosystem for pedigree analysis. In particular, it imports pedprobr for calculating pedigree likelihoods and forrel for estimating pairwise relatedness.
This package provides a suite of diagnostic tools for univariate point processes. This includes tools for simulating and fitting both common and more complex temporal point processes. We also include functions to visualise these point processes and collect existing diagnostic tools of Brown et al. (2002) <doi:10.1162/08997660252741149> and Wu et al. (2021) <doi:10.1002/9781119821588.ch7>, which can be used to assess the fit of a chosen point process model.
Most of the time floating point arithmetic does approximately the right thing. When adding sums or having products of numbers that greatly differ in magnitude, the floating point arithmetic may be incorrect. This package implements the Kahan (1965) sum <doi:10.1145/363707.363723>, Neumaier (1974) sum <doi:10.1002/zamm.19740540106>, pairwise-sum (adapted from NumPy', See Castaldo (2008) <doi:10.1137/070679946> for a discussion of accuracy), and arbitrary precision sum (adapted from the fsum in Python ; Shewchuk (1997) <https://people.eecs.berkeley.edu/~jrs/papers/robustr.pdf>). In addition, products are changed to long double precision for accuracy, or changed into a log-sum for accuracy.
Classification based analysis of DNA sequences to taxonomic groupings. This package primarily implements Naive Bayesian Classifier from the Ribosomal Database Project. This approach has traditionally been used to classify 16S rRNA gene sequences to bacterial taxonomic outlines; however, it can be used for any type of gene sequence. The method was originally described by Wang, Garrity, Tiedje, and Cole in Applied and Environmental Microbiology 73(16):5261-7 <doi:10.1128/AEM.00062-07>. The package also provides functions to read in FASTA'-formatted sequence data.
Collection of functions to get files in parquet format. Parquet is a columnar storage file format <https://parquet.apache.org/>. The files to convert can be of several formats ("csv", "RData", "rds", "RSQLite", "json", "ndjson", "SAS", "SPSS"...).
Provide estimation for particular cases of the power series cure rate model <doi:10.1080/03610918.2011.639971>. For the distribution of the concurrent causes the alternative models are the Poisson, logarithmic, negative binomial and Bernoulli (which are includes in the original work), the polylogarithm model <doi:10.1080/00949655.2018.1451850> and the Flory-Schulz <doi:10.3390/math10244643>. The estimation procedure is based on the EM algorithm discussed in <doi:10.1080/03610918.2016.1202276>. For the distribution of the time-to-event the alternative models are slash half-normal, Weibull, gamma and Birnbaum-Saunders distributions.
This package provides functions for graph-based multiple-sample testing and visualization of microbiome data, in particular data stored in phyloseq objects. The tests are based on those described in Friedman and Rafsky (1979) <http://www.jstor.org/stable/2958919>, and the tests are described in more detail in Callahan et al. (2016) <doi:10.12688/f1000research.8986.1>.
This package provides some easy-to-use functions for spatial analyses of (plant-) phenological data sets and satellite observations of vegetation.
This package provides functions for constructing dashboards for business process monitoring. Building on the event log objects class from package bupaR'. Allows the use to assemble custom shiny dashboards based on process data.
Miscellaneous small utilities are provided to mitigate issues with messy, inconsistent or high dimensional data and help for preprocessing and preparing analyses.
Package for learning and evaluating (subgroup) policies via doubly robust loss functions. Policy learning methods include doubly robust blip/conditional average treatment effect learning and sequential policy tree learning. Methods for (subgroup) policy evaluation include doubly robust cross-fitting and online estimation/sequential validation. See Nordland and Holst (2022) <doi:10.48550/arXiv.2212.02335> for documentation and references.
This package provides functions and mined database from UniProt focusing on post-translational modifications to do single enrichment analysis (SEA) and protein set enrichment analysis (PSEA). Payman Nickchi, Uladzislau Vadadokhau, Mehdi Mirzaie, Marc Baumann, Amir Ata Saei, Mohieddin Jafari (2025) <doi:10.1002/pmic.202400238>.
An implementation of a non-parametric statistical model using a parallelised Monte Carlo sampling scheme. The method implemented in this package allows non-parametric inference to be regularized for small sample sizes, while also being more accurate than approximations such as variational Bayes. The concentration parameter is an effective sample size parameter, determining the faith we have in the model versus the data. When the concentration is low, the samples are close to the exact Bayesian logistic regression method; when the concentration is high, the samples are close to the simplified variational Bayes logistic regression. The method is described in full in the paper Lyddon, Walker, and Holmes (2018), "Nonparametric learning from Bayesian models with randomized objective functions" <arXiv:1806.11544>.
Smoothing splines with penalties on order m derivatives.
Conduct power analyses and inference of marginal effects. Uses plug-in estimation and influence functions to perform robust inference, optionally leveraging historical data to increase precision with prognostic covariate adjustment. The methods are described in Højbjerre-Frandsen et al. (2025) <doi:10.48550/arXiv.2503.22284>.
This package provides functions for fitting abundance distributions over environmental gradients to the species in ecological communities, and tools for simulating the fossil assemblages from those abundance models for such communities, as well as simulating assemblages across various patterns of sedimentary history and sampling. These tools are for particular use with fossil records with detailed age models and abundance distributions used for calculating environmental gradients from ordinations or other indices based on fossil assemblages.
Presentation of a new goodness-of-fit normality test based on the Lilliefors method. For details on this method see: Sulewski (2019) <doi:10.1080/03610918.2019.1664580>.
Reviews other packages during code review by looking at their dependencies, code style, code complexity, and how internally defined functions interact with one another.
Calculations of an information criterion are proposed to check the quality of simulations results of Agent-based models (ABM/IBM) or other non-linear rule-based models. The POMDEV measure (Pattern Oriented Modelling DEViance) is based on the Kullback-Leibler divergence and likelihood theory. It basically indicates the deviance of simulation results from field observations. Once POMDEV scores and metropolis-hasting sampling on different model versions are effectuated, POMIC scores (Pattern Oriented Modelling Information Criterion) can be calculated. This method could be further developed to incorporate multiple patterns assessment. Piou C, U Berger and V Grimm (2009) <doi:10.1016/j.ecolmodel.2009.05.003>.