Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel search send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Named after the Irish name for weather, this package contains tidied data from the Irish Meteorological Service's hourly observations for 2017. In all, the data sets include observations from 25 weather stations, and also latitude and longitude coordinates for each weather station. Now includes energy generation data for Ireland and Northern Ireland (2017), including Wind Generation data.
This package implements adaptive gPCA, as described in: Fukuyama, J. (2017) <arXiv:1702.00501>. The package also includes functionality for applying the method to phyloseq objects so that the method can be easily applied to microbiome data and a shiny app for interactive visualization.
Wraps the Ace editor in a HTML widget. The Ace editor has support for many languages. It can be opened in the viewer pane of RStudio', and this provides a second source editor.
This package provides a set of fast and convenient functions to help conducting accessibility analyses. Given a pre-computed travel cost matrix and a land use dataset (containing the location of jobs, healthcare and population, for example), the package allows one to calculate accessibility levels and accessibility poverty and inequality. The package covers the majority of the most commonly used accessibility measures (such as cumulative opportunities, gravity-based and floating catchment areas methods), as well as the most frequently used inequality and poverty metrics (such as the Palma ratio, the concentration and Theil indices and the FGT family of measures).
Survival analysis is employed to model the time it takes for events to occur. Survival model examines the relationship between survival and one or more predictors, usually termed covariates in the survival-analysis literature. To this end, Cox-proportional (Cox-PH) hazard rate model introduced in a seminal paper by Cox (1972) <doi:10.1111/j.2517-6161.1972.tb00899.x>, is a broadly applicable and the most widely used method of survival analysis. This package can be used to estimate the effect of fixed and time-dependent covariates and also to compute the survival probabilities of the lactation of dairy animal. This package has been developed using algorithm of Klein and Moeschberger (2003) <doi:10.1007/b97377>.
Estimate and plot confounder-adjusted survival curves using either Direct Adjustment', Direct Adjustment with Pseudo-Values', various forms of Inverse Probability of Treatment Weighting', two forms of Augmented Inverse Probability of Treatment Weighting', Empirical Likelihood Estimation or Targeted Maximum Likelihood Estimation'. Also includes a significance test for the difference between two adjusted survival curves and the calculation of adjusted restricted mean survival times. Additionally enables the user to estimate and plot cause-specific confounder-adjusted cumulative incidence functions in the competing risks setting using the same methods (with some exceptions). For details, see Denz et. al (2023) <doi:10.1002/sim.9681>.
You can use this package to create custom pipeline badges in a standard svg format. This is useful for a company to use internally, where it may not be possible to create badges through external providers. This project was inspired by the anybadge library in python.
The ggarrow package is a ggplot2 extension that plots a variety of different arrow segments with many options to customize. The arrowheadr package makes it easy to create custom arrowheads and fins within the parameters that ggarrow functions expect. It has preset arrowheads and a collection of functions to create and transform data for customizing arrows.
The normal process of creating clinical study slides is that a statistician manually type in the numbers from outputs and a separate statistician to double check the typed in numbers. This process is time consuming, resource intensive, and error prone. Automatic slide generation is a solution to address these issues. It reduces the amount of work and the required time when creating slides, and reduces the risk of errors from manually typing or copying numbers from the output to slides. It also helps users to avoid unnecessary stress when creating large amounts of slide decks in a short time window.
Provides: (1) Tools to infer dominance hierarchies based on calculating Elo scores, but with custom functions to improve estimates in animals with relatively stable dominance ranks. (2) Tools to plot the shape of the dominance hierarchy and estimate the uncertainty of a given data set.
The maximum likelihood estimator (MLE) is a technology: under regularity conditions, any MLE is asymptotically normal with variance given by the inverse Fisher information. This package exploits that structure by defining an algebra over MLEs. Compose independent estimators into joint MLEs via block-diagonal covariance ('joint'), optimally combine repeated estimates via inverse-variance weighting ('combine'), propagate transformations via the delta method ('rmap'), and bridge to distribution algebra via conversion to normal or multivariate normal objects ('as_dist'). Supports asymptotic ('mle', mle_numerical') and bootstrap ('mle_boot') estimators with a unified interface for inference: confidence intervals, standard errors, AIC, Fisher information, and predictive intervals. For background on maximum likelihood estimation, see Casella and Berger (2002, ISBN:978-0534243128). For the delta method and variance estimation, see Lehmann and Casella (1998, ISBN:978-0387985022).
Estimate ideal efficiencies of aerosol sampling through sample lines. Functions were developed consistent with the approach described in Hogue, Mark; Thompson, Martha; Farfan, Eduardo; Hadlock, Dennis, (2014), "Hand Calculations for Transport of Radioactive Aerosols through Sampling Systems" Health Phys 106, 5, S78-S87, <doi:10.1097/HP.0000000000000092>.
Developed for Computing the probability density function, cumulative distribution function, random generation, estimating the parameters of asymmetric exponential power distribution, and robust regression analysis with error term that follows asymmetric exponential power distribution. The asymmetric exponential power distribution studied here is a special case of that introduced by Dongming and Zinde-Walsh (2009) <doi:10.1016/j.jeconom.2008.09.038>.
Checks function arguments, ideally for use in R packages. Uses a simple interface and produces clean, informative error messages using cli'.
Aids the programming of Clinical Data Standards Interchange Consortium (CDISC) compliant Ophthalmology Analysis Data Model (ADaM) datasets in R. ADaM datasets are a mandatory part of any New Drug or Biologics License Application submitted to the United States Food and Drug Administration (FDA). Analysis derivations are implemented in accordance with the "Analysis Data Model Implementation Guide" (CDISC Analysis Data Model Team, 2021, <https://www.cdisc.org/standards/foundational/adam/adamig-v1-3-release-package>).
Presents a series of molecular and genetic routines in the R environment with the aim of assisting in analytical pipelines before and after the use of asreml or another library to perform analyses such as Genomic Selection or Genome-Wide Association Analyses. Methods and examples are described in Gezan, Oliveira, Galli, and Murray (2022) <https://asreml.kb.vsni.co.uk/wp-content/uploads/sites/3/ASRgenomics_Manual.pdf>.
Scraping content from archived web pages stored in the Internet Archive (<https://archive.org>) using a systematic workflow. Get an overview of the mementos available from the respective homepage, retrieve the Urls and links of the page and finally scrape the content. The final output is stored in tibbles, which can be then easily used for further analysis.
This toolkit implements a numerical solution algorithm to invert a quality of life measure from observed data. Unlike the traditional Rosen-Roback measure, this measure accounts for mobility frictionsâ generated by idiosyncratic tastes and local ties â and trade frictions â generated by trade costs and non-tradable services, thereby reducing non-classical measurement error. The QoL measure is based on Ahlfeldt, Bald, Roth, Seidel (2024) <https://econpapers.repec.org/RePEc:boc:bocode:s459382> "Measuring Quality of Life under Spatial Frictions". When using this programme or the toolkit in your work, please cite the paper.
This package provides a tidy text corpus of Aesop's Fables sourced from the Library of Congress, along with analysis-ready datasets for sentiment, emotion, and linguistic analysis of moral storytelling. The package includes both full narrative texts and word-level representations to support exploratory text analysis and teaching workflows.
This software solves an Advection Bi-Flux Diffusive Problem using the Finite Difference Method FDM. Vasconcellos, J.F.V., Marinho, G.M., Zanni, J.H., 2016, Numerical analysis of an anomalous diffusion with a bimodal flux distribution. <doi:10.1016/j.rimni.2016.05.001>. Silva, L.G., Knupp, D.C., Bevilacqua, L., Galeao, A.C.N.R., Silva Neto, A.J., 2014, Formulation and solution of an Inverse Anomalous Diffusion Problem with Stochastic Techniques. <doi:10.5902/2179460X13184>. In this version, it is possible to include a source as a function depending on space and time, that is, s(x,t).
This package provides the functions for planning and conducting a clinical trial with adaptive sample size determination. Maximal statistical efficiency will be exploited even when dramatic or multiple adaptations are made. Such a trial consists of adaptive determination of sample size at an interim analysis and implementation of frequentist statistical test at the interim and final analysis with a prefixed significance level. The required assumptions for the stage-wise test statistics are independent and stationary increments and normality. Predetermination of adaptation rule is not required.
Loss reserving generally focuses on identifying a single model that can generate superior predictive performance. However, different loss reserving models specialise in capturing different aspects of loss data. This is recognised in practice in the sense that results from different models are often considered, and sometimes combined. For instance, actuaries may take a weighted average of the prediction outcomes from various loss reserving models, often based on subjective assessments. This package allows for the use of a systematic framework to objectively combine (i.e. ensemble) multiple stochastic loss reserving models such that the strengths offered by different models can be utilised effectively. Our framework is developed in Avanzi et al. (2023). Firstly, our criteria model combination considers the full distributional properties of the ensemble and not just the central estimate - which is of particular importance in the reserving context. Secondly, our framework is that it is tailored for the features inherent to reserving data. These include, for instance, accident, development, calendar, and claim maturity effects. Crucially, the relative importance and scarcity of data across accident periods renders the problem distinct from the traditional ensemble techniques in statistical learning. Our framework is illustrated with a complex synthetic dataset. In the results, the optimised ensemble outperforms both (i) traditional model selection strategies, and (ii) an equally weighted ensemble. In particular, the improvement occurs not only with central estimates but also relevant quantiles, such as the 75th percentile of reserves (typically of interest to both insurers and regulators). Reference: Avanzi B, Li Y, Wong B, Xian A (2023) "Ensemble distributional forecasting for insurance loss reserving" <doi:10.48550/arXiv.2206.08541>.
An interface for performing all stages of ADMIXTOOLS analyses (<https://github.com/dreichlab/admixtools>) entirely from R. Wrapper functions (D, f4, f3, etc.) completely automate the generation of intermediate configuration files, run ADMIXTOOLS programs on the command-line, and parse output files to extract values of interest. This allows users to focus on the analysis itself instead of worrying about low-level technical details. A set of complementary functions for processing and filtering of data in the EIGENSTRAT format is also provided.
Solves the problem of identifying the densest submatrix in a given or sampled binary matrix, Bombina et al. (2019) <arXiv:1904.03272>.