Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
This package creates plots showing scored HR experiments and plots of distribution of means of ranks of HR score from bootstrapping. Authors (2019) <doi:10.5281/zenodo.3374507>.
This package provides a collection of functions to analyse, visualize and interpret wind data and to calculate the potential energy production of wind turbines.
In p >> n settings, full posterior sampling using existing Markov chain Monte Carlo (MCMC) algorithms is highly inefficient and often not feasible from a practical perspective. To overcome this problem, we propose a scalable stochastic search algorithm that is called the Simplified Shotgun Stochastic Search (S5) and aimed at rapidly explore interesting regions of model space and finding the maximum a posteriori(MAP) model. Also, the S5 provides an approximation of posterior probability of each model (including the marginal inclusion probabilities). This algorithm is a part of an article titled "Scalable Bayesian Variable Selection Using Nonlocal Prior Densities in Ultrahigh-dimensional Settings" (2018) by Minsuk Shin, Anirban Bhattacharya, and Valen E. Johnson and "Nonlocal Functional Priors for Nonparametric Hypothesis Testing and High-dimensional Model Selection" (2020+) by Minsuk Shin and Anirban Bhattacharya.
Bit-level reading and writing are necessary when dealing with many file formats e.g. compressed data and binary files. Currently, R connections are manipulated at the byte level. This package wraps existing connections and raw vectors so that it is possible to read bits, bit sequences, unaligned bytes and low-bit representations of integers.
An improved multiple testing procedure for controlling false discovery rates which is developed based on the Bonferroni procedure with integrated estimates from the Benjamini-Hochberg procedure and the Storey's q-value procedure. It controls false discovery rates through controlling the expected number of false discoveries.
Code for backShift', an algorithm to estimate the connectivity matrix of a directed (possibly cyclic) graph with hidden variables. The underlying system is required to be linear and we assume that observations under different shift interventions are available. For more details, see <arXiv:1506.02494>.
This package provides tools for constructing board/grid based games, as well as readily available game(s) for your entertainment.
This package provides tools to model and forecast multivariate time series including Bayesian Vector heterogeneous autoregressive (VHAR) model by Kim & Baek (2023) (<doi:10.1080/00949655.2023.2281644>). bvhar can model Vector Autoregressive (VAR), VHAR, Bayesian VAR (BVAR), and Bayesian VHAR (BVHAR) models.
Collect your data on digital marketing campaigns from bing Ads using the Windsor.ai API <https://windsor.ai/api-fields/>.
Toolkit for Bayesian estimation of the dependence structure in multivariate extreme value parametric models, following Sabourin and Naveau (2014) <doi:10.1016/j.csda.2013.04.021> and Sabourin, Naveau and Fougeres (2013) <doi:10.1007/s10687-012-0163-0>.
Provide a tool to easily build customized data flows to pre-process large volumes of information from different sources. To this end, bdpar allows to (i) easily use and create new functionalities and (ii) develop new data source extractors according to the user needs. Additionally, the package provides by default a predefined data flow to extract and pre-process the most relevant information (tokens, dates, ... ) from some textual sources (SMS, Email, YouTube comments).
Package provides functions for estimation and inference in Bayesian quantile regression with ordinal outcomes. An ordinal model with 3 or more outcomes (labeled OR1 model) is estimated by a combination of Gibbs sampling and Metropolis-Hastings (MH) algorithm. Whereas an ordinal model with exactly 3 outcomes (labeled OR2 model) is estimated using a Gibbs sampling algorithm. The summary output presents the posterior mean, posterior standard deviation, 95% credible intervals, and the inefficiency factors along with the two model comparison measures â logarithm of marginal likelihood and the deviance information criterion (DIC). The package also provides functions for computing the covariate effects and other functions that aids either the estimation or inference in quantile ordinal models. Rahman, M. A. (2016).â Bayesian Quantile Regression for Ordinal Models.â Bayesian Analysis, 11(1): 1-24 <doi: 10.1214/15-BA939>. Yu, K., and Moyeed, R. A. (2001). â Bayesian Quantile Regression.â Statistics and Probability Letters, 54(4): 437â 447 <doi: 10.1016/S0167-7152(01)00124-9>. Koenker, R., and Bassett, G. (1978).â Regression Quantiles.â Econometrica, 46(1): 33-50 <doi: 10.2307/1913643>. Chib, S. (1995). â Marginal likelihood from the Gibbs output.â Journal of the American Statistical Association, 90(432):1313â 1321, 1995. <doi: 10.1080/01621459.1995.10476635>. Chib, S., and Jeliazkov, I. (2001). â Marginal likelihood from the Metropolis-Hastings output.â Journal of the American Statistical Association, 96(453):270â 281, 2001. <doi: 10.1198/016214501750332848>.
This package provides a method to filter correlation and covariance matrices by averaging bootstrapped filtered hierarchical clustering and boosting. See Ch. Bongiorno and D. Challet, Covariance matrix filtering with bootstrapped hierarchies (2020) <arXiv:2003.05807> and Ch. Bongiorno and D. Challet, Reactive Global Minimum Variance Portfolios with k-BAHC covariance cleaning (2020) <arXiv:2005.08703>.
This package implements variable selection for high dimensional datasets with a binary response variable using the EM algorithm. Both probit and logit models are supported. Also included is a useful function to generate high dimensional data with correlated variables.
Network meta-analyses using Bayesian framework following Dias et al. (2013) <DOI:10.1177/0272989X12458724>. Based on the data input, creates prior, model file, and initial values needed to run models in rjags'. Able to handle binomial, normal and multinomial arm-level data. Can handle multi-arm trials and includes methods to incorporate covariate and baseline risk effects. Includes standard diagnostics and visualization tools to evaluate the results.
Generic Extraction of main text content from HTML files; removal of ads, sidebars and headers using the boilerpipe <https://github.com/kohlschutter/boilerpipe> Java library. The extraction heuristics from boilerpipe show a robust performance for a wide range of web site templates.
Modelling of population growth under static and dynamic environmental conditions. Includes functions for model fitting and making prediction under isothermal and dynamic conditions. The methods (algorithms & models) are based on predictive microbiology (See Perez-Rodriguez and Valero (2012, ISBN:978-1-4614-5519-6)).
Fetches monthly financial tables and banking sector data published on the official website of the Banking Regulation and Supervision Agency of Turkey and also enables you to save it as an Excel file. It is a R implementation of the Python package <https://pypi.org/project/bddkdata/>.
The Bayesian optimal interval (BOIN) design is a novel phase I clinical trial design for finding the maximum tolerated dose (MTD). It can be used to design both single-agent and drug-combination trials. The BOIN design is motivated by the top priority and concern of clinicians when testing a new drug, which is to effectively treat patients and minimize the chance of exposing them to subtherapeutic or overly toxic doses. The prominent advantage of the BOIN design is that it achieves simplicity and superior performance at the same time. The BOIN design is algorithm-based and can be implemented in a simple way similar to the traditional 3+3 design. The BOIN design yields an average performance that is comparable to that of the continual reassessment method (CRM, one of the best model-based designs) in terms of selecting the MTD, but has a substantially lower risk of assigning patients to subtherapeutic or overly toxic doses. For tutorial, please check Yan et al. (2020) <doi:10.18637/jss.v094.i13>.
This package creates bivariate choropleth maps using Leaflet'. This package provides tools for visualizing the relationship between two variables through a color matrix representation on an interactive map.
Approximates best-subset selection (L0) regression with an iteratively adaptive Ridge (L2) penalty for large-scale models. This package uses Cyclops for an efficient implementation and the iterative method is described in Kawaguchi et al (2020) <doi:10.1002/sim.8438> and Li et al (2021) <doi:10.1016/j.jspi.2020.12.001>.
Query information and generate badge for using in README and GitHub Pages.
Using numeric or raster data, this package contains functions to calculate: complete water balance, bioclimatic balance, bioclimatic intensities, reports for individual locations, multi-layered rasters for spatial analysis.
We perform linear, logistic, and cox regression using the base functions lm(), glm(), and coxph() in the R software and the survival package. Likewise, we can use ols(), lrm() and cph() from the rms package for the same functionality. Each of these two sets of commands has a different focus. In many cases, we need to use both sets of commands in the same situation, e.g. we need to filter the full subset model using AIC, and we need to build a visualization graph for the final model. base.rms package can help you to switch between the two sets of commands easily.