Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
List of english scrabble words as listed in the OTCWL2014 <https://www.scrabbleplayers.org/w/Official_Tournament_and_Club_Word_List_2014_Edition>. Words are collated from the Word Game Dictionary <https://www.wordgamedictionary.com/word-lists/>.
Computes inequality measures of a given variable taking into account weights. Suitable for ratio, interval and ordered scale. Includes Gini, Theil, Leti index, Palma ratio, 20:20 ratio, Allison and Foster index, Jenkins index, Cowell and Flechaire index, Abul Naga and Yalcin index, Apouey index, Blair and Lacy index. Bootstrap provides distribution of inequality measures enabling significance tests.
Additional options for making graphics in the context of analyzing high-throughput data are available here. This includes automatic segmenting of the current device (eg window) to accommodate multiple new plots, automatic checking for optimal location of legends in plots, small histograms to insert as legends, histograms re-transforming axis labels to linear when plotting log2-transformed data, a violin-plot <doi:10.1080/00031305.1998.10480559> function for a wide variety of input-formats, principal components analysis (PCA) <doi:10.1080/14786440109462720> with bag-plots <doi:10.1080/00031305.1999.10474494> to highlight and compare the center areas for groups of samples, generic MA-plots (differential- versus average-value plots) <doi:10.1093/nar/30.4.e15>, staggered count plots and generation of mouse-over interactive html pages.
Utility functions to convert between the Spatial classes specified by the package sp', and the well-known binary (WKB) representation for geometry specified by the Open Geospatial Consortium'. Supports Spatial objects of class SpatialPoints', SpatialPointsDataFrame', SpatialLines', SpatialLinesDataFrame', SpatialPolygons', and SpatialPolygonsDataFrame'. Supports WKB geometry types Point', LineString', Polygon', MultiPoint', MultiLineString', and MultiPolygon'. Includes extensions to enable creation of maps with TIBCO Spotfire'.
This package provides a Shiny web application for life data analysis that depends on WeibullR by David Silkworth and Jurgen Symynck (2022) <https://CRAN.R-project.org/package=WeibullR>, an R package for Weibull analysis.
Easily override the default visual choices in ggplot2 to make your time series plots look more like the Wall Street Journal. Specific theme design choices include omitting x-axis grid lines and displaying sparse light grey y-axis grid lines. Additionally, this allows to label the y-axis scales with your units only displayed on the top-most number, while also removing the bottom most number (unless specifically overridden). The goal is visual simplicity, because who has time to waste looking at a cluttered graph?
The genetic algorithm is designed to optimize wind farms of any shape. It requires a predefined amount of turbines, a unified rotor radius and an average wind speed value for each incoming wind direction. A terrain effect model can be included that downloads an SRTM elevation model and loads a Corine Land Cover raster to approximate surface roughness.
This package provides a workflow for your analysis projects by combining literate programming ('knitr and rmarkdown') and version control ('Git', via git2r') to generate a website containing time-stamped, versioned, and documented results.
Computes Bayesian wavelet shrinkage credible intervals for nonparametric regression. The method uses cumulants to derive Bayesian credible intervals for wavelet regression estimates. The first four cumulants of the posterior distribution of the estimates are expressed in terms of the observed data and integer powers of the mother wavelet functions. These powers are closely approximated by linear combinations of wavelet scaling functions at an appropriate finer scale. Hence, a suitable modification of the discrete wavelet transform allows the posterior cumulants to be found efficiently for any data set. Johnson transformations then yield the credible intervals themselves. Barber, S., Nason, G.P. and Silverman, B.W. (2002) <doi:10.1111/1467-9868.00332>.
This package provides functions to create factor variables with contrasts based on weighted effect coding, and their interactions. In weighted effect coding the estimates from a first order regression model show the deviations per group from the sample mean. This is especially useful when a researcher has no directional hypotheses and uses a sample from a population in which the number of observation per group is different.
Read from, interogate, and write to Wikidata <https://www.wikidata.org> - the multilingual, interdisciplinary, semantic knowledgebase. Includes functions to: read from wikidata (single items, properties, or properties); query wikidata (retrieving all items that match a set of criterial via Wikidata SPARQL query service); write to Wikidata (adding new items or statements via QuickStatements); and handle and manipulate Wikidata objects (as lists and tibbles). Uses the Wikidata and Quickstatements APIs.
World Flora Online is an online flora of all known plants, available from <https://www.worldfloraonline.org/>. Methods are provided of matching a list of plant names (scientific names, taxonomic names, botanical names) against a static copy of the World Flora Online Taxonomic Backbone data that can be downloaded from the World Flora Online website. The World Flora Online Taxonomic Backbone is an updated version of The Plant List (<http://www.theplantlist.org/>), a working list of plant names that has become static since 2013.
This package provides a toolbox of common robust statistical tests, including robust descriptives, robust t-tests, and robust ANOVA. It is also available as a module for jamovi (see <https://www.jamovi.org> for more information). Walrus is based on the WRS2 package by Patrick Mair, which is in turn based on the scripts and work of Rand Wilcox. These analyses are described in depth in the book Introduction to Robust Estimation & Hypothesis Testing'.
Post-construction fatality monitoring studies at wind facilities are based on data from searches for bird and bat carcasses in plots beneath turbines. Bird and bat carcasses can fall outside of the search plot. Bird and bat carcasses from wind turbines often fall outside of the searched area. To compensate, area correction (AC) estimations are calculated to estimate the percentage of fatalities that fall within the searched area versus those that fall outside of it. This package provides two likelihood based methods and one physics based method (Hull and Muir (2010) <doi:10.1080/14486563.2010.9725253>, Huso and Dalthorp (2014) <doi:10.1002/jwmg.663>) to estimate the carcass fall distribution. There are also functions for calculating the proportion of area searched within one unit annuli, log logistic distribution functions, and truncated distribution functions.
Organizational framework for web development in R including functions to serve static and dynamic content via HTTP methods, includes the html5 package to create HTML pages, and offers other utility functions for common tasks related to web development.
An implementation of the Weighted Portmanteau Tests described in "New Weighted Portmanteau Statistics for Time Series Goodness-of-Fit Testing" published by the Journal of the American Statistical Association, Volume 107, Issue 498, pages 777-787, 2012.
Apply Wordpiece (<arXiv:1609.08144>) tokenization to input text, given an appropriate vocabulary. The BERT (<arXiv:1810.04805>) tokenization conventions are used by default.
The weighted ensemble method is a valuable approach for combining forecasts. This algorithm employs several optimization techniques to generate optimized weights. This package has been developed using algorithm of Armstrong (1989) <doi:10.1016/0024-6301(90)90317-W>.
This package provides maximum likelihood estimation methods for eight modified Weibull-type distributions. It returns parameter estimates, log-likelihood, AIC, and BIC, and also supports model fitting, validation, and comparison across different distributional forms. These methods can be applied to reliability, survival, and lifetime data analysis, making the package useful for researchers and practitioners in statistics, engineering, and medicine. The following distributions are included: Rangoli2023, Peng2014, Lai2003, Xie1996, Sarhan2009, Rangoli2025, Mustafa2012, and Alwasel2009.
This package provides a Stata-style `webuse()` function for importing named datasets from Stata's online collection.
R interface to a W3C Markup Validation service. See <https://validator.w3.org/> for more information.
Weighted Piecewise Kernel Density Estimation for large data.
This package provides tools for writing and debugging R code. Provides: %.>% dot-pipe (an S3 configurable pipe), unpack/to (R style multiple assignment/return), build_frame()'/'draw_frame() ('data.frame example tools), qc() (quoting concatenate), := (named map builder), let() (converts non-standard evaluation interfaces to parametric standard evaluation interfaces, inspired by gtools::strmacro() and base::bquote()'), and more.
Time series outlier detection with non parametric test. This is a new outlier detection methodology (washer): efficient for time saving elaboration and implementation procedures, adaptable for general assumptions and for needing very short time series, reliable and effective as involving robust non parametric test. You can find two approaches: single time series (a vector) and grouped time series (a data frame). For other informations: Andrea Venturini (2011) Statistica - Universita di Bologna, Vol.71, pp.329-344. For an informal explanation look at R-bloggers on web.