Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Based on data of real user-agent strings, we can set filtering conditions and randomly sample user-agent strings from the user-agent string pool.
Encapsulates functions to streamline calls from R to the REDCap API. REDCap (Research Electronic Data CAPture) is a web application for building and managing online surveys and databases developed at Vanderbilt University. The Application Programming Interface (API) offers an avenue to access and modify data programmatically, improving the capacity for literate and reproducible programming.
Converts data to STL (stereolithography) files that can be used to feed a 3-dimensional printer. The 3-dimensional output from a function can be materialized into a solid surface in a plastic material, therefore allowing more detailed examination. There are many possible uses for this new tool, such as to examine mathematical expressions with very irregular shapes, to aid teaching people with impaired vision, to create raised relief maps from digital elevation maps (DEMs), to bridge the gap between mathematical tools and rapid prototyping, and many more. Ian Walker created the function r2stl() and Jose Gama assembled the package.
This package provides a single method implementing multiple approaches to generate pseudo-random vectors whose components sum up to one (see, e.g., Maziero (2015) <doi:10.1007/s13538-015-0337-8>). The components of such vectors can for example be used for weighting objectives when reducing multi-objective optimisation problems to a single-objective problem in the socalled weighted sum scalarisation approach.
This package provides tools for basic and advance cancer statistics and graphics. Groups individual data, merges registry data and population data, calculates age-specific rate, age-standardized rate, cumulative risk, estimated annual percentage rate with standards error. Creates graphics across variable and time, such as age-specific trends, bar chart and period-cohort trends.
Function for generating random gender and ethnicity correct first and/or last names. Names are chosen proportionally based upon their probability of appearing in a large scale data base of real names.
It provides external jars required for the rjdverse (as rjd3toolkit', rjd3x13 and rjd3tramoseats').
Applies a given global-fit method to random partial tanglegrams of a fixed size to identify the associations, terminals, and nodes that maximize phylogenetic (in)congruence. It also includes functions to compute more easily the confidence intervals of classification metrics and plot results, reducing computational time. See Llaberia-Robledillo et al., (2023) <doi:10.1093/sysbio/syad016>.
This package provides a carbon-water coupled model (TRIPLEX-CW-Flux) is based on two well-established models, TRIPLEX-Flux model and Penmanâ Monteith model, integrates soil water and water vapor pressure deficits into the stomata conductance submodule to estimate net ecosystem production and evapotranspiration in forest ecosystems.<https://github.com/ShulanSun/rTRIPLEX_CW_Flux>.
Applies methods used to estimate animal homerange, but instead of geospatial coordinates, we use isotopic coordinates. The estimation methods include: 1) 2-dimensional bivariate normal kernel utilization density estimator, 2) bivariate normal ellipse estimator, and 3) minimum convex polygon estimator, all applied to stable isotope data. Additionally, functions to determine niche area, polygon overlap between groups and levels (confidence contours) and plotting capabilities.
R infrastructure for optimally robust estimation in general smoothly parameterized models using S4 classes and methods as described Kohl, M., Ruckdeschel, P., and Rieder, H. (2010), <doi:10.1007/s10260-010-0133-0>, and in Rieder, H., Kohl, M., and Ruckdeschel, P. (2008), <doi:10.1007/s10260-007-0047-7>.
Build regular expressions using grammar and functionality inspired by <https://github.com/VerbalExpressions>. Usage of the %>% is encouraged to build expressions in a chain-like fashion.
Within this package the XML-RPC API to NEOS <https://neos-server.org/neos/> is implemented. This enables the user to pass optimization problems to NEOS and retrieve results within R.
Render scenes using pathtracing. Build 3D scenes out of spheres, cubes, planes, disks, triangles, cones, curves, line segments, cylinders, ellipsoids, and 3D models in the Wavefront OBJ file format or the PLY Polygon File Format. Supports several material types, textures, multicore rendering, and tone-mapping. Based on the "Ray Tracing in One Weekend" book series. Peter Shirley (2018) <https://raytracing.github.io>.
High resolution vector country boundaries derived from Natural Earth data, can be plotted in rworldmap.
Parameter estimation, computation of probability, information, and (log-)likelihood, and visualization of item/test characteristic curves and item/test information functions for three uni-dimensional item response theory models: the 3-parameter-logistic model, generalized partial credit model, and graded response model. The full documentation and tutorials are at <https://github.com/xluo11/Rirt>.
This package provides a set of functions to perform pathway analysis and meta-analysis from multiple gene expression datasets, as well as visualization of the results. This package wraps functionality from the following packages: Ritchie et al. (2015) <doi:10.1093/nar/gkv007>, Love et al. (2014) <doi:10.1186/s13059-014-0550-8>, Robinson et al. (2010) <doi:10.1093/bioinformatics/btp616>, Korotkevich et al. (2016) <arxiv:10.1101/060012>, Efron et al. (2015) <https://CRAN.R-project.org/package=GSA>, and Gu et al. (2012) <https://CRAN.R-project.org/package=CePa>.
Load data from vk.com api about your communiti users and views, ads performance, post on user wall and etc. For more information see API Documentation <https://vk.com/dev/first_guide>.
Assists in the manipulation and processing of linear features with the help of the sf package. Makes use of linear referencing to extract data from most shape files. Reference for this packages methods: Albeke, S.E. et al. (2010) <doi:10.1007/s10980-010-9528-4>.
Enhances the R Optimization Infrastructure ('ROI') package by registering the free GLPK solver. It allows for solving mixed integer linear programming ('MILP') problems as well as all variants/combinations of LP', IP'.
This package provides portable access from R to biomedical image processing toolbox ANTs by Avants et al. (2009) <doi:10.54294/uvnhin> via seamless integration with the Python implementation ANTsPy'. Allows biomedical images to be processed in Python and analyzed in R', and vice versa via shared memory. See citation("rpyANTs") for more reference information.
The RBP curve is a visual tool to assess the performance of prediction models.
Rho is used to test the generalization of inter rater reliability (IRR) statistics. Calculating rho starts by generating a large number of simulated, fully-coded data sets: a sizable collection of hypothetical populations, all of which have a kappa value below a given threshold -- which indicates unacceptable agreement. Then kappa is calculated on a sample from each of those sets in the collection to see if it is equal to or higher than the kappa in then real sample. If less than five percent of the distribution of samples from the simulated data sets is greater than actual observed kappa, the null hypothesis is rejected and one can conclude that if the two raters had coded the rest of the data, we would have acceptable agreement (kappa above the threshold).
This package provides a custom implementation of the apriori algorithm and binomial tests to identify combinations of features (genes, variants etc) significantly enriched for simultaneous mutations/events from sparse Boolean input, see Vijay Kumar Pounraja, Santhosh Girirajan (2021). Version 1.1 includes a minor adjustment to the number of combinations to be considered for multiple testing correction. This updated version is more conservative in its approach and hence more selective. <doi:10.1101/2021.10.01.462832>.