Local structure in genomic data often induces dependence between observations taken at different genomic locations. Ignoring this dependence leads to underestimation of the standard error of parameter estimates. This package uses block bootstrapping to estimate asymptotically correct standard errors of parameters from any standard generalised linear model that may be fit by the glm()
function.
This package provides a collection of functions to set up Google Public Data Explorer <https://www.google.com/publicdata/> data visualization tool with your own data, building automatically the corresponding DataSet
Publishing Language file, or DSPL (XML), metadata file jointly with the CSV files. All zip-up and ready to be published in Public Data Explorer'.
Ensemble functions for outlier/anomaly detection. There is a new ensemble method proposed using Item Response Theory. Existing outlier ensemble methods from Schubert et al (2012) <doi:10.1137/1.9781611972825.90>, Chiang et al (2017) <doi:10.1016/j.jal.2016.12.002> and Aggarwal and Sathe (2015) <doi:10.1145/2830544.2830549> are also included.
Flask-RESTX is an extension for Flask that adds support for quickly building REST APIs. Flask-RESTX encourages best practices with minimal setup. If you are familiar with Flask, Flask-RESTX should be easy to pick up. It provides a coherent collection of decorators and tools to describe your API and expose its documentation properly using Swagger.
Scraping content from archived web pages stored in the Internet Archive (<https://archive.org>) using a systematic workflow. Get an overview of the mementos available from the respective homepage, retrieve the Urls and links of the page and finally scrape the content. The final output is stored in tibbles, which can be then easily used for further analysis.
You can apply image processing effects that modifies the perceived material properties of objects in photos, such as gloss, smoothness, and blemishes. This is an implementation of the algorithm proposed by Boyadzhiev et al. (2015) "Band-Sifting Decomposition for Image Based Material Editing". Documentation and practical tips of the package is available at <https://github.com/tsuda16k/materialmodifier>.
Extract the signed backbones of intrinsically dense weighted networks based on the significance filter and vigor filter as described in the following paper. Please cite it if you find this software useful in your work. Furkan Gursoy and Bertan Badur. "Extracting the signed backbone of intrinsically dense weighted networks." Journal of Complex Networks. <arXiv:2012.05216>
.
Visualize clonal expansion via circle-packing. APackOfTheClones
extends scRepertoire
to produce a publication-ready visualization of clonal expansion at a single cell resolution, by representing expanded clones as differently sized circles. The method was originally implemented by Murray Christian and Ben Murrell in the following immunology study: Ma et al. (2021) <doi:10.1126/sciimmunol.abg6356>.
This package provides YAML parser/emitter that supports roundtrip preservation of comments, seq/map flow style, and map key order. It is a derivative of Kirill Simonov’s PyYAML 3.11. It supports YAML 1.2 and has round-trip loaders and dumpers. It supports comments. Block style and key ordering are kept, so you can diff the source.
This package provides YAML parser/emitter that supports roundtrip preservation of comments, seq/map flow style, and map key order. It is a derivative of Kirill Simonov’s PyYAML 3.11. It supports YAML 1.2 and has round-trip loaders and dumpers. It supports comments. Block style and key ordering are kept, so you can diff the source.
This package provides a set of functions and a class to connect, extract and upload information from the LSEG Datastream database. This package uses the DSWS API and server used by the Datastream DFO addin'. Details of this API are available at <https://www.lseg.com/en/data-analytics>. Please report issues at <https://github.com/CharlesCara/DatastreamDSWS2R/issues>
.
Calculates landscape metrics for categorical landscape patterns in a tidy workflow. landscapemetrics reimplements the most common metrics from FRAGSTATS (<https://www.fragstats.org/>) and new ones from the current literature on landscape metrics. This package supports terra SpatRaster
objects as input arguments. It further provides utility functions to visualize patches, select metrics and building blocks to develop new metrics.
This package provides a selection of tools that make it easier to place elements onto a (base R) plot exactly where you want them. It allows users to identify points and distances on a plot in terms of inches, pixels, margin lines, data units, and proportions of the plotting space, all in a manner more simple than manipulating par()
.
Make acoustic cues to use with the R package ndl
. The package implements functions used in the PLoS ONE paper "Words from spontaneous conversational speech can be recognized with human-like accuracy by an error-driven learning algorithm that discriminates between meanings straight from smart acoustic features, bypassing the phoneme as recognition unit." doi:10.1371/journal.pone.0174623
Brings a set of tools to help and automatically realise the description of principal component analyses (from FactoMineR
functions). Detection of existing outliers, identification of the informative components, graphical views and dimensions description are performed threw dedicated functions. The Investigate()
function performs all these functions in one, and returns the result as a report document (Word, PDF or HTML).
This tiny crate checks that the running or installed rustc meets some version requirements. The version is queried by calling the Rust compiler with --version
. The path to the compiler is determined first via the RUSTC
environment variable. If it is not set, then rustc
is used. If that fails, no determination is made, and calls return None.
This tiny crate checks that the running or installed rustc meets some version requirements. The version is queried by calling the Rust compiler with --version
. The path to the compiler is determined first via the RUSTC
environment variable. If it is not set, then rustc
is used. If that fails, no determination is made, and calls return None.
The package uses PStricks and pst-solides3d
to draw three dimensional ribbons on a cylinder, torus, sphere, cone or paraboloid. The width of the ribbon, the number of turns, the colour of the outer and the inner surface of the ribbon may be set. In the case of circular and conical helices, one may also choose the number of ribbons.
Set of tools to simplify application of atomic forecast verification metrics for (comparative) verification of ensemble forecasts to large data sets. The forecast metrics are imported from the SpecsVerification
package, and additional forecast metrics are provided with this package. Alternatively, new user-defined forecast scores can be implemented using the example scores provided and applied using the functionality of this package.
Non linear dot plots are diagrams that allow dots of varying size to be constructed, so that columns with a large number of samples are reduced in height. Implementation of algorithm described in: Nils Rodrigues and Daniel Weiskopf, "Nonlinear Dot Plots", IEEE Transactions on Visualization and Computer Graphics, vol. 24, no. 1, pp. 616-625, 2018. <doi:10.1109/TVCG.2017.2744018>.
This crate is an async version of std::process
. A background thread named async-process
is lazily created on first use, which waits for spawned child processes to exit and then calls the wait()
syscall to clean up the ``zombie'' processes.
This is unlike the process API in the standard library, where dropping a running Child leaks its resources.
This crate is an async version of std::process
. A background thread named async-process
is lazily created on first use, which waits for spawned child processes to exit and then calls the wait()
syscall to clean up the ``zombie'' processes.
This is unlike the process API in the standard library, where dropping a running Child leaks its resources.
Fits Bayesian models (amongst others) to dissolution data sets that can be used for dissolution testing. The package was originally constructed to include only the Bayesian models outlined in Pourmohamad et al. (2022) <doi:10.1111/rssc.12535>. However, additional Bayesian and non-Bayesian models (based on bootstrapping and generalized pivotal quanties) have also been added. More models may be added over time.
These routines create multiple imputations of missing at random categorical data, and create multiply imputed synthesis of categorical data, with or without structural zeros. Imputations and syntheses are based on Dirichlet process mixtures of multinomial distributions, which is a non-parametric Bayesian modeling approach that allows for flexible joint modeling, described in Manrique-Vallier and Reiter (2014) <doi:10.1080/10618600.2013.844700>.