This package provides functions for importing, creating, editing and exporting FSK files <https://foodrisklabs.bfr.bund.de/fskx-food-safety-knowledge-exchange-format/> using the R programming environment. Furthermore, it enables users to run simulations contained in the FSK files and visualize the results.
This package provides functions for modeling and forecasting time series data. Forecasting is based on the innovations algorithm. A description of the innovations algorithm can be found in the textbook "Introduction to Time Series and Forecasting" by Peter J. Brockwell and Richard A. Davis.
Temporary and permanent message queues for R. Built on top of SQLite databases. SQLite provides locking, and makes it possible to detect crashed consumers. Crashed jobs can be automatically marked as "failed", or put in the queue again, potentially a limited number of times.
Investigate the evolution of biological processes by capturing evolutionary signatures in transcriptomes (Drost et al. (2018) <doi:10.1093/bioinformatics/btx835>). This package aims to provide a transcriptome analysis environment to quantify the average evolutionary age of genes contributing to a transcriptome of interest.
Computes robust and bias-corrected sandwich variance estimators for multi-state Cox models with clustered time-to-event data. The methodology extends the marginal Cox model bias-correction framework of Wang et al. (2023) <doi:10.1002/bimj.202200113> to the multi-state setting.
Interface with the One Health VBD (vector-borne disease) Hub <https://vbdhub.org/> and related repositories (VectorByte <https://www.vectorbyte.org>, GBIF <https://www.gbif.org> and AREAdata <https://pearselab.github.io/areadata/>) directly to find, download, and subset vector-borne disease data.
This is an implementation of the partial profile score feature selection (PPSFS) approach to generalized linear (interaction) models. The PPSFS is highly scalable even for ultra-high-dimensional feature space. See the paper by Xu, Luo and Chen (2022) <doi:10.4310/21-SII706>.
Given a SpatialPolygonsDataFrame and a set of populations for each polygon, compute a population density estimate based on Tobler's pycnophylactic interpolation algorithm. The result is a SpatialGridDataFrame. Methods are described in Tobler Waldo R. (1979) <doi:10.1080/01621459.1979.10481647>.
This package provides tools for retrieving, organizing, and analyzing environmental data from the System Wide Monitoring Program of the National Estuarine Research Reserve System <https://cdmo.baruch.sc.edu/>. These tools address common challenges associated with continuous time series data for environmental decision making.
Handles both vector and matrices, using a flexible S4 class for automatic differentiation. The method used is forward automatic differentiation. Many functions and methods have been defined, so that in most cases, functions written without automatic differentiation in mind can be used without change.
Regularized version of partial least square approaches providing sparse, group, and sparse group versions of partial least square regression models (Liquet, B., Lafaye de Micheaux, P., Hejblum B., Thiebaut, R. (2016) <doi:10.1093/bioinformatics/btv535>). Version of PLS Discriminant analysis is also provided.
This package provides a fast and accurate pipeline for single-cell analyses. The scDHA software package can perform clustering, dimension reduction and visualization, classification, and time-trajectory inference on single-cell data (Tran et.al. (2021) <DOI:10.1038/s41467-021-21312-2>).
This package provides tools for reading, visualising and processing Magnetic Resonance Spectroscopy data. The package includes methods for spectral fitting: Wilson (2021) <DOI:10.1002/mrm.28385>, Wilson (2025) <DOI:10.1002/mrm.30462> and spectral alignment: Wilson (2018) <DOI:10.1002/mrm.27605>.
This package provides a novel and fast two stage method for simultaneous multiple change point detection and variable selection for piecewise stationary autoregressive (PSAR) processes and linear regression model. It also simultaneously performs variable selection for each autoregressive model and hence the order selection.
This package provides a simple type annotation for R that is usable in scripts, in the R console and in packages. It is intended as a convention to allow other packages to use the type information to provide error checking, automatic documentation or optimizations.
Accompanies the book Rainer Schlittgen and Cristina Sattarhoff (2020) <https://www.degruyter.com/view/title/575978> "Angewandte Zeitreihenanalyse mit R, 4. Auflage" . The package contains the time series and functions used therein. It was developed over many years teaching courses about time series analysis.
Estimates Pareto-optimal solution for personnel selection with 3 objectives using Normal Boundary Intersection (NBI) algorithm introduced by Das and Dennis (1998) <doi:10.1137/S1052623496307510>. Takes predictor intercorrelations and predictor-objective relations as input and generates a series of solutions containing predictor weights as output. Accepts between 3 and 10 selection predictors. Maximum 2 objectives could be adverse impact objectives. Partially modeled after De Corte (2006) TROFSS Fortran program <https://users.ugent.be/~wdecorte/trofss.pdf> and updated from ParetoR package described in Song et al. (2017) <doi:10.1037/apl0000240>. For details, see Study 3 of Zhang et al. (2023).
This package contains functions to retrieve, organize, and visualize weather data from the NCEP/NCAR Reanalysis (<https://psl.noaa.gov/data/gridded/data.ncep.reanalysis.html>) and NCEP/DOE Reanalysis II (<https://psl.noaa.gov/data/gridded/data.ncep.reanalysis2.html>) datasets. Data are queried via the Internet and may be obtained for a specified spatial and temporal extent or interpolated to a point in space and time. We also provide functions to visualize these weather data on a map. There are also functions to simulate flight trajectories according to specified behavior using either NCEP wind data or data specified by the user.
With this package we provide an easy method to compute robust and conditional Data Envelopment Analysis (DEA), Free Disposal Hull (FDH) and Benefit of the Doubt (BOD) scores. The robust approach is based on the work of Cazals, Florens and Simar (2002) <doi:10.1016/S0304-4076(01)00080-X>. The conditional approach is based on Daraio and Simar (2007) <doi:10.1007/s11123-007-0049-3>. Besides we provide graphs to help with the choice of m. We relay on the Benchmarking package to compute the efficiency scores and on the np package to compute non parametric estimation of similarity among units.
Fit (exponential or diffusion) response-time extended multinomial processing tree (RT-MPT) models by Klauer and Kellen (2018) <doi:10.1016/j.jmp.2017.12.003> and Klauer, Hartmann, and Meyer-Grant (submitted). The RT-MPT class not only incorporate frequencies like traditional multinomial processing tree (MPT) models, but also latencies. This enables it to estimate process completion times and encoding plus motor execution times next to the process probabilities of traditional MPTs. rtmpt is a hierarchical Bayesian framework and posterior samples are sampled using a Metropolis-within-Gibbs sampler (for exponential RT-MPTs) or Hamiltonian-within-Gibbs sampler (for diffusion RT-MPTs).
Sensitivity (or recall or true positive rate), false positive rate, specificity, precision (or positive predictive value), negative predictive value, misclassification rate, accuracy, F-score---these are popular metrics for assessing performance of binary classifiers for certain thresholds. These metrics are calculated at certain threshold values. Receiver operating characteristic (ROC) curve is a common tool for assessing overall diagnostic ability of the binary classifier. Unlike depending on a certain threshold, area under ROC curve (also known as AUC), is a summary statistic about how well a binary classifier performs overall for the classification task. The ROCit package provides flexibility to easily evaluate threshold-bound metrics.
This package contains many functions useful for data analysis, high-level graphics, utility operations, functions for computing sample size and power, importing and annotating datasets, imputing missing values, advanced table making, variable clustering, character string manipulation, conversion of R objects to LaTeX code, and recoding variables.
Optimized XML (Ox) is a fast XML parser and object serializer for Ruby written as a native C extension. It was designed to be an alternative to Nokogiri and other Ruby XML parsers for generic XML parsing and as an alternative to Marshal for Object serialization.
Strand specific peak-pair calling in ChIP-exo replicates. The cumulative Skellam distribution function is used to detect significant normalised count differences of opposed sign at each DNA strand (peak-pairs). Then, irreproducible discovery rate for overlapping peak-pairs across biological replicates is computed.