Discriminant Analysis (DA) for evolutionary inference (Qin, X. et al, 2020, <doi:10.22541/au.159256808.83862168>), especially for population genetic structure and community structure inference. This package incorporates the commonly used linear and non-linear, local and global supervised learning approaches (discriminant analysis), including Linear Discriminant Analysis of Kernel Principal Components (LDAKPC), Local (Fisher) Linear Discriminant Analysis (LFDA), Local (Fisher) Discriminant Analysis of Kernel Principal Components (LFDAKPC) and Kernel Local (Fisher) Discriminant Analysis (KLFDA). These discriminant analyses can be used to do ecological and evolutionary inference, including demography inference, species identification, and population/community structure inference.
Cross-validation methods of regression models that exploit features of various modeling functions to improve speed. Some of the methods implemented in the package are novel, as described in the package vignettes; for general introductions to cross-validation, see, for example, Gareth James, Daniela Witten, Trevor Hastie, and Robert Tibshirani (2021, ISBN 978-1-0716-1417-4, Secs. 5.1, 5.3), "An Introduction to Statistical Learning with Applications in R, Second Edition", and Trevor Hastie, Robert Tibshirani, and Jerome Friedman (2009, ISBN 978-0-387-84857-0, Sec. 7.10), "The Elements of Statistical Learning, Second Edition".
We provide a toolbox to conduct a Bayesian meta-analysis for estimating the current expansion rate of the Universe, called the Hubble constant H0, via time delay cosmography. The input data are Fermat potential difference and time delay estimates. For a robust inference, we assume a Student's t error for these inputs. Given these inputs, the meta-analysis produces posterior samples of the model parameters including the Hubble constant via Metropolis-Hastings within Gibbs. The package provides an option to implement repelling-attracting Metropolis-Hastings within Gibbs in a case where the parameter space has multiple modes.
Generate the James Blinding Index, as described in James et al (1996) <https://pubmed.ncbi.nlm.nih.gov/8841652/> and the Bang Blinding Index, as described in Bang et al (2004) <https://pubmed.ncbi.nlm.nih.gov/15020033/>. These are measures to assess whether or not satisfactory blinding has been maintained in a randomized, controlled, clinical trial. These can be generated for trial subjects, research coordinators and principal investigators, based upon standardized questionnaires that have been administered, to assess whether or not they can correctly guess to which treatment arm (e.g. placebo or treatment) subjects were assigned at randomization.
Flexible general-purpose toolbox implementing genetic algorithms (GAs) for stochastic optimisation. Binary, real-valued, and permutation representations are available to optimize a fitness function, i.e., a function provided by users depending on their objective function. Several genetic operators are available and can be combined to explore the best settings for the current task. Furthermore, users can define new genetic operators and easily evaluate their performances. Local search using general-purpose optimisation algorithms can be applied stochastically to exploit interesting regions. GAs can be run sequentially or in parallel, using an explicit master-slave parallelisation or a coarse-grain islands approach.
User-friendly functions for extracting a data table (row for each match, column for each group) from non-tabular text data using regular expressions, and for melting columns that match a regular expression. Patterns are defined using a readable syntax that makes it easy to build complex patterns in terms of simpler, re-usable sub-patterns. Named R arguments are translated to column names in the output; capture groups without names are used internally in order to provide a standard interface to three regular expression C libraries ('PCRE', RE2', ICU'). Output can also include numeric columns via user-specified type conversion functions.
Facilitates easy analysis of factorial experiments, including purely within-Ss designs (a.k.a. "repeated measures"), purely between-Ss designs, and mixed within-and-between-Ss designs. The functions in this package aim to provide simple, intuitive and consistent specification of data analysis and visualization. Visualization functions also include design visualization for pre-analysis data auditing, and correlation matrix visualization. Finally, this package includes functions for non-parametric analysis, including permutation tests and bootstrap resampling. The bootstrap function obtains predictions either by cell means or by more advanced/powerful mixed effects models, yielding predictions and confidence intervals that may be easily visualized at any level of the experiment's design.
This package provides a set of utilities for calculating the Deficit (frailty) Index (DI) in gerontological studies. The deficit index was first proposed by Arnold Mitnitski and Kenneth Rockwood and represents a proxy measure of aging and also can be served as a sensitive predictor of survival. For more information, see (i)"Accumulation of Deficits as a Proxy Measure of Aging" by Arnold B. Mitnitski et al. (2001), The Scientific World Journal 1, <DOI:10.1100/tsw.2001.58>; (ii) "Frailty, fitness and late-life mortality in relation to chronological and biological age" by Arnold B Mitnitski et al. (2001), BMC Geriatrics2002 2(1), <DOI:10.1186/1471-2318-2-1>.
Geographical detectors for measuring spatial stratified heterogeneity, as described in Jinfeng Wang (2010) <doi:10.1080/13658810802443457> and Jinfeng Wang (2016) <doi:10.1016/j.ecolind.2016.02.052>. Includes the optimal discretization of continuous data, four primary functions of geographical detectors, comparison of size effects of spatial unit and the visualizations of results. To use the package and to refer the descriptions of the package, methods and case datasets, please cite Yongze Song (2020) <doi:10.1080/15481603.2020.1760434>. The model has been applied in factor exploration of road performance and multi-scale spatial segmentation for network data, as described in Yongze Song (2018) <doi:10.3390/rs10111696> and Yongze Song (2020) <doi:10.1109/TITS.2020.3001193>, respectively.
Some tools for developing general equilibrium models and some general equilibrium models. These models can be used for teaching economic theory and are built by the methods of new structural economics (see LI Wu, 2019, ISBN: 9787521804225, General Equilibrium and Structural Dynamics: Perspectives of New Structural Economics. Beijing: Economic Science Press). The model form and mathematical methods can be traced back to J. von Neumann (1945, A Model of General Economic Equilibrium. The Review of Economic Studies, 13. pp. 1-9), J. G. Kemeny, O. Morgenstern and G. L. Thompson (1956, A Generalization of the von Neumann Model of an Expanding Economy, Econometrica, 24, pp. 115-135) et al. By the way, J. G. Kemeny is a co-inventor of the computer language BASIC.
The solution to some common problems is proposed, as well as a summary of some small functions. In particular, it provides a useful function for some problems in chemistry. For example, monoa()
, monob()
and mono()
function can be used to calculate The pH
of weak acid/base. The ggpng()
function can save the PNG format with transparent background. The period_table()
function will show the periodic table. Also the show_ruler()
function will show the ruler. The show_color()
function is funny and easier to show colors. I also provide the symb()
function to generate multiple symbols at once. The csv2vcf()
function provides an easy method to generate a file. The sym2poly()
and sym2coef()
function can extract coefficients from polynomials.
The aim of od is to provide tools and example datasets for working with origin-destination ('OD') datasets of the type used to describe aggregate urban mobility patterns (Carey et al. 1981) <doi:10.1287/trsc.15.1.32>. The package builds on functions for working with OD data in the package stplanr', (Lovelace and Ellison 2018) <doi:10.32614/RJ-2018-053> with a focus on computational efficiency and support for the sf class system (Pebesma 2018) <doi:10.32614/RJ-2018-009>. With few dependencies and a simple class system based on data frames, the package is intended to facilitate efficient analysis of OD datasets and to provide a place for developing new functions. The package enables the creation and analysis of geographic entities representing large scale mobility patterns, from daily travel between zones in cities to migration between countries.
This package performs (Adaptive) Boosting Trees for Poisson distributed response variables, using log-link function. The code approach is similar to the one used in gbm'/'gbm3'. Moreover, each tree in the expansion is built thanks to the rpart package. This package is based on following books and articles Denuit, M., Hainaut, D., Trufin, J. (2019) <doi:10.1007/978-3-030-25820-7> Denuit, M., Hainaut, D., Trufin, J. (2019) <doi:10.1007/978-3-030-57556-4> Denuit, M., Hainaut, D., Trufin, J. (2019) <doi:10.1007/978-3-030-25827-6> Denuit, M., Hainaut, D., Trufin, J. (2022) <doi:10.1080/03461238.2022.2037016> Denuit, M., Huyghe, J., Trufin, J. (2022) <https://dial.uclouvain.be/pr/boreal/fr/object/boreal%3A244325/datastream/PDF_01/view> Denuit, M., Trufin, J., Verdebout, T. (2022) <https://dial.uclouvain.be/pr/boreal/fr/object/boreal%3A268577>.
In mathematics, rejection sampling is a basic technique used to generate observations from a distribution. It is also commonly called the Acceptance-Rejection method or Accept-Reject algorithm and is a type of Monte Carlo method. Acceptance-Rejection method is based on the observation that to sample a random variable one can perform a uniformly random sampling of the 2D cartesian graph, and keep the samples in the region under the graph of its density function. Package AR is able to generate/simulate random data from a probability density function by Acceptance-Rejection method. Moreover, this package is a useful teaching resource for graphical presentation of Acceptance-Rejection method. From the practical point of view, the user needs to calculate a constant in Acceptance-Rejection method, which package AR is able to compute this constant by optimization tools. Several numerical examples are provided to illustrate the graphical presentation for the Acceptance-Rejection Method.
Reduced-rank regression, diagnostics and graphics.
R implementation of the common parsing tools lex and yacc'.
Extract the implied risk neutral density from options using various methods.
Play the classic game of tic-tac-toe (naughts and crosses).
This package lets you rarefy data, calculate diversity and plot the results.
DBI/RJDBC interface to h2 database. h2 version 2.3.232 is included.
Probabilistic analysis of probe reliability and differential gene expression on short oligonucleotide arrays.
Deprecated.
The rpx package implements an interface to proteomics data submitted to the ProteomeXchange
consortium.
The ROI is a framework for handling optimization problems in R.