Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Cronbach's alpha and various formulas for confidence intervals. The relevant paper is Tsagris M., Frangos C.C. and Frangos C.C. (2013). "Confidence intervals for Cronbach's reliability coefficient". Recent Techniques in Educational Science, 14-16 May, Athens, Greece.
With this package you can run ConMET locally in R. ConMET is an R-shiny application that facilitates performing and evaluating confirmatory factor analyses (CFAs) and is useful for running and reporting typical measurement models in applied psychology and management journals. ConMET automatically creates, compares and summarizes CFA models. Most common fit indices (E.g., CFI and SRMR) are put in an overview table. ConMET also allows to test for common method variance. The application is particularly useful for teaching and instruction of measurement issues in survey research. The application uses the lavaan package (Rosseel, 2012) to run CFAs.
Clustering multi-subject resting state functional Magnetic Resonance Imaging data. This methods enables the clustering of subjects based on multi-subject resting state functional Magnetic Resonance Imaging data. Objects are clustered based on similarities and differences in cluster-specific estimated components obtained by Independent Component Analysis.
Gather boxscore and play-by-play data from the Canadian Elite Basketball League (CEBL) <https://www.cebl.ca> to create a repository of basic and advanced statistics for teams and players.
Balancing and rounding matrices subject to restrictions. Adjustment of matrices so that columns and rows add up to given vectors, rounding of a matrix while keeping the column and/or row totals, performing these by blocks...
An interactive document on the topic of classification tree analysis using rmarkdown and shiny packages. Runtime examples are provided in the package function as well as at <https://kartikeyab.shinyapps.io/CTShiny/>.
Perform evaluation of automatic subject indexing methods. The main focus of the package is to enable efficient computation of set retrieval and ranked retrieval metrics across multiple dimensions of a dataset, e.g. document strata or subsets of the label set. The package also provides the possibility of computing bootstrap confidence intervals for all major metrics, with seamless integration of parallel computation and propensity scored variants of standard metrics.
Includes commands for bootstrapping and permutation tests, a command for created grouped bar plots, and a demo of the quantile-normal plot for data drawn from different distributions.
This package performs least squares constrained optimization on a linear objective function. It contains a number of algorithms to choose from and offers a formula syntax similar to lm().
This package provides a convenient interface for making requests directly to the Civis Platform API <https://www.civisanalytics.com/platform>. Full documentation available here <https://civisanalytics.github.io/civis-r/>.
This package provides equations commonly used in clinical pharmacokinetics and clinical pharmacology, such as equations for dose individualization, compartmental pharmacokinetics, drug exposure, anthropomorphic calculations, clinical chemistry, and conversion of common clinical parameters. Where possible and relevant, it provides multiple published and peer-reviewed equations within the respective R function.
Predict the course of clinical trial with a time-to-event endpoint for both two-arm and single-arm design. Each of the four primary study design parameters (the expected number of observed events, the number of subjects enrolled, the observation time, and the censoring parameter) can be derived analytically given the other three parameters. And the simulation datasets can be generated based on the design settings.
Access Cloudstor via their WebDAV API. This package can read, write, and navigate Cloudstor from R.
This package contains functions to help in selecting and exploring features ( or variables ) in binary classification problems. Provides functions to compute and display information value and weight of evidence (WoE) of the variables , and to convert numeric variables to categorical variables by binning. Functions are also provided to determine which levels ( or categories ) of a categorical variable can be collapsed (or combined ) based on their response rates. The functions provided only work for binary classification problems.
This package contains tools for working with data during statistical analysis, promoting flexible, intuitive, and reproducible workflows. There are functions designated for specific statistical tasks such building a custom univariate descriptive table, computing pairwise association statistics, etc. These are built on a collection of data manipulation tools designed for general use that are motivated by the functional programming concept.
Utilize the shiny interface to generate Goodness of Fit (GOF) plots and tables for Non-Linear Mixed Effects (NLME / NONMEM) pharmacometric models. From the interface, users can customize model diagnostics and generate the underlying R code to reproduce the diagnostic plots and tables outside of the shiny session. Model diagnostics can be included in a rmarkdown document and rendered to desired output format.
Implementation of a procedure---Domingue (2012) <https://eric.ed.gov/?id=ED548657>, Domingue (2014) <doi:10.1007/s11336-013-9342-4>; see also Karabatsos (2001) <https://psycnet.apa.org/record/2002-01665-005> and Kyngdon (2011) <doi:10.1348/2044-8317.002004>---to test the single and double cancellation axioms of conjoint measure in data that is dichotomously coded and measured with error.
This package provides ability to control how many times in function calls conditions are thrown (shown to the user). Includes control of warnings and messages.
To improve estimation accuracy and stability in statistical modeling, catalytic prior distributions are employed, integrating observed data with synthetic data generated from a simpler model's predictive distribution. This approach enhances model robustness, stability, and flexibility in complex data scenarios. The catalytic prior distributions are introduced by Huang et al. (2020, <doi:10.1073/pnas.1920913117>), Li and Huang (2023, <doi:10.48550/arXiv.2312.01411>).
Support for import from and export to the CSVY file format. CSVY is a file format that combines the simplicity of CSV (comma-separated values) with the metadata of other plain text and binary formats (JSON, XML, Stata, etc.) by placing a YAML header on top of a regular CSV.
An R DataBase Interface ('DBI') compatible interface to various database platforms ('PostgreSQL', Oracle', Microsoft SQL Server', Amazon Redshift', Microsoft Parallel Database Warehouse', IBM Netezza', Apache Impala', Google BigQuery', Snowflake', Spark', SQLite', and InterSystems IRIS'). Also includes support for fetching data as Andromeda objects. Uses either Java Database Connectivity ('JDBC') or other DBI drivers to connect to databases.
Efficient covariate-adjusted estimators of quantities that are useful for establishing the effects of treatments on ordinal outcomes.
Compare detrital zircon suites by uploading univariate, U-Pb age, or bivariate, U-Pb age and Lu-Hf data, in a shiny'-based user-interface. Outputs publication quality figures using ggplot2', and tables of statistics currently in use in the detrital zircon geochronology community.
The function takes a DNA sequence, a start point, an end point in the sequence, dot size and dot color and draws a fractal image of the sequence. The fractal starts in the center of the canvas. The image is drawn by moving base by base along the sequence and dropping a midpoint between the actual point and the corner designated by the actual base. For more details see Jeffrey (1990) <doi:10.1093/nar/18.8.2163>, Hill, Schisler, and Singh (1992) <doi:10.1007/BF00178602>, and Löchel and Heider (2021) <doi:10.1016/j.csbj.2021.11.008>.