Enter the query into the form above. You can look for specific version of a package by using @ symbol like this: gcc@10.
API method:
GET /api/packages?search=hello&page=1&limit=20
where search is your query, page is a page number and limit is a number of items on a single page. Pagination information (such as a number of pages and etc) is returned
in response headers.
If you'd like to join our channel webring send a patch to ~whereiseveryone/toys@lists.sr.ht adding your channel as an entry in channels.scm.
Fast fitting of generalised linear models on moderately large datasets, by taking an initial sample, fitting in memory, then evaluating the score function for the full data in the database. Thomas Lumley <doi:10.1080/10618600.2019.1610312>.
Function to test spatial segregation and association based in contingency table analysis of nearest neighbour counts following Dixon (2002) <doi:10.1080/11956860.2002.11682700>. Some Fortran code has been included to the original dixon2002() function of the ecespa package to improve speed.
Statistical hypothesis testing using the Delta method as proposed by Deng et al. (2018) <doi:10.1145/3219819.3219919>. This method replaces the standard variance estimation formula in the Z-test with an approximate formula derived via the Delta method, which can account for within-user correlation.
Query database tables over a DBI connection using data.table syntax. Attach database schemas to the search path. Automatically merge using foreign key constraints.
Attempt to repair inconsistencies and missing values in data records by using information from valid values and validation rules restricting the data.
The gap statistic approach is extended to estimate the number of clusters for categorical response format data. This approach and accompanying software is designed to be used with the output of any clustering algorithm and with distances specifically designed for categorical (i.e. multiple choice) or ordinal survey response data.
Collects a diverse range of symbolic data and offers a comprehensive set of functions that facilitate the conversion of traditional data into the symbolic data format.
Likelihood-based inference methods with doubly-truncated data are developed under various models. Nonparametric models are based on Efron and Petrosian (1999) <doi:10.1080/01621459.1999.10474187> and Emura, Konno, and Michimae (2015) <doi:10.1007/s10985-014-9297-5>. Parametric models from the special exponential family (SEF) are based on Hu and Emura (2015) <doi:10.1007/s00180-015-0564-z> and Emura, Hu and Konno (2017) <doi:10.1007/s00362-015-0730-y>. The parametric location-scale models are based on Dorre et al. (2021) <doi:10.1007/s00180-020-01027-6>.
Cancer genomes contain large numbers of somatic alterations but few genes drive tumor development. Identifying cancer driver genes is critical for precision oncology. Most of current approaches either identify driver genes based on mutational recurrence or using estimated scores predicting the functional consequences of mutations. driveR is a tool for personalized or batch analysis of genomic data for driver gene prioritization by combining genomic information and prior biological knowledge. As features, driveR uses coding impact metaprediction scores, non-coding impact scores, somatic copy number alteration scores, hotspot gene/double-hit gene condition, phenolyzer gene scores and memberships to cancer-related KEGG pathways. It uses these features to estimate cancer-type-specific probability for each gene of being a cancer driver using the related task of a multi-task learning classification model. The method is described in detail in Ulgen E, Sezerman OU. 2021. driveR: driveR: a novel method for prioritizing cancer driver genes using somatic genomics data. BMC Bioinformatics <doi:10.1186/s12859-021-04203-7>.
Analyzes group patterns using discourse analysis data with graph theory mathematics. Takes the order of which individuals talk and converts it to a network edge and weight list. Returns the density, centrality, centralization, and subgroup information for each group. Based on the analytical framework laid out in Chai et al. (2019) <doi:10.1187/cbe.18-11-0222>.
Extremely fast and memory efficient computation of the DER (or PaF) income polarization index as proposed by Duclos J. Y., Esteban, J. and Ray D. (2004). "Polarization: concepts, measurement, estimation". Econometrica, 72(6): 1737--1772. <doi:10.1111/j.1468-0262.2004.00552.x>. The index may be computed for a single or for a range of values of the alpha-parameter and bootstrapping is also available.
Interface for Rcpp users to dlib <http://dlib.net> which is a C++ toolkit containing machine learning algorithms and computer vision tools. It is used in a wide range of domains including robotics, embedded devices, mobile phones, and large high performance computing environments. This package allows R users to use dlib through Rcpp'.
This package provides a facility to generate efficient designs for order-of-additions experiments under pair-wise-order model, see Dennis K. J. Lin and Jiayu Peng (2019)."Order-of-addition experiments: A review and some new thoughts". Quality Engineering, 31:1, 49-59, <doi:10.1080/08982112.2018.1548021>. It also provides a facility to generate component orthogonal arrays under component position model, see Jian-Feng Yang, Fasheng Sun & Hongquan Xu (2020): "A Component Position Model, Analysis and Design for Order-of-Addition Experiments". Technometrics, <doi:10.1080/00401706.2020.1764394>.
Simultaneously detect the number and locations of change points in piecewise linear models under stationary Gaussian noise allowing autocorrelated random noise. The core idea is to transform the problem of detecting change points into the detection of local extrema (local maxima and local minima)through kernel smoothing and differentiation of the data sequence, see Cheng et al. (2020) <doi:10.1214/20-EJS1751>. A low-computational and fast algorithm call dSTEM is introduced to detect change points based on the STEM algorithm in D. Cheng and A. Schwartzman (2017) <doi:10.1214/16-AOS1458>.
Diffusion Weighted Imaging (DWI) is a Magnetic Resonance Imaging modality, that measures diffusion of water in tissues like the human brain. The package contains R-functions to process diffusion-weighted data. The functionality includes diffusion tensor imaging (DTI), diffusion kurtosis imaging (DKI), modeling for high angular resolution diffusion weighted imaging (HARDI) using Q-ball-reconstruction and tensor mixture models, several methods for structural adaptive smoothing including POAS and msPOAS, and a streamline fiber tracking for tensor and tensor mixture models. The package provides functionality to manipulate and visualize results in 2D and 3D.
This package provides methods for valuation of life insurance premiums and reserves (including variable-benefit and fractional coverage) based on "Actuarial Mathematics" by Bowers, H.U. Gerber, J.C. Hickman, D.A. Jones and C.J. Nesbitt (1997, ISBN: 978-0938959465), "Actuarial Mathematics for Life Contingent Risks" by Dickson, David C. M., Hardy, Mary R. and Waters, Howard R (2009) <doi:10.1017/CBO9780511800146> and "Life Contingencies" by Jordan, C. W (1952) <doi:10.1017/S002026810005410X>. It also contains functions for equivalent interest and discount rate calculation, present and future values of annuities, and loan amortization schedule.
Real life data is muddy, fuzzy and unpredictable. This makes data manipulations tedious and bringing the data to right shape alone is a major chunk of work. Functions in this package help us get an understanding of dataframes to dramatically reduces data coding time.
By systematically aggregating and processing textual reports from earthquakes, floods, storms, wildfires, and other natural disasters, the framework enables a holistic assessment of crisis narratives. Intelligent cleaning and normalization techniques transform raw commentary into structured data, ensuring precise extraction of disaster-specific insights. Collective sentiments of affected communities are quantitatively scored and qualitatively categorized, providing a multifaceted view of societal responses under duress. Interactive geographic maps and temporal charts illustrate the evolution and spatial dispersion of emotional reactions and impact indicators.
Compares two dataframes with a common key and returns the delta records. The package will return three dataframes that contain the added, changed, and deleted records.
Access and manage the application programming interface (API) of the United Nations Office for the Coordination of Humanitarian Affairs (OCHA) ReliefWeb disaster events at <https://reliefweb.int/disasters>. The package requires a minimal number of dependencies. It offers functionality to retrieve a user-defined sample of disaster events from ReliefWeb, providing an easy alternative to scraping the ReliefWeb website. It enables a seamless integration of regular data updates into the research work flow.
Lightweight utility functions used for the R package development infrastructure inside the data integration centers ('DIZ') to standardize and facilitate repetitive tasks such as setting up a database connection or issuing notification messages and to avoid redundancy.
Calculates distances from point locations to features. The usual approach for eg. resource selection function analyses is to generate a complete distance to features surface then sample it with your observed and random points. Since these raster based approaches can be pretty costly with large areas, and often lead to memory issues in R, the distanceto package opts to compute these distances using efficient, vector based approaches. As a helper, there's a decidedly low-res raster based approach for visually inspecting your region's distance surface. But the workhorse is distance_to.
Discrete splines are a class of univariate piecewise polynomial functions which are analogous to splines, but whose smoothness is defined via divided differences rather than derivatives. Tools for efficient computations relating to discrete splines are provided here. These tools include discrete differentiation and integration, various matrix computations with discrete derivative or discrete spline bases matrices, and interpolation within discrete spline spaces. These techniques are described in Tibshirani (2020) <doi:10.48550/arXiv.2003.03886>.
Allows to simulate SNP data using genlight objects. For example, it is straight forward to simulate a simple drift scenario with exchange of individuals between two populations or create a new genlight object based on allele frequencies of an existing genlight object.