This package provides an interface to a large number of classification and regression techniques. These techniques include machine-readable parameter descriptions. There is also an experimental extension for survival analysis, clustering and general, example-specific cost-sensitive learning. Also included:
Generic resampling, including cross-validation, bootstrapping and subsampling;
Hyperparameter tuning with modern optimization techniques, for single- and multi-objective problems;
Filter and wrapper methods for feature selection;
Extension of basic learners with additional operations common in machine learning, also allowing for easy nested resampling.
Most operations can be parallelized.
mlr3
enables efficient, object-oriented programming on the building blocks of machine learning. It provides R6
objects for tasks, learners, resamplings, and measures. The package is geared towards scalability and larger datasets by supporting parallelization and out-of-memory data-backends like databases. While mlr3
focuses on the core computational operations, add-on packages provide additional functionality.
Plug-in and difference-based long-run covariance matrix estimation for time series regression. Two applications of hypothesis testing are also provided. The first one is for testing for structural stability in coefficient functions. The second one is aimed at detecting long memory in time series regression. Lujia Bai and Weichi Wu (2024)<doi:10.3150/23-BEJ1680> Zhou Zhou and Wei Biao Wu(2010)<doi:10.1111/j.1467-9868.2010.00743.x> Jianqing Fan and Wenyang Zhang<doi:10.1214/aos/1017939139> Lujia Bai and Weichi Wu(2024)<doi:10.1093/biomet/asae013> Dimitris N. Politis, Joseph P. Romano, Michael Wolf(1999)<doi:10.1007/978-1-4612-1554-7> Weichi Wu and Zhou Zhou(2018)<doi:10.1214/17-AOS1582>.
The stepwise regression with assumptions checking and the possible Box-Cox transformation.
Toolset that enriches mlr with a diverse set of preprocessing operators. Composable Preprocessing Operators ("CPO"s) are first-class R objects that can be applied to data.frames and mlr "Task"s to modify data, can be attached to mlr "Learner"s to add preprocessing to machine learning algorithms, and can be composed to form preprocessing pipelines.
Extends the mlr3 package with a backend to transparently work with databases such as SQLite', DuckDB
', MySQL
', MariaDB
', or PostgreSQL
'. The package provides two additional backends: DataBackendDplyr
relies on the abstraction of package dbplyr to interact with most DBMS. DataBackendDuckDB
operates on DuckDB
data bases and also on Apache Parquet files.
This package is a flexible and comprehensive R toolbox for model-based optimization. It implements Efficient Global Optimization Algorithm for single- and multi-objective optimization. It supports mixed parameters. The machine learning toolbox mlr offers regression learners. It provides various infill criteria and features batch proposal, parallel execution, visualization, and logging. Its modular implementation allows easy customization by the user.
This package provides an interface to OpenML.org
to list and download machine learning data, tasks and experiments. The OpenML
objects can be automatically converted to mlr3 objects. For a more sophisticated interface with more upload options, see the OpenML
package.
This package offers features plots for mlr3 objects such as tasks, learners, predictions, benchmark results, tuning instances and filters via the autoplot()
generic of ggplot2. The mlr3viz package draws plots with the viridis color palette and applies the minimal theme. Visualizations include barplots, boxplots, histograms, ROC curves, and precision-recall curves.
Extends the mlr3 ecosystem to functional analysis by adding support for irregular and regular functional data as defined in the tf package. The package provides PipeOps
for preprocessing functional columns and for extracting scalar features, thereby allowing standard machine learning algorithms to be applied afterwards. Available operations include simple functional features such as the mean or maximum, smoothing, interpolation, flattening, and functional PCA'.
This package provides a flexible approach to Bayesian optimization / model based optimization building on the bbotk package. The mlr3mbo is a toolbox providing both ready-to-use optimization algorithms as well as their fundamental building blocks allowing for straightforward implementation of custom algorithms. Single- and multi-objective optimization is supported as well as mixed continuous, categorical and conditional search spaces. Moreover, using mlr3mbo for hyperparameter optimization of machine learning models within the mlr3 ecosystem is straightforward via mlr3tuning.
mlr3misc
provides frequently used helper functions and assertions used in mlr3
and its companion packages. It comes with helper functions for functional programming, for printing, to work with data.table
, as well as some generally useful R6
classes. This package also supersedes the package BBmisc
.
This package provides a small collection of interesting and educational machine learning data sets which are used as examples in the mlr3 book Applied machine learning using mlr3 in R https://mlr3book.mlr-org.com, the use case gallery https://mlr3gallery.mlr-org.com, or in other examples. All data sets are properly preprocessed and ready to be analyzed by most machine learning algorithms. Data sets are automatically added to the dictionary of tasks if mlr3 is loaded.
This package provides a web-based graphical user interface to provide the basic steps of a machine learning workflow. It uses the functionalities of the mlr3 framework.
The mlr3 package family is a set of packages for machine-learning purposes built in a modular fashion. This wrapper package is aimed to simplify the installation and loading of the core mlr3 packages.
Deep Learning library that extends the mlr3 framework by building upon the torch package. It allows to conveniently build, train, and evaluate deep learning models without having to worry about low level details. Custom architectures can be created using the graph language defined in mlr3pipelines'.
mlr3tuning
implements methods for hyperparameter tuning, e.g. Grid Search, Random Search, or Simulated Annealing. Various termination criteria can be set and combined. The class AutoTuner
provides a convenient way to perform nested resampling in combination with mlr3
.
This package extends the mlr3 package with cluster analysis.
This package extends mlr3
with support for performing ordinal regression.
Extends the mlr3 ML framework with methods for spatial objects. Data storage and prediction are supported for packages terra', raster and stars'.
This package extends mlr3 with filter methods for feature selection. Besides standalone filter methods built-in methods of any machine-learning algorithm are supported. Partial scoring of multivariate filter methods is supported.
Concise and interpretable summaries for machine learning models and learners of the mlr3 ecosystem. The package takes inspiration from the summary function for (generalized) linear models but extends it to non-parametric machine learning models, based on generalization performance, model complexity, feature importances and effects, and fairness metrics.
This package is a feature selection package of the mlr3 ecosystem. It selects the optimal feature set for any mlr3 learner. The package works with several optimization algorithms e.g. random search, Recursive feature elimination, and genetic search. Moreover, it can automatically optimize learners and estimate the performance of optimized feature sets with nested resampling.
The mlrMBO
package can ordinarily not be used for optimization within mlr3', because of incompatibilities of their respective class systems. mlrintermbo offers a compatibility interface that provides mlrMBO
as an mlr3tuning Tuner object, for tuning of machine learning algorithms within mlr3', as well as a bbotk Optimizer object for optimization of general objective functions using the bbotk black box optimization framework. The control parameters of mlrMBO
are faithfully reproduced as a paradox ParamSet
'.