Examples: visualization, C++, networks, data cleaning, html widgets, ropensci.

Found 1541 packages in 0.14 seconds

randomForest — by Andy Liaw, 2 years ago

Breiman and Cutler's Random Forests for Classification and Regression

Classification and regression based on a forest of trees using random inputs, based on Breiman (2001) .

randomForestSRC — by Udaya B. Kogalur, 5 months ago

Fast Unified Random Forests for Survival, Regression, and Classification (RF-SRC)

Fast OpenMP parallel computing of Breiman's random forests for univariate, multivariate, unsupervised, survival, competing risks, class imbalanced classification and quantile regression. New Mahalanobis splitting for correlated outcomes. Extreme random forests and randomized splitting. Suite of imputation methods for missing data. Fast random forests using subsampling. Confidence regions and standard errors for variable importance. New improved holdout importance. Case-specific importance. Minimal depth variable importance. Visualize trees on your Safari or Google Chrome browser. Anonymous random forests for data privacy.

RRF — by Houtao Deng, 2 years ago

Regularized Random Forest

Feature Selection with Regularized Random Forest. This package is based on the 'randomForest' package by Andy Liaw. The key difference is the RRF() function that builds a regularized random forest. Fortran original by Leo Breiman and Adele Cutler, R port by Andy Liaw and Matthew Wiener, Regularized random forest for classification by Houtao Deng, Regularized random forest for regression by Xin Guan. Reference: Houtao Deng (2013) .

missForest — by Daniel J. Stekhoven, 2 years ago

Nonparametric Missing Value Imputation using Random Forest

The function 'missForest' in this package is used to impute missing values particularly in the case of mixed-type data. It uses a random forest trained on the observed values of a data matrix to predict the missing values. It can be used to impute continuous and/or categorical data including complex interactions and non-linear relations. It yields an out-of-bag (OOB) imputation error estimate without the need of a test set or elaborate cross-validation. It can be run in parallel to save computation time.

drf — by Loris Michel, 3 years ago

Distributional Random Forests

An implementation of distributional random forests as introduced in Cevid & Michel & Meinshausen & Buhlmann (2020) .

arf — by Marvin N. Wright, 3 months ago

Adversarial Random Forests

Adversarial random forests (ARFs) recursively partition data into fully factorized leaves, where features are jointly independent. The procedure is iterative, with alternating rounds of generation and discrimination. Data becomes increasingly realistic at each round, until original and synthetic samples can no longer be reliably distinguished. This is useful for several unsupervised learning tasks, such as density estimation and data synthesis. Methods for both are implemented in this package. ARFs naturally handle unstructured data with mixed continuous and categorical covariates. They inherit many of the benefits of random forests, including speed, flexibility, and solid performance with default parameters. For details, see Watson et al. (2022) .

orf — by Gabriel Okasa, 2 years ago

Ordered Random Forests

An implementation of the Ordered Forest estimator as developed in Lechner & Okasa (2019) . The Ordered Forest flexibly estimates the conditional probabilities of models with ordered categorical outcomes (so-called ordered choice models). Additionally to common machine learning algorithms the 'orf' package provides functions for estimating marginal effects as well as statistical inference thereof and thus provides similar output as in standard econometric models for ordered choice. The core forest algorithm relies on the fast C++ forest implementation from the 'ranger' package (Wright & Ziegler, 2017) .

grf — by Erik Sverdrup, 2 months ago

Generalized Random Forests

Forest-based statistical estimation and inference. GRF provides non-parametric methods for heterogeneous treatment effects estimation (optionally using right-censored outcomes, multiple treatment arms or outcomes, or instrumental variables), as well as least-squares regression, quantile regression, and survival regression, all with support for missing covariates.

morf — by Riccardo Di Francesco, a year ago

Modified Ordered Random Forest

Nonparametric estimator of the ordered choice model using random forests. The estimator modifies a standard random forest splitting criterion to build a collection of forests, each estimating the conditional probability of a single class. The package also implements a nonparametric estimator of the covariates’ marginal effects.

LongituRF — by Louis Capitaine, 4 years ago

Random Forests for Longitudinal Data

Random forests are a statistical learning method widely used in many areas of scientific research essentially for its ability to learn complex relationships between input and output variables and also its capacity to handle high-dimensional data. However, current random forests approaches are not flexible enough to handle longitudinal data. In this package, we propose a general approach of random forests for high-dimensional longitudinal data. It includes a flexible stochastic model which allows the covariance structure to vary over time. Furthermore, we introduce a new method which takes intra-individual covariance into consideration to build random forests. The method is fully detailled in Capitaine et.al. (2020) Random forests for high-dimensional longitudinal data.