Examples: visualization, C++, networks, data cleaning, html widgets, ropensci.

Found 1537 packages in 0.07 seconds

fuzzyforest — by Daniel Conn, 4 years ago

Fuzzy Forests

Fuzzy forests, a new algorithm based on random forests, is designed to reduce the bias seen in random forest feature selection caused by the presence of correlated features. Fuzzy forests uses recursive feature elimination random forests to select features from separate blocks of correlated features where the correlation within each block of features is high and the correlation between blocks of features is low. One final random forest is fit using the surviving features. This package fits random forests using the 'randomForest' package and allows for easy use of 'WGCNA' to split features into distinct blocks. See D. Conn, Ngun, T., C. Ramirez, and G. Li (2019) for further details.

quantregRanger — by Philipp Probst, 6 years ago

Quantile Regression Forests for 'ranger'

This is the implementation of quantile regression forests for the fast random forest package 'ranger'.

rpms — by Daniell Toth, 3 years ago

Recursive Partitioning for Modeling Survey Data

Functions to allow users to build and analyze design consistent tree and random forest models using survey data from a complex sample design. The tree model algorithm can fit a linear model to survey data in each node obtained by recursively partitioning the data. The splitting variables and selected splits are obtained using a randomized permutation test procedure which adjusted for complex sample design features used to obtain the data. Likewise the model fitting algorithm produces design-consistent coefficients to any specified least squares linear model between the dependent and independent variables used in the end nodes. The main functions return the resulting binary tree or random forest as an object of "rpms" or "rpms_forest" type. The package also provides methods modeling a "boosted" tree or forest model and a tree model for zero-inflated data as well as a number of functions and methods available for use with these object types.

bonsai — by Simon Couch, a year ago

Model Wrappers for Tree-Based Models

Bindings for additional tree-based model engines for use with the 'parsnip' package. Models include gradient boosted decision trees with 'LightGBM' (Ke et al, 2017.) and conditional inference trees and conditional random forests with 'partykit' (Hothorn and Zeileis, 2015. and Hothorn et al, 2006. ).

kernelFactory — by Michel Ballings, 9 years ago

Kernel Factory: An Ensemble of Kernel Machines

Binary classification based on an ensemble of kernel machines ("Ballings, M. and Van den Poel, D. (2013), Kernel Factory: An Ensemble of Kernel Machines. Expert Systems With Applications, 40(8), 2904-2913"). Kernel factory is an ensemble method where each base classifier (random forest) is fit on the kernel matrix of a subset of the training data.

SpatialML — by Stamatis Kalogirou, 18 days ago

Spatial Machine Learning

Implements a spatial extension of the random forest algorithm (Georganos et al. (2019) ). Allows for a geographically weighted random forest regression including a function to find the optical bandwidth. (Georganos and Kalogirou (2022) < https://www.mdpi.com/2220-9964/11/9/471>).

DMLLZU — by Junchang Zhao, 3 years ago

Double Machine Learning

Yang(2020,) come up with Double Machine Learning model ,it is based on this model, using four machine learning methods-- bagging, Boosting, random forest and neural network, and then based on the four models for two different combinations of the integrated model -- linear model combination and random forest .

SurvivalClusteringTree — by Lu You, 7 months ago

Clustering Analysis Using Survival Tree and Forest Algorithms

An outcome-guided algorithm is developed to identify clusters of samples with similar characteristics and survival rate. The algorithm first builds a random forest and then defines distances between samples based on the fitted random forest. Given the distances, we can apply hierarchical clustering algorithms to define clusters. Details about this method is described in < https://github.com/luyouepiusf/SurvivalClusteringTree>.

JOUSBoost — by Matthew Olson, 7 years ago

Implements Under/Oversampling for Probability Estimation

Implements under/oversampling for probability estimation. To be used with machine learning methods such as AdaBoost, random forests, etc.

outForest — by Michael Mayer, a year ago

Multivariate Outlier Detection and Replacement

Provides a random forest based implementation of the method described in Chapter 7.1.2 (Regression model based anomaly detection) of Chandola et al. (2009) . It works as follows: Each numeric variable is regressed onto all other variables by a random forest. If the scaled absolute difference between observed value and out-of-bag prediction of the corresponding random forest is suspiciously large, then a value is considered an outlier. The package offers different options to replace such outliers, e.g. by realistic values found via predictive mean matching. Once the method is trained on a reference data, it can be applied to new data.