Random forests
Users: 4 - Average Rating: 4.75
This article describes the Random Forest (RF) learning method, a combination of tree predictors, each grown on a subset (sampling with replacement) of the original training set. The article mostly focuses on classification’s problems: each tree makes its prediction, i.e., votes for a class, and the most voted class is chosen as prediction of the forest for that instance. The paper shows with case studies that growing an ensemble of trees results in significant improvements in classification accuracy respect to single trees and obtain good results also respects to Adaboost.
First, the article gives some theoretical background for RFs and then carry out different examples to reinforce and explore the theoretical concepts. In particular, it focuses on two hot topics: how many features have to be randomly selected at each node to determine the split and the relationship between strength of the individual trees in the forest and correlation between trees.
First, the article gives some theoretical background for RFs and then carry out different examples to reinforce and explore the theoretical concepts. In particular, it focuses on two hot topics: how many features have to be randomly selected at each node to determine the split and the relationship between strength of the individual trees in the forest and correlation between trees.
Type:
Scientific Paper
Area:
Data Analytics, Machine Learning
Target Group:
Basic
DOI:
https://doi.org/10.1023/A:1010933404324
Cite as:
Breiman, L., Random forests, Machine learning 45.1 (2001): 5-32.
Author of the review:
Giulia Cademartori
University of Genoa
You have to login to leave a comment. If you are not registered click here
Giulia Cademartori
Pablo Guerrero-Garcia
Antonio Cinà