top of page

Random Forest (Machine Learning Algorithm)

Writer's picture: Danielle Costa NakanoDanielle Costa Nakano

Description:  In Random Forest, we’ve collection of decision trees (so known as “Forest”). To classify a new object based on attributes, each tree gives a classification and we say the tree “votes” for that class. The forest chooses the classification having the most votes (over all the trees in the forest).

Random Forest is a trademark term for an ensemble of decision trees. Decision trees works for both categorical and continuous dependent variables.


Algorithm: Each tree is planted & grown as follows:

1.  If the number of cases in the training set is N, then sample of N cases is taken at random but with replacement. This sample will be the training set for growing the tree. 2.  If there are M input variables, a number m<<M is specified such that at each node, m variables are selected at random out of the M and the best split on these m is used to split the node. The value of m is held constant during the forest growing. 3.   Each tree is grown to the largest extent possible. There is no pruning.

7 views0 comments

Recent Posts

See All

Data Products: Too good to be true

If you are productizing a predictive model at work or playing around with MLE in R for the first time, always check the data. Roles When...

Comments


Subscribe Form

Thanks for submitting!

©2025 by Danielle Costa Nakano.

  • LinkedIn
bottom of page