Boosted Random Forest

Yohei MISHINA  Ryuei MURATA  Yuji YAMAUCHI  Takayoshi YAMASHITA  Hironobu FUJIYOSHI  

Publication
IEICE TRANSACTIONS on Information and Systems   Vol.E98-D   No.9   pp.1630-1636
Publication Date: 2015/09/01
Online ISSN: 1745-1361
DOI: 10.1587/transinf.2014OPP0004
Type of Manuscript: Special Section PAPER (Special Section on Optimization and Learning Algorithms of Small Embedded Devices and Related Software/Hardware Implementation)
Category: 
Keyword: 
Boosting,  Random Forest,  machine learning,  pattern recognition,  

Full Text: PDF(744.9KB)>>
Buy this Article




Summary: 
Machine learning is used in various fields and demand for implementations is increasing. Within machine learning, a Random Forest is a multi-class classifier with high-performance classification, achieved using bagging and feature selection, and is capable of high-speed training and classification. However, as a type of ensemble learning, Random Forest determines classifications using the majority of multiple trees; so many decision trees must be built. Performance increases with the number of decision trees, requiring memory, and decreases if the number of decision trees is decreased. Because of this, the algorithm is not well suited to implementation on small-scale hardware as an embedded system. As such, we have proposed Boosted Random Forest, which introduces a boosting algorithm into the Random Forest learning method to produce high-performance decision trees that are smaller. When evaluated using databases from the UCI Machine learning Repository, Boosted Random Forest achieved performance as good or better than ordinary Random Forest, while able to reduce memory use by 47%. Thus, it is suitable for implementing Random Forests on embedded hardware with limited memory.