Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

BoostTree and BoostForest for Ensemble Learning (2003.09737v3)

Published 21 Mar 2020 in cs.LG and stat.ML

Abstract: Bootstrap aggregating (Bagging) and boosting are two popular ensemble learning approaches, which combine multiple base learners to generate a composite model for more accurate and more reliable performance. They have been widely used in biology, engineering, healthcare, etc. This paper proposes BoostForest, which is an ensemble learning approach using BoostTree as base learners and can be used for both classification and regression. BoostTree constructs a tree model by gradient boosting. It increases the randomness (diversity) by drawing the cut-points randomly at node splitting. BoostForest further increases the randomness by bootstrapping the training data in constructing different BoostTrees. BoostForest generally outperformed four classical ensemble learning approaches (Random Forest, Extra-Trees, XGBoost and LightGBM) on 35 classification and regression datasets. Remarkably, BoostForest tunes its parameters by simply sampling them randomly from a parameter pool, which can be easily specified, and its ensemble learning framework can also be used to combine many other base learners.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Changming Zhao (9 papers)
  2. Dongrui Wu (94 papers)
  3. Jian Huang (165 papers)
  4. Ye Yuan (274 papers)
  5. Hai-Tao Zhang (13 papers)
  6. Ruimin Peng (3 papers)
  7. Zhenhua Shi (9 papers)
Citations (26)

Summary

We haven't generated a summary for this paper yet.