jump to main area
:::
A- A A+

Seminars

Asymptotic Properties of High-Dimensional Random Forests

  • 2022-01-17 (Mon.), 10:30 AM
  • The tea reception will be held at 10:10 at the Auditorium, B1F, Institute of Statistical Science.
  • Lecture in English. Online live streaming through Microsoft Teams will be available.
  • Dr. Chien-Ming Chi
  • Institute of Statistics, National Tsing Hua University, Taiwan; Data Sciences and Operations Department, Marshall School of Business, USC, USA.

Abstract

As a flexible nonparametric learning tool, random forests has been widely applied to various real applications with appealing empirical performance, even in the presence of high-dimensional feature space. Unveiling the underlying mechanisms has led to some important recent theoretical results on the consistency of the random forests algorithm and its variants. However, to our knowledge, all existing works concerning random forests consistency under the setting of high dimensionality were done for various modified random forests models where the splitting rules are independent of the response. In light of this, in this paper we derive the consistency rates for the random forests algorithm associated with the sample CART splitting criterion, which is the one used in the original version of the algorithm in Breiman (2001), in a general high-dimensional nonparametric regression setting through a bias-variance decomposition analysis. Our new theoretical results show that random forests can indeed adapt to high dimensionality and allow for discontinuous regression function. Our bias analysis characterizes explicitly how the random forests bias depends on the sample size, tree height, and column subsampling parameter. Some limitations of our current results are also discussed.
 

Download

01171030學術EN-紀建名.pdf
Update:2022-01-10 09:23
scroll to top