High-dimensional model selection via Chebyshev's greedy algorithm
- 2024-03-25 (Mon.), 10:30 AM
- Auditorium, B1F, Institute of Statistical Science;The tea reception will be held at 10:10.
- Lecture in English. Online live streaming through Cisco Webex will be available.
- Prof. Chien-Tong Lin
- Department of Statistics, Feng Chia University
Abstract
Assuming sparsity on the regression coefficient is fundamental to ultra-high dimensional variable selection. However, the true sparsity of practical data is typically uncertain, making it necessary to device a variable selection technique that performs well under various sparsity settings. In this talk, we investigate the convergence rate of Chebyshev's greedy algorithm (CGA) for regression models when the true coefficient vector satisfies a general weak sparsity condition. We determine the iteration number of CGA using our developed data-driven approach and demonstrate that the optimal convergence rate can be achieved even when the actual sparsity level is unknown. Our convergence theory relies on the convexity and the smoothness of the population loss function, allowing for the analysis of a broad family of regression models and providing optimality guarantees under weak assumptions. As a specific example, we apply our method to generalized linear models (GLM) and composite quantile regression (CQR) models, and offer the sufficient conditions under which the optimal rate can be achieved. Thorough simulation studies as well as data analysis are provided to support the obtained theory.
Please click here for participating the talk online.
Please click here for participating the talk online.
Download
Update:2024-03-20 13:55