Sufficient dimension reduction reduces the dimension of a regression model without loss of information by replacing the original predictor with its lower-dimensional linear combinations. Partial (sufficient) dimension reduction arises when the predictors naturally fall into two sets X and W, and pursues a partial dimension reduction of X. Though partial dimension reduction is a very general problem, only very few research results are available when W is continuous. To the best of our knowledge, none can deal with the situation where the reduced lower-dimensional subspace of X varies with W. To address such issue, we in this paper propose a novel variable-dependent partial dimension reduction framework and adapt classical sufficient dimension reduction methods into this general paradigm. The asymptotic consistency of our method is investigated. Extensive numerical studies and real data analysis show that our variable-dependent partial dimension reduction method has superior performance compared to the existing methods.
L. Li et al., "Variable-Dependent Partial Dimension Reduction," Test, Springer, Jan 2023.
The definitive version is available at https://doi.org/10.1007/s11749-022-00841-y
Mathematics and Statistics
Keywords and Phrases
Directional Regression; Order Determination; Sliced Average Variance Estimation; Sliced Inverse Regression; Sufficient Dimension Reduction
International Standard Serial Number (ISSN)
Article - Journal
© 2023 Springer, All rights reserved.
01 Jan 2023
National Natural Science Foundation of China, Grant 22JC1400800