Abstract

Sufficient dimension reduction reduces the dimension of a regression model without loss of information by replacing the original predictor with its lower-dimensional linear combinations. Partial (sufficient) dimension reduction arises when the predictors naturally fall into two sets X and W, and pursues a partial dimension reduction of X. Though partial dimension reduction is a very general problem, only very few research results are available when W is continuous. To the best of our knowledge, none can deal with the situation where the reduced lower-dimensional subspace of X varies with W. To address such issue, we in this paper propose a novel variable-dependent partial dimension reduction framework and adapt classical sufficient dimension reduction methods into this general paradigm. The asymptotic consistency of our method is investigated. Extensive numerical studies and real data analysis show that our variable-dependent partial dimension reduction method has superior performance compared to the existing methods.

Department(s)

Mathematics and Statistics

Comments

National Natural Science Foundation of China, Grant 22JC1400800

Keywords and Phrases

Directional Regression; Order Determination; Sliced Average Variance Estimation; Sliced Inverse Regression; Sufficient Dimension Reduction

International Standard Serial Number (ISSN)

1863-8260; 1133-0686

Document Type

Article - Journal

Document Version

Final Version

File Type

text

Language(s)

English

Rights

© 2023 Springer, All rights reserved.

Publication Date

01 Jan 2023

Share

 
COinS