Dimension folding PCA and PFC for matrix-valued predictors

Shanshan Ding, R. Dennis Cook

Research output: Contribution to journalArticlepeer-review

36 Scopus citations

Abstract

Conventional dimension reduction methods deal mainly with simple data structure and are inappropriate for data with matrix-valued predictors. Li, Kim, and Altman (2010) proposed dimension folding methods that effectively improve major moment-based dimension reduction techniques for the more complex data structure. Their methods, however, are moment-based and rely on slicing the responses to gain information about the conditional distribution of X|Y. This can be inadequate when the number of slices is not chosen properly. We propose model-based dimension folding methods that can be treated as extensions of conventional principal components analysis (PCA) and principal fitted components (PFC). We refer to them as dimension folding PCA and dimension folding PFC. The proposed methods can simultaneously reduce a predictor's multiple dimensions and inherit asymptotic properties from maximum likelihood estimation. Dimension folding PFC gains further efficiency by effective use of the response information. Both methods can provide robust estimation and are computationally efficient. We demonstrated their advantages by both simulation and data analysis.

Original languageEnglish (US)
Pages (from-to)463-492
Number of pages30
JournalStatistica Sinica
Volume24
Issue number1
DOIs
StatePublished - Jan 2014

Keywords

  • Central dimension folding subspace
  • Central subspace
  • Inverse regression
  • Matrix normal distribution
  • Sufficient dimension reduction

Fingerprint

Dive into the research topics of 'Dimension folding PCA and PFC for matrix-valued predictors'. Together they form a unique fingerprint.

Cite this