A1 Refereed original research article in a scientific journal
Supervised invariant coordinate selection
Authors: Liski E, Nordhausen K, Oja H
Publisher: Taylor & Francis LTD
Publication year: 2014
Journal: Statistics
Journal name in source: STATISTICS
Journal acronym: Statistics
Volume: 48
Issue: 4
First page : 711
Last page: 731
Number of pages: 21
ISSN: 0233-1888
DOI: https://doi.org/10.1080/02331888.2013.800067
Dimension reduction plays an important role in high-dimensional data analysis. Principal component analysis, independent component analysis, and sliced inverse regression (SIR) are well known but very different analysis tools for the dimension reduction. It appears that these three approaches can all be seen as a comparison of two different scatter matrices S-1 and S-2. The components for dimension reduction are then given by the eigenvectors of S-1 S-2. In SIR, the second scatter matrix is supervised and therefore the choice of the components is based on the dependence between the observed random vector and a response variable. Based on these notions, we extend the invariant coordinate selection (ICS), allowing the second scatter matrix S-2 to be supervised; supervised ICS can then be used in supervised dimension reduction. It is remarkable that many supervised dimension reduction methods proposed in the literature such as the linear discriminant analysis, canonical correlation analysis, SIR, sliced average variance estimate, directional regression, and principal Hessian directions can be reformulated in this way. Several families of supervised scatter matrices are discussed, and their use in supervised dimension reduction is illustrated with a real data example and simulations.