A new optimization criterion for discriminant analysis is presented. The new criterion extends the optimization criteria of the classical linear discriminant analysis (LDA) by introducing the pseudo-inverse when the scatter matrices are singular. It is applicable regardless of the relative sizes of the data dimension and sample size, overcoming a limitation of the classical LDA. Recently, a new algorithm called LDA/GSVD for structure-preserving dimension reduction has been introduced, which extends the classical LDA to very high-dimensional undersampled problems by using the generalized singular value decomposition (GSVD). The solution from the LDA/GSVD algorithm is a special case of the solution for our generalized criterion in this paper, which is also based on GSVD. We also present an approximate solution for our GSVDbased solution, which reduces computational complexity by finding sub-clusters of each cluster, and using their centroids to capture the structure of each cluster. This reduced problem yields much smaller matrices of which the GSVD can be applied efficiently. Experiments on text data, with up to 7000 dimensions, show that the approximation algorithm produces results that are close to those produced by the exact algorithm.
|Original language||English (US)|
|Title of host publication||Proceedings - 3rd IEEE International Conference on Data Mining, ICDM 2003|
|Number of pages||8|
|State||Published - 2003|
|Event||3rd IEEE International Conference on Data Mining, ICDM '03 - Melbourne, FL, United States|
Duration: Nov 19 2003 → Nov 22 2003
|Name||Proceedings - IEEE International Conference on Data Mining, ICDM|
|Other||3rd IEEE International Conference on Data Mining, ICDM '03|
|Period||11/19/03 → 11/22/03|
Copyright 2010 Elsevier B.V., All rights reserved.