Alternating proximal gradient method for sparse nonnegative Tucker decomposition

Yangyang Xu

Research output: Contribution to journalArticlepeer-review

64 Scopus citations

Abstract

Multi-way data arises in many applications such as electroencephalography classification, face recognition, text mining and hyperspectral data analysis. Tensor decomposition has been commonly used to find the hidden factors and elicit the intrinsic structures of the multi-way data. This paper considers sparse nonnegative Tucker decomposition (NTD), which is to decompose a given tensor into the product of a core tensor and several factor matrices with sparsity and nonnegativity constraints. An alternating proximal gradient method is applied to solve the problem. The algorithm is then modified to sparse NTD with missing values. Per-iteration cost of the algorithm is estimated scalable about the data size, and global convergence is established under fairly loose conditions. Numerical experiments on both synthetic and real world data demonstrate its superiority over a few state-of-the-art methods for (sparse) NTD from partial and/or full observations. The MATLAB code along with demos are accessible from the author’s homepage.

Original languageEnglish (US)
Pages (from-to)39-70
Number of pages32
JournalMathematical Programming Computation
Volume7
Issue number1
DOIs
StatePublished - Mar 2015
Externally publishedYes

Bibliographical note

Publisher Copyright:
© 2014, Springer-Verlag Berlin Heidelberg and Mathematical Optimization Society.

Keywords

  • Alternating proximal gradient method
  • Non-convex optimization
  • Sparse nonnegative Tucker decomposition
  • Sparse optimization

Fingerprint

Dive into the research topics of 'Alternating proximal gradient method for sparse nonnegative Tucker decomposition'. Together they form a unique fingerprint.

Cite this