Multi-view learning arouses vast amount of interest in the past decades with numerous real-world applications in web page analysis, bioinformatics, image processing and so on. Unlike the most previous works following the idea of co-training, in this paper we propose a new generative model for Multi-view Learning via Probabilistic Latent Semantic Analysis, called MVPLSA. In this model, we jointly model the co-occurrences of features and documents from different views. Specifically, in the model there are two latent variables y for the latent topic and z for the document cluster, and three visible variables d for the document, f for the feature, and v for the view label. The conditional probability p(zd), which is independent of v, is used as the bridge to share knowledge among multiple views. Also, we have p(yz, v) and p(fy, v), which are dependent of v, to capture the specifical structures inside each view. Experiments are conducted on four real-world data sets to demonstrate the effectiveness and superiority of our model.
- Generative model
- Multi-view learning
- Probabilistic Latent Semantic Analysis (PLSA)