discLDA: Discriminative Learning for Dimensionality Reduction and Classification

From statwiki
Jump to navigation Jump to search

Introduction

Dimensionality reduction is a common and often necessary step in most machine learning applications and high-dimensional data analyses. There exists some linear methods for dimensionality reduction such as principal component analysis (PCA) and Fisher discriminant analysis (FDA) and some nonlinear procedures such as kernelized versions of PCA and FDA as well as manifold learning algorithms.

A recent trend in dimensionality reduction is to focus on probabilistic models. These models, which include generative topological mapping, factor analysis, independent component analysis and probabilistic latent semantic analysis (pLSA), are generally specified in terms of an underlying independence assumption or low-rank assumption. The models are generally fit with maximum likelihood, although Bayesian methods are sometimes used.

1)[math]\displaystyle{ \theta_d }[/math] ~ [math]\displaystyle{ Dir (\alpha) }[/math]

2)zdn ~ Multi [math]\displaystyle{ (\theta_d) }[/math]