adaptive dimension reduction for clustering high dimensional data

From statwiki
Jump to navigation Jump to search
The printable version is no longer supported and may have rendering errors. Please update your browser bookmarks and please use the default browser print function instead.

1. Introduction

Clustering methods such as the K-means and EM suffer from local minima problems. In high dimensional space, the cost function surface is very rugged and it is easy to get trapped somewhere close to the initial configurations. The conventional method is to try a number of initial values, and pick up the best one of the results.

The Paper proposes a approach to tackle this problem. The approach utilizes the idea of dimension reduction. In the paper, (i) they approach dimension reduction as a dynamic process that is adaptively adjusted and integrated with clustering process (ii) make effective use of cluster membership to connect reduced dimensional space and full dimension space.

The paper focuses on K-means and EM algorithms using mixture model of spherical Gaussian components, which retain identical model parameters in reduced low-dimensional subspace as in high dimensional space.

2. Effective Dimension for Clustering

Low dimensional subspace is interpreted as containing relevant attributes( linear combinations of coordinates). The dimensionality r of subspace should satisfy: [math]\displaystyle{ y \leq K-1 }[/math] based on linear discriminant analysis. Thus the effective clustering dimensions for the K spherical Gaussians are spanned by the K centers [math]\displaystyle{ \mu_1, ...,\mu_k }[/math], for which r=K-1. We call the relevant dimensions passing through all K centers to the r-dim subspace. The effective dimensionality of the relevant subspace could be less than K-1. This happens when the K cluster centers lie in a subspace with dimension [math]\displaystyle{ r\lt K-1 }[/math]

3. EM in relevant subspace

4. Adaptive Dimension Reduction for EM

5. Adaptive Dimension Reduction for K-means