# paper 13

## Random mapping (Random Projection)

The following text is a short review and brief introduction about Random Mapping. <ref> Samuel Kaski; Dimensionality Reduction by Random Mapping:Fast Similarity Computation for Clustering, 1998 IEEE </ref> and some important results for estimating the Intrinsic Dimensionality of the manifold embedding in high dimensional space using $n$ sample points as well as results about the upper an lower bounds for Intrinsic Dimensionality of projected manifold and minimum number of random projections. <ref> Chinmay Hedg and Colleagues: Random Projection for Manifold Learning (2007)</ref>

### Introduction

For vector $x \in R^{N}$ and a Random matrix $R$, $Rx=y$ is called Random map (projection) of $x$, where $y\in R^{M}$. $y$ can be written as
$y= \sum_{i}x_{i}r_{i}$
where $r_{i}'s$ are the columns of $R$. In Random mapping orthogonal directions are replaced by random directions
By definition the number of columns of matrix $R$ called the Number of Random projections ($M$) required to project a K-dimensional manifold from $R^{N}$ into $R^{M}$
For Large $N$ applying adaptive (data dependent) algorithms like Isomap, LLE, etc are too much costly and not feasible so, it is reasonable to use random projection (non-adaptive method)to map data into lower dimension ($M$) and then apply clustering algorithms and also other dimensionality reduction algorithms, because Random Mapping almost preserves the metric structure. Specially, If $R$ is orthonormal then similarity under random mapping is exactly preserved.
For example if Cosine of the angle made by two vectors is the measure for their similarity, suppose $y_{1}$ and $y_{2}$ are the random maps of $x_{1}$ and $x_{2}$ then we have
$y^{T}_{1}y_{2}=x^{T}_{1}R^{T}Rx_{2}$
where $R$ can be written as
$R^{T}R=I+\epsilon$ in which

$\epsilon_{ij}= \left\{\begin{matrix} r^{T}_{i}r_{j} & \text{if } j \neq i \\ 0 & \text{if} i=j \end{matrix}\right.$
now for special case where $R$ is orthonormal,
$R^{T}R=I$
since $\epsilon_{ij}=r^{T}_{i}r_{j}=0$
so
$y^{T}_{1}y_{2}=x^{T}_{1}x_{2}$

As an interesting result about random matrix $R$ it can be shown that when $M$ (reduced dimension) is large then $\epsilon_{ij}$ are independently distributed as Normal distribution wit mean $0$ and Variance $1/M$

### Estimating The Intrinsic Dimension

Intrinsic Dimension ($K$) is an important input for almost all metric learning algorithms. Grossberger-Proccacia (GP) algorithm can help us to estimate $K$ using $n$ sample points selected from $K$-dimensional manifold embedded in $R^{N}$. Define

$C_{n}(r)=\frac{1}{n(n-1)} \sum_{i\neq j}I \left\|x_{i}-x_{j}\right\|\lt r$

which basically measures the probability of a random pairs of points havinng distances less than $r$. Then estimate of $K$ is obtained the following formula (known as the Scale-dependent Correlation Dimension of $X=(x_{1},x_{2},...,x_{n})$):

$\hat{D}_{corr}(r_{1},r_{2})=\frac{log(C_{n}(r_{1}))-log(C_{n}(r_{2}))}{log(r_{1})-log(r_{2})}$

when $r_{1}$ and $r_{2}$ are selected such that they cover the biggest range of graph at which $C_{n}(r)$ is linear, GP gives the best estimate. another point is that $\widehat{K}$ obtained GP algorithm based on finite sample points is Biased estimate.

The reason why the above formula works may be explained as follows: if the intrinsic dimensionality is 1 the number of points within a (1-dim) sphere around a given point would be proportional to $r$. If the intrinsic dimensionality is 2 and the points are (at least locally) distributed on a plane, then the number of points within a (2-dim) sphere around a given point would be proportional to $\pi r^2$, i.e., the area of the corresponding circle. Similarly in the case of a 3-dim sphere we will have proportionality to $4/3 \pi r^3$, and so on. So we if we take logarithm of this function, we expect the number of intrinsic dimensionality to be reflected in the slope of the resulting function, as implied by the the formula above.

### Upper and Lower Bound for estimator $K$

Baraniuk and Wakin (2007) showed that if the number of random projection $(M)$ is large we can be sure that the projection of a smooth manifold into lower dimensional space will be a near-isometric embedding. The point is that $M$ is only neede to have linear relation with $K$ and Logarithmic relationship with $N$ like
$M=c K log(N)$
Also we mentioned that, since random mapping almost preserves metric structure so we need to estimate the ID of the projected data.
Chinmay Hedg and colleagues (2007) showed that under some conditions, for random orthoprojector $R$ (consists of $M$ orthogonalized vectors of length $N$) and sequence of sample $X=\left\{x_{1},x_{2},...\right\}$ it can be shown that with probability $1-\rho$

$(1-\delta)\hat{K}\leq \hat{K}_{R}\leq (1+\delta)\hat{K}$

for fix $0 \lt \delta \lt 1$ and $0 \lt \rho \lt 1$

### Lower Bound for the number of Random Projections

We can also find the minimum number of projection needed so that, the difference between residuals obtained from applying Isomap on projected data and original data does not exceed an arbitary value.

Let $\Gamma=max_{1\leq i,j\leq n}d_{iso}(x_{i},x_{j})$ and call it Diameter of data set $X=\left\{x_{1},x_{2},...,x_{n}\right\}$ where $d_{iso}(x_{i},x_{j})$ is the Isomap estimate of geodesic distance, and also let $L$ and $L_{R}$ be the residuals obtained when Isomap generated $K$-dimentional embeding of $X$ and its progection $RX$, respectively. then we have,

$L_{R}\lt L+C(n)\Gamma^{2}\epsilon$

in which $C(n)$ is a function of the number of sample points.

### practical algorithm to determine $K$ and $M$

The above theorems are not able to give us the estimates without some prior information (for example about the volume and condition number of manifold) so mostly, they are called Theoretical Theorems. There is an algorithm to make it practically possible which is called ML-RP Algorithm.

let $M=1$ and select a random vector for $R$
While residual variance $\gt \delta$ do

 $(i)$ for $X=\left\{x_{1},x_{2},...,x_{n}\right\}$ estimate $K$ using GP algorithm.
$(ii)$ Use Isomap to produce an embedding in to $\hat{K}$-dimensional space.
$(iii)$ let $M=M+1$ and add one row to $R$.


EndWhile
Print $M$ and $\hat{K}$

### Applications

The main goal of dimension reduction methods is to preserve as much information as possible from the high dimension to lower dimensional space. But it is also important to consider the computational complexity of the algorithm. Random projections preserves the similarity between the data vectors when the data is projected in fewer dimensions and the projections can also be computed faster.
In [1] <ref> Ella Bingham and Heikki Mannila:Random projection in dimensionality reduction: Applications to image and text data. </ref> , they applied random projections technique of dimensonality reduction to image (both, noisy and noiseless) and text data and conclude that the random lower dimension subspace yields results comparable to conventional dimensional reduction method, such as PCA but the computational time is significantly reduced. They also demonstarte that using a sparse random matrix leads to a further savings in computational complexity. The benefit of random projections is applicable only in applications where the distance between the points in the higher dimension have some meaning but if the distances are themselves suspect, the there is no point in preserving them in low dimensional space. They provide an example," In a Euclidean space, every dimension is equally important and independent of the others, whereas e.g. in a process monitoring application some measured quantities (that is, dimensions) might be closely related to others,and the interpoint distances do not necessarily bear a clear meaning."
Random projections do not suffer from the curse of dimensionality.
Random projections also have application in the area of signal processing, where the device can perform tasks directly on random projections and thereby save storage and processing costs.

<references/>