Difference between revisions of "stat841"
(→Neural Networks (NN)  October 30, 2009) 
(→Neural Networks (NN)  October 30, 2009) 

Line 1,872:  Line 1,872:  
==Neural Networks (NN)  October 30, 2009 ==  ==Neural Networks (NN)  October 30, 2009 ==  
−  <math>\sum_{i=1}^  +  <math>\sum_{i=1}^k a_i\cdot\cfrac{\partial u_j}{\partial a_i}<\math> 
=== Backpropagation ===  === Backpropagation ===  
The idea is that we first feed an input (we can normalize the data before feeding) from the training set to the Neural Network, then find the error rate at the output and then we propagate the error to previous layers and for each edge of weight <math>\,u_{ij}</math> we find <math>\frac{\partial \mathrm{err}}{\partial u_{ij}}</math>. Having the error rates at hand we adjust the weight of each edge by taking steps proportional to the negative of the gradient to decrease the error at output. The next step is to apply the next input from the training set and go through the described adjustment procedure.  The idea is that we first feed an input (we can normalize the data before feeding) from the training set to the Neural Network, then find the error rate at the output and then we propagate the error to previous layers and for each edge of weight <math>\,u_{ij}</math> we find <math>\frac{\partial \mathrm{err}}{\partial u_{ij}}</math>. Having the error rates at hand we adjust the weight of each edge by taking steps proportional to the negative of the gradient to decrease the error at output. The next step is to apply the next input from the training set and go through the described adjustment procedure. 
Revision as of 01:37, 18 October 2011
Contents
 1 Proposal
 2 Mark your contribution here
 3 Scribe sign up
 4 Classfication2009.9.30
 5 Linear and Quadratic Discriminant Analysis  October 2,2009
 6 Linear and Quadratic Discriminant Analysis cont'd  October 5, 2009
 7 LDA and QDA in Matlab  October 7, 2009
 8 Trick: Using LDA to do QDA  October 7, 2009
 9 Introduction to Fisher's Discriminant Analysis  October 7, 2009
 10 Fisher's Discriminant Analysis (FDA)  October 9, 2009
 11 FDA for Multiclass Problems  October 14, 2009
 12 Linear Regression Models  October 14, 2009
 13 Logistic Regression October 16, 2009
 14 Logistic Regression(2)  October 19, 2009
 15 2009.10.21
 16 The Perceptron (Lecture October 23, 2009)
 17 Neural Networks (NN)  October 28, 2009
 18 Neural Networks (NN)  October 30, 2009
 19 Complexity Control October 30, 2009
 20 Complexity Control  Nov 2, 2009
 21 CrossValidation (CV)  Introduction
 22 Complexity Control  Nov 4, 2009
 23 Crossvalidation
 24 Regularization for Neural Network — Weight Decay
 25 Radial Basis Function (RBF) Networks  November 6, 2009
 26 Radial Basis Function (RBF) Networks  November 9th, 2009
 27 Model Selection(Stein's Unbiased Risk Estimate) November 11th, 2009
 28 SURE for RBF network & Support Vector Machine  November 13th, 2009
 29 Optimizing The Support Vector Machine  November 16th, 2009
 30 The Support Vector Machine algorithm  November 18, 2009
 31 Nonlinear hypersurfaces and NonSeparable classes  November 20, 2009
 32 Support Vector Machine algorithm for nonseparable cases  November 23, 2009
 33 Finishing up SVM  November 25, 2009
 34 Naive Bayes, Decision Trees, K Nearest Neighbours, Boosting, and Bagging  November 25, 2009
Proposal
Mark your contribution here
Scribe sign up
Classfication2009.9.30
Classification
With the rise of fields such as datamining, bioinformatics, and machine learning, classification has becomes a fastdeveloping topic. In the age of information, vast amounts of data are generated constantly, and the goal of classification is to learn from data. Potential application areas include handwritten post codes recognition, medical diagnosis, face recognition, human language processing and so on.
Definition: The problem of Prediction a discrete random variable [math] \mathcal{Y} [/math] from another random variable [math] \mathcal{X} [/math] is called Classification.
In classification,, we attempt to approximate a function [math]\,h[/math], by using a training data set, which will then be able to accurately classify new data inputs.
Given [math]\mathcal{X} \subset \mathbb{R}^{d}[/math], a subset of the [math]d[/math]dimensional real vectors and [math] \mathcal{Y} [/math], a finite set of labels, We try to determine a 'classification rule' [math]\,h[/math] such that,
 [math]\,h: \mathcal{X} \mapsto \mathcal{Y} [/math]
We use [math]\,n[/math] ordered pairs of training data which are identical independent distributions, [math]\,\{(X_{1},Y_{1}), (X_{2},Y_{2}), \dots , (X_{n},Y_{n})\}[/math] where [math]\,X_{i} \in \mathcal{X}[/math],[math]\,Y_{i} \in \mathcal{Y} [/math], to approximate [math]\,h[/math].
Thus, given a new input, [math]\,X \in \mathcal{X} [/math]
by using the classification rule we can predict a corresponding [math]\,\hat{Y}=h(X)[/math].
 Example Suppose we wish to classify fruits into apples and oranges by considering certain features of the fruit, for instance, color, diameter, and weight.
Let [math]\mathcal{X}= (\mathrm{colour}, \mathrm{diameter}, \mathrm{weight})[/math] and [math]\mathcal{Y}=\{\mathrm{apple}, \mathrm{orange}\}[/math]. The goal is to find a classification rule such that when a new fruit [math]\,X[/math] is presented based on its features, [math](\,X_{\mathrm{color}}, X_{\mathrm{diameter}}, X{_\mathrm{weight}})[/math], our classification rule [math]\,h[/math] can classify it as either an apple or an orange, i.e., [math]\,h(X_{\mathrm{color}}, X_{\mathrm{diameter}}, X_{\mathrm{weight}})[/math] be the fruit type of [math]\,X[/math].
Error rate
 The true error rate' [math]\,L(h)[/math] of a classifier having classification rule [math]\,h[/math] is defined as the probability that [math]\,h[/math] does not correctly classify any new data input, i.e., it is defined as [math]\,L(h)=P(h(X) \neq Y)[/math]. Here, [math]\,X \in \mathcal{X}[/math] and [math]\,Y \in \mathcal{Y}[/math] are the known feature values and the true class of that input, respectively.
 The empirical error rate (or training error rate) of a classifier having classification rule [math]\,h[/math] is defined as the frequency at which [math]\,h[/math] does not correctly classify the data inputs in the training set, i.e., it is defined as
[math]\,\hat{L}_{n} = \frac{1}{n} \sum_{i=1}^{n} I(h(X_{i}) \neq Y_{i})[/math], where [math]\,I[/math] is an indicator variable and [math]\,I = \left\{\begin{matrix} 1 &\text{if } h(X_i) \neq Y_i \\ 0 &\text{if } h(X_i) = Y_i \end{matrix}\right.[/math]. Here, [math]\,X_{i} \in \mathcal{X}[/math] and [math]\,Y_{i} \in \mathcal{Y}[/math] are the known feature values and the true class of the [math]\,i_th[/math] training input, respectively.
Bayes Classifier
The principle of Bayes Classifier is to calculate the posterior probability of a given object from its prior probability via Bayes formula, and then place the object in the class with the largest posterior probability<ref> http://www.wikicoursenote.com/wiki/Stat841f11#Bayes_Classifier </ref>
Intuitively speaking, to classify [math]\,x\in \mathcal{X}[/math] we find [math]y \in \mathcal{Y}[/math] such that [math]\,P(Y=yX=x)[/math] is maximum over all the members of [math]\mathcal{Y}[/math].
Mathematically, for [math]\,k[/math] classes and given object [math]\,X=x[/math], we find [math]\,y\in \mathcal{Y}[/math] which maximizes [math]\,P(Y=yX=x)[/math], and classify [math]\,X[/math] into class [math]\,y[/math]. In order to calculate the value of [math]\,P(Y=yX=x)[/math], we use Bayes formula
 [math] \begin{align} P(Y=yX=x) &= \frac{P(X=xY=y)P(Y=y)}{P(X=x)} \\ &=\frac{P(X=xY=y)P(Y=y)}{\Sigma_{\forall y \in \mathcal{Y}}P(X=xY=y)P(Y=y)} \end{align} [/math]
where [math]\,P(Y=yX=x)[/math] is referred to as the posterior probability, [math]\,P(Y=y)[/math] as the prior probability, [math]\,P(X=xY=y)[/math] as the likelihood, and [math]\,P(X=x)[/math] as the evidence.
For the special case that [math]\,Y[/math] has only two classes, that is, [math]\, \mathcal{Y}=\{0, 1\}[/math]. Consider the probability that [math]\,r(X)=P\{Y=1X=x\}[/math]. Given [math]\,X=x[/math], By Bayes formula, we have
 [math] \begin{align} r(X)&=P(Y=1X=x) \\ &=\frac{P(X=xY=1)P(Y=1)}{P(X=x)}\\ &=\frac{P(X=xY=1)P(Y=1)}{P(X=xY=1)P(Y=1)+P(X=xY=0)P(Y=0)} \end{align} [/math]
Definition:
The Bayes classification rule [math]\,h[/math] is_{}
 [math]\, h(X)= \left\{\begin{matrix} 1 & r(x)\gt \frac{1}{2} \\ 0 & \mathrm{otherwise} \end{matrix}\right.[/math]
3 different approaches to classification:
1) Empirical Risk Minimization: Choose a set fo classifier [math]\mathcal{H}[/math] and find [math]\,h^*\in \mathcal{H}[/math] that minimizes some estimate of [math]\,L(h)[/math]
2) Regression: Find an estimate [math] (\hat r) [/math] of the function [math] r [/math] and define
 [math]\, h(X)= \left\{\begin{matrix} 1 & \hat r(x)\gt \frac{1}{2} \\ 0 & \mathrm{otherwise} \end{matrix}\right.[/math]
3) Density Estimation: estimate [math]\,P(X=xY=0)[/math] and [math]\,P(X=xY=1)[/math] (less popular in highdimension cases)
Bayes Classification Rule Optimality Theorem: The Bayes rule is optimal in true error rate, that is for any other classification rule [math]\, \overline{h}[/math], we have [math]\,L(h) \le L(\overline{h})[/math]. Intuitively speaking this theorem is saying we cannot do better than classifying [math]\,x\in \mathcal{X}[/math] to [math]\,y[/math] when[math][/math] the probability of being of type [math]\,y[/math] for [math]\,x[/math] is more than probability of being any other type.
Definition:
The set [math]\,D(h)=\{x: P(Y=1X=x)=P(Y=0X=x)\}[/math] is called the decision boundary.
 [math]\, h^*(X)= \left\{\begin{matrix} 1 & if P(Y=1X=x)\gt P(Y=0X=x) \\ 0 & \mathrm{otherwise} \end{matrix}\right.[/math]
Remark:
1)Bayes classification rule is optimal. Proof:[1]
2)We still need any other method, since we cannot define prior probability in realistic.
Example:
We’re going to predict if a particular student will pass STAT441/841.
We have data on past student performance. For each student we know:
If student’s GPA > 3.0 (G)
If student had a strong math background (M)
If student is a hard worker (H)
If student passed or failed course
[math]\, \mathcal{Y}= \{ 0,1 \} [/math], where 1 refers to pass and 0 refers to fail. Assume that [math]\,P(Y=1)=P(Y=0)=0.5[/math]
For a new student comes along with values [math]\,G=0, M=1, H=0[/math], we calculate [math]\,r(X)=P(Y=1X=(0,1,0))[/math] as
[math]\,r(X)=P(Y=1X=(0,1,0))=\frac{P(X=(0,1,0)Y=1)P(Y=1)}{P(X=(0,1,0)Y=1)P(Y=1)+P(X=(0,1,0)Y=0)P(Y=0)}=\frac{0.025}{0.125}=0.2\lt \frac{1}{2}[/math]
Thus, we classify the new student into class 0, namely, we predict him to fail in this course.
 Notice: Although the Bayes rule is optimal, we still need other methods, since it is generally impossible for us to know the prior [math]\,P(Y=1)[/math], and class conditional density [math]\,P(X=xY=1)[/math] and ultimately calculate the value of [math]\,r(X)[/math], which makes Bayes rule inconvenient in practice.
Currently, there are four primary classifier based on Bayes Classifier: Naive Bayes classifier[2], treeaugmented naive Bayes (TAN), Bayesian network augmented naive Bayes (BAN) and general Bayesian network (GBN).
useful link:Decision Theory, Bayes Classifier
Bayesian vs. Frequentist
Intuitively, to solve a twoclass problem, we may have the following two approaches:
1) If [math]\,P(Y=1X=x)\gt P(Y=0X=x)[/math], then [math]\,h(x)=1[/math], otherwise [math]\,h(x)=0[/math].
2) If [math]\,P(X=xY=1)\gt P(X=xY=0)[/math], then [math]\,h(x)=1[/math], otherwise [math]\,h(x)=0[/math].
One obvious difference between these two methods is that the first one considers probability as changing based on observation while the second one considers probablity as having objective existence. Actually, they represent two different schools in statistics.
During the history of statistics, there are two major classification methods : Bayesian and frequentist. The two methods represent two different ways of thoughts and hold different view to define probability. The followings are the main differences between Bayes and Frequentist.
Frequentist
 Probability is objective.
 Data is a repeatable random sample(there is a frequency).
 Parameters are fixed and unknown constant.
 Not applicable to single event. For example, a frequentist cannot predict the weather of tomorrow because tomorrow is only one unique event, and cannot be referred to a frequency in a lot of samples.
Bayesian
 Probability is subjective.
 Data are fixed.
 Parameters are unknown and random variables that have a given distribution and other probability statements can be made about them.
 Can be applied to single events based on degree of confidence or beliefs. For example, Bayesian can predict tomorrow's weather, such as having the probability of [math]\,50%[/math] of rain.
Example
Suppose there is a man named Jack. In Bayesian method, at first, one can see this man (object), and then judge whether his name is Jack (label). On the other hand, in Frequentist method, one doesn’t see the man (object), but can see the photos (label) of this man to judge whether he is Jack.
Linear and Quadratic Discriminant Analysis  October 2,2009
Introduction
Notation
Let us first introduce some new notation for the following sections.
Multiclass Classification:
Y takes on more than two values.
Recall that in the discussion of the Bayes Classifier, we introduced Bayes Formula:
 [math] \begin{align} P(Y=yX=x) &=\frac{P(X=xY=y)P(Y=y)}{\Sigma_{\forall y \in \mathcal{Y}}P(X=xY=y)P(Y=y)} \end{align} [/math]
We will use new labels for the following equivalent formula:
 [math] \begin{align} P(Y=kX=x) &=\frac{f_k(x)\pi_k}{\Sigma_kf_k(x)\pi_k} \end{align} [/math]
 [math]\,f_k[/math] is called the class conditional density; also referred to previously as the likelihood function. Essentially, this is the function that allows us to reason about a parameter given a certain outcome.
 [math]\,\pi_k[/math] is called the prior probability. This is a probability distribution that represents what we know (or believe we know) about a population.
 [math]\,\Sigma_k[/math] is the sum with respect to all [math]\,k[/math] classes.
Approaches
Representing the optimal method, Bayes classifier cannot be used in the most practical situations though, since usually the prior probability is unknown. Fortunately, other methods of classification have been evolved. These methods fall into three general categories.
1 Empirical Risk Minimization:Choose a set fo classifier [math]\mathcal{H}[/math] and find [math]\,h^* \epsilon H[/math], minimize some estimate of [math]\,L(H)[/math].
2 Regression:Find an estimate [math] (\hat r) [/math] of the function [math]\ r [/math] and deifne
 [math]\, h(X)= \left\{\begin{matrix} 1 & \hat r(x)\gt \frac{1}{2} \\ 0 & \mathrm{otherwise} \end{matrix}\right.[/math]
3 Density estimation, estimate [math]\ P(X = xY = 0)[/math] and [math]\ P(X = xY = 1)[/math]
Note:
The third approach, in this form, is not popular because density estimation doesn't work very well with dimension greater than 2. However this approach is the simplest and we can assume a parametric model for the densities.
Linear Discriminate Analysis and Quadratic Discriminate Analysis are examples of the third approach, density estimation.
LDA
Motivation
The Bayes classifier is optimal. Unfortunately, the prior and conditional density of most data is not known. Some estimation of these should be made if we want to classify some data.
The simplest way to achieve this is to assume that all the class densities are approximately a multivariate normal distribution, find the parameters of each such distribution, and use them to calculate the conditional density and prior for unknown points, thus approximating the Bayesian classifier to choose the most likely class. In addition, if the covariance of each class density is assumed to be the same, the number of unknown parameters is reduced and the model is easy to fit and use, as seen later.
History
The name Linear Discriminant Analysis comes from the fact that these simplifications produce a linear model, which is used to discriminate between classes. In many cases, this simple model is sufficient to provide a near optimal classification  for example, the ZScore credit risk model, designed by Edward Altman in 1968, which is essentially a weighted LDA, revisited in 2000, has shown an 8590% success rate predicting bankruptcy, and is still in use today.
Purpose
1 feature selection
2 which classification rule best seperate the classes
Definition
To perform LDA we make two assumptions.
 The clusters belonging to all classes each follow a multivariate normal distribution.
[math]x \in \mathbb{R}^d[/math] [math]f_k(x)=\frac{1}{ (2\pi)^{d/2}\Sigma_k^{1/2} }\exp\left( \frac{1}{2} [x  \mu_k]^\top \Sigma_k^{1} [x  \mu_k] \right)[/math]
where [math]\ f_k(x)[/math] is a class conditional density
 Simplification Assumption: Each cluster has the same covariance matrix [math]\,\Sigma[/math] equal to the covariance of [math]\Sigma_k \forall k[/math].
We wish to solve for the decision boundary where the error rates for classifying a point are equal, where one side of the boundary gives a lower error rate for one class and the other side gives a lower error rate for the other class.
So we solve [math]\,r_k(x)=r_l(x)[/math] for all the pairwise combinations of classes.
[math]\,\Rightarrow Pr(Y=kX=x)=Pr(Y=lX=x)[/math]
[math]\,\Rightarrow \frac{Pr(X=xY=k)Pr(Y=k)}{Pr(X=x)}=\frac{Pr(X=xY=l)Pr(Y=l)}{Pr(X=x)}[/math] using Bayes' Theorem
[math]\,\Rightarrow Pr(X=xY=k)Pr(Y=k)=Pr(X=xY=l)Pr(Y=l)[/math] by canceling denominators
[math]\,\Rightarrow f_k(x)\pi_k=f_l(x)\pi_l[/math]
[math]\,\Rightarrow \frac{1}{ (2\pi)^{d/2}\Sigma^{1/2} }\exp\left( \frac{1}{2} [x  \mu_k]^\top \Sigma^{1} [x  \mu_k] \right)\pi_k=\frac{1}{ (2\pi)^{d/2}\Sigma^{1/2} }\exp\left( \frac{1}{2} [x  \mu_l]^\top \Sigma^{1} [x  \mu_l] \right)\pi_l[/math]
[math]\,\Rightarrow \exp\left( \frac{1}{2} [x  \mu_k]^\top \Sigma^{1} [x  \mu_k] \right)\pi_k=\exp\left( \frac{1}{2} [x  \mu_l]^\top \Sigma^{1} [x  \mu_l] \right)\pi_l[/math] Since both [math]\Sigma[/math] are equal based on the assumptions specific to LDA.
[math]\,\Rightarrow \frac{1}{2} [x  \mu_k]^\top \Sigma^{1} [x  \mu_k] + \log(\pi_k)=\frac{1}{2} [x  \mu_l]^\top \Sigma^{1} [x  \mu_l] +\log(\pi_l)[/math] taking the log of both sides.
[math]\,\Rightarrow \log(\frac{\pi_k}{\pi_l})\frac{1}{2}\left( x^\top\Sigma^{1}x + \mu_k^\top\Sigma^{1}\mu_k  2x^\top\Sigma^{1}\mu_k  x^\top\Sigma^{1}x  \mu_l^\top\Sigma^{1}\mu_l + 2x^\top\Sigma^{1}\mu_l \right)=0[/math] by expanding out
[math]\,\Rightarrow \log(\frac{\pi_k}{\pi_l})\frac{1}{2}\left( \mu_k^\top\Sigma^{1}\mu_k\mu_l^\top\Sigma^{1}\mu_l  2x^\top\Sigma^{1}(\mu_k\mu_l) \right)=0[/math] after canceling out like terms and factoring.
We can see that this is a linear function in [math]\ x [/math] with general form [math]\,ax+b=0[/math].
Actually, this linear log function shows that the decision boundary between class [math]\ k [/math] and class [math]\ l [/math], i.e. [math]\ P(G=kX=x)=P(G=lX=x) [/math], is linear in [math]\ x[/math]. Given any pair of classes, decision boundaries are always linear. In [math]\ d[/math] dimensions, we separate regions by hyperplanes.
In the special case where the number of samples from each class are equal ([math]\,\pi_k=\pi_l[/math]), the boundary surface or line lies halfway between [math]\,\mu_l[/math] and [math]\,\mu_k[/math]
Limitation
 LDA implicitly assumes Gaussian distribution of data.
 LDA implicitly assumes that the mean is the discriminating factor, not variance.
 LDA may overfit the data.
QDA
The concept uses a same idea as LDA of finding a boundary where the error rate for classification between classes are equal, except the assumption that each cluster has the same variance [math]\,\Sigma[/math] equal to the mean variance of [math]\Sigma_k \forall k[/math] is removed. We can use a hypothesis test with [math]\ H_0 [/math]: [math]\Sigma_k \forall k [/math]=[math]\,\Sigma[/math].The best method is likelihood ratio test.
Following along from where QDA diverges from LDA.
[math]\,f_k(x)\pi_k=f_l(x)\pi_l[/math]
[math]\,\Rightarrow \frac{1}{ (2\pi)^{d/2}\Sigma_k^{1/2} }\exp\left( \frac{1}{2} [x  \mu_k]^\top \Sigma_k^{1} [x  \mu_k] \right)\pi_k=\frac{1}{ (2\pi)^{d/2}\Sigma_l^{1/2} }\exp\left( \frac{1}{2} [x  \mu_l]^\top \Sigma_l^{1} [x  \mu_l] \right)\pi_l[/math]
[math]\,\Rightarrow \frac{1}{\Sigma_k^{1/2} }\exp\left( \frac{1}{2} [x  \mu_k]^\top \Sigma_k^{1} [x  \mu_k] \right)\pi_k=\frac{1}{\Sigma_l^{1/2} }\exp\left( \frac{1}{2} [x  \mu_l]^\top \Sigma_l^{1} [x  \mu_l] \right)\pi_l[/math] by cancellation
[math]\,\Rightarrow \frac{1}{2}\log(\Sigma_k)\frac{1}{2} [x  \mu_k]^\top \Sigma_k^{1} [x  \mu_k]+\log(\pi_k)=\frac{1}{2}\log(\Sigma_l)\frac{1}{2} [x  \mu_l]^\top \Sigma_l^{1} [x  \mu_l]+\log(\pi_l)[/math] by taking the log of both sides
[math]\,\Rightarrow \log(\frac{\pi_k}{\pi_l})\frac{1}{2}\log(\frac{\Sigma_k}{\Sigma_l})\frac{1}{2}\left( x^\top\Sigma_k^{1}x + \mu_k^\top\Sigma_k^{1}\mu_k  2x^\top\Sigma_k^{1}\mu_k  x^\top\Sigma_l^{1}x  \mu_l^\top\Sigma_l^{1}\mu_l + 2x^\top\Sigma_l^{1}\mu_l \right)=0[/math] by expanding out
[math]\,\Rightarrow \log(\frac{\pi_k}{\pi_l})\frac{1}{2}\log(\frac{\Sigma_k}{\Sigma_l})\frac{1}{2}\left( x^\top(\Sigma_k^{1}\Sigma_l^{1})x + \mu_k^\top\Sigma_k^{1}\mu_k  \mu_l^\top\Sigma_l^{1}\mu_l  2x^\top(\Sigma_k^{1}\mu_k\Sigma_l^{1}\mu_l) \right)=0[/math] this time there are no cancellations, so we can only factor
The final result is a quadratic equation specifying a curved boundary between classes with general form [math]\,ax^2+bx+c=0[/math].
It is quadratic because there is no boundaries.
Linear and Quadratic Discriminant Analysis cont'd  October 5, 2009
Linear discriminant analysis[3] is a statistical method used to find the linear combination of features which best separate two or more classes of objects or events. It is widely applied in classifying diseases, positioning, product management, and marketing research.
Quadratic Discriminant Analysis[4], on the other had, aims to find the quadratic combination of features. It is more general than Linear discriminant analysis. Unlike LDA however, in QDA there is no assumption that the covariance of each of the classes is identical. When the assumption is true, the best possible test for the hypothesis that a given measurement is from a given class is the likelihood ratio test. Suppose the means of each class are known to be [math] \mu_{y=0},\mu_{y=1} [/math] and the covariances [math] \Sigma_{y=0}, \Sigma_{y=1} [/math]. Then the likelihood ratio will be given by
 Likelihood ratio = [math] \frac{ \sqrt{2 \pi \Sigma_{y=1}}^{1} \exp \left( \frac{1}{2}(x\mu_{y=1})^T \Sigma_{y=1}^{1} (x\mu_{y=1}) \right) }{ \sqrt{2 \pi \Sigma_{y=0}}^{1} \exp \left( \frac{1}{2}(x\mu_{y=0})^T \Sigma_{y=0}^{1} (x\mu_{y=0}) \right)} \lt t [/math]
for some threshold t. After some rearrangement, it can be shown that the resulting separating surface between the classes is a quadratic.
Summarizing LDA and QDA
We can summarize what we have learned on LDA and QDA so far into the following theorem.
Theorem:
Suppose that [math]\,Y \in \{1,\dots,k\}[/math], if [math]\,f_k(x) = Pr(X=xY=k)[/math] is Gaussian, the Bayes Classifier rule is
 [math]\,h(X) = \arg\max_{k} \delta_k(x)[/math]
where
 [math] \,\delta_k =  \frac{1}{2}log(\Sigma_k)  \frac{1}{2}(x\mu_k)^\top\Sigma_k^{1}(x\mu_k) + log (\pi_k) [/math] (quadratic)
 Note The decision boundary between classes [math]k[/math] and [math]l[/math] is quadratic in [math]x[/math].
If the covariance of the Gaussians are the same, this becomes
 [math] \,\delta_k = x^\top\Sigma^{1}\mu_k  \frac{1}{2}\mu_k^\top\Sigma^{1}\mu_k + log (\pi_k) [/math] (linear)
 Note [math]\,\arg\max_{k} \delta_k(x)[/math]returns the set of k for which [math]\,\delta_k(x)[/math] attains its largest value.
In practice
We need to estimate the prior, so in order to do this, we use the sample estimates of [math]\,\pi,\mu_k,\Sigma_k[/math] in place of the true values, i.e.
[math]\,\hat{\pi_k} = \hat{Pr}(y=k) = \frac{n_k}{n}[/math]
[math]\,\hat{\mu_k} = \frac{1}{n_k}\sum_{i:y_i=k}x_i[/math]
[math]\,\hat{\Sigma_k} = \frac{1}{n_k}\sum_{i:y_i=k}(x_i\hat{\mu_k})(x_i\hat{\mu_k})^\top[/math]
In the case where we have a common covariance matrix, we get the ML estimate to be
[math]\,\Sigma=\frac{\sum_{r=1}^{k}(n_r\Sigma_r)}{\sum_{l=1}^{k}(n_l)} [/math]
This is a Maximum Likelihood estimate.
Computation
Case 1: (Example) [math]\, \Sigma_k = I [/math]'
This means that the data is distributed symmetrically around the center [math]\mu[/math], i.e. the isocontours are all circles.
We have:
[math] \,\delta_k =  \frac{1}{2}log(I)  \frac{1}{2}(x\mu_k)^\top I(x\mu_k) + log (\pi_k) [/math]
We see that the first term in the above equation, [math]\,\frac{1}{2}log(I)[/math], is zero since [math]\ I [/math] is the determine and [math]\ I=1 [/math]. The second term contains [math]\, (x\mu_k)^\top I(x\mu_k) = (x\mu_k)^\top(x\mu_k) [/math], which is the squared Euclidean distance between [math]\,x[/math] and [math]\,\mu_k[/math]. Therefore we can find the distance between a point and each center and adjust it with the log of the prior, [math]\,log(\pi_k)[/math]. The class that has the minimum distance will maximise [math]\,\delta_k[/math]. According to the theorem, we can then classify the point to a specific class [math]\,k[/math]. In addition, [math]\, \Sigma_k = I [/math] implies that our data is spherical.
Case 2: (General Case) [math]\, \Sigma_k \ne I [/math]
We can decompose this as:
[math] \, \Sigma_k = USV^\top = USU^\top [/math] (In general when [math]\,X=USV^\top[/math], [math]\,U[/math] is the eigenvectors of [math]\,XX^T[/math] and [math]\,V[/math] is the eigenvectors of [math]\,X^\top X[/math]. So if [math]\, X[/math] is symmetric, we will have [math]\, U=V[/math]. Here [math]\, \Sigma [/math] is symmetric)
and the inverse of [math]\,\Sigma_k[/math] is
[math] \, \Sigma_k^{1} = (USU^\top)^{1} = (U^\top)^{1}S^{1}U^{1} = US^{1}U^\top [/math] (since [math]\,U[/math] is orthonormal)
So from the formula for [math]\,\delta_k[/math], the second term is
 [math]\begin{align} (x\mu_k)^\top\Sigma_k^{1}(x\mu_k)&= (x\mu_k)^\top US^{1}U^T(x\mu_k)\\ & = (U^\top xU^\top\mu_k)^\top S^{1}(U^\top xU^\top \mu_k)\\ & = (U^\top xU^\top\mu_k)^\top S^{\frac{1}{2}}S^{\frac{1}{2}}(U^\top xU^\top\mu_k) \\ & = (S^{\frac{1}{2}}U^\top xS^{\frac{1}{2}}U^\top\mu_k)^\top I(S^{\frac{1}{2}}U^\top xS^{\frac{1}{2}}U^\top \mu_k) \\ & = (S^{\frac{1}{2}}U^\top xS^{\frac{1}{2}}U^\top\mu_k)^\top(S^{\frac{1}{2}}U^\top xS^{\frac{1}{2}}U^\top \mu_k) \\ \end{align} [/math]
where we have the Euclidean distance between [math] \, S^{\frac{1}{2}}U^\top x [/math] and [math]\, S^{\frac{1}{2}}U^\top\mu_k[/math].
A transformation of all the data points can be done from [math]\,x[/math] to [math]\,x^*[/math] where [math] \, x^* \leftarrow S^{\frac{1}{2}}U^\top x [/math].
It is now possible to do classification with [math]\,x^*[/math], treating it as in Case 1 above.
Note that when we have multiple classes, they must all have the same transformation, else, ahead of time we would have to assume a data point belongs to one class or the other. All classes therefore need to have the same shape for classification to be applicable using this method. So this method works for LDA.
If the classes have different shapes, in another word, have different covariance [math]\,\Sigma_k[/math], can we use the same method to transform all data points [math]\,x[/math] to [math]\,x^*[/math]?
The answer is NO. Consider that you have two classes with different shapes, then consider transforming them to the same shape. Given a data point, justify which class this point belongs to. The question is, which transformation can you use? For example, if you use the transformation of class A, then you have assumed that this data point belongs to class A.
Kernel QDA In real life, QDA is always better fit the data then LDA because QDA relaxes does not have the assumption made by LDA that the covariance matrix for each class is identical. However, QDA still assumes that the class conditional distribution is Gaussian which does not be the real case in practical. Another methodkernel QDA does not have the Gaussian distribution assumption and it works better.
The Number of Parameters in LDA and QDA
Both LDA and QDA require us to estimate parameters. The more estimation we have to do, the less robust our classification algorithm will be.
LDA: Since we just need to compare the differences between one given class and remaining [math]\,K1[/math] classes, totally, there are [math]\,K1[/math] differences. For each of them, [math]\,a^{T}x+b[/math] requires [math]\,d+1[/math] parameters. Therefore, there are [math]\,(K1)\times(d+1)[/math] parameters.
QDA: For each of the differences, [math]\,x^{T}ax + b^{T}x + c[/math] requires [math]\frac{1}{2}(d+1)\times d + d + 1 = \frac{d(d+3)}{2}+1[/math] parameters. Therefore, there are [math](K1)(\frac{d(d+3)}{2}+1)[/math] parameters.
related link:
LDA:[5]
Regularized linear discriminant analysis and its application in microarrays
MATHEMATICAL OPERATIONS OF LDA
Application in face recognition and in market
QDA:[7]
LDA and QDA in Matlab  October 7, 2009
We have examined the theory behind Linear Discriminant Analysis (LDA) and Quadratic Discriminant Analysis (QDA) above; how do we use these algorithms in practice? Matlab offers us a function called classify
that allows us to perform LDA and QDA quickly and easily.
In class, we were shown an example of using LDA and QDA on the 2_3 data that is used in the first assignment. The code below reproduces that example, slightly modified, and explains each step.
>> load 2_3; >> [U, sample] = princomp(X'); >> sample = sample(:,1:2);
 First, we do principal component analysis (PCA) on the 2_3 data to reduce the dimensionality of the original data from 64 dimensions to 2. Doing this makes it much easier to visualize the results of the LDA and QDA algorithms.
>> plot (sample(1:200,1), sample(1:200,2), '.'); >> hold on; >> plot (sample(201:400,1), sample(201:400,2), 'r.');
 Recall that in the 2_3 data, the first 200 elements are images of the number two handwritten and the last 200 elements are images of the number three handwritten. This code sets up a plot of the data such that the points that represent a 2 are blue, while the points that represent a 3 are red.
 Before using
classify
we can set up a vector that contains the actual labels for our data, to train the classification algorithm. If we don't know the labels for the data, then the element in thegroup
vector should be an empty string orNaN
. (See grouping data for more information.)
>> group = ones(400,1); >> group(201:400) = 2;
 We can now classify our data.
>> [class, error, POSTERIOR, logp, coeff] = classify(sample, sample, group, 'linear');
 The full details of this line can be examined in the Matlab help file linked above. What we care about are
class
, which contains the labels that the algorithm thinks that each data point belongs to, andcoeff
, which contains information about the line that algorithm created to separate the data into each class.
 We can see the efficacy of the algorithm by comparing
class
togroup
.
>> sum (class==group) ans = 369
 This compares the value in
class
to the value ingroup
. The answer of 369 tells us that the algorithm correctly determined the class of the point 369 times, out of a possible 400 data points. This gives us an empirical error rate of 0.0775.
 We can see the line produced by LDA using
coeff
.
>> k = coeff(1,2).const; >> l = coeff(1,2).linear; >> f = sprintf('0 = %g+%g*x+%g*y', k, l(1), l(2)); >> ezplot(f, [min(sample(:,1)), max(sample(:,1)), min(sample(:,2)), max(sample(:,2))]);
 Those familiar with the programming language C will find the
sprintf
line refreshingly familiar; those with no exposure to C are directed to Matlab'ssprintf
page. Essentially, this code sets up the equation of the line in the form0 = a + bx + cy
. We then use theezplot
function to plot the line.
 Let's perform the same steps, except this time using QDA. The main difference with QDA is a slightly different call to
classify
, and a more complicated procedure to plot the line.
>> [class, error, POSTERIOR, logp, coeff] = classify(sample, sample, group, 'quadratic'); >> sum (class==group) ans = 371 >> k = coeff(1,2).const; >> l = coeff(1,2).linear; >> q = coeff(1,2).quadratic; >> f = sprintf('0 = %g+%g*x+%g*y+%g*x^2+%g*x.*y+%g*y.^2', k, l, q(1,1), q(1,2)+q(2,1), q(2,2)); >> ezplot(f, [min(sample(:,1)), max(sample(:,1)), min(sample(:,2)), max(sample(:,2))]);
classify
can also be used with other discriminant analysis algorithms. The steps laid out above would only need to be modified slightly for those algorithms.
Recall: An analysis of the function of princomp
in matlab.
In our assignment 1, we have learnt that how to perform Principal Component Analysis using SVD method. In fact, the matlab offers us a function called princomp
which can perform PCA conveniently. From the matlab help file on princomp
, you can find the details about this function. But here we will analyze the code of the function of princomp()
in matlab to find something different when comparing with SVD method. The following is the code of princomp and explanations to some emphasized steps.
function [pc, score, latent, tsquare] = princomp(x); % PRINCOMP Principal Component Analysis (centered and scaled data). % [PC, SCORE, LATENT, TSQUARE] = PRINCOMP(X) takes a data matrix X and % returns the principal components in PC, the socalled Zscores in SC % ORES, the eigenvalues of the covariance matrix of X in LATENT, % and Hotelling's Tsquared statistic for each data point in TSQUARE. % Reference: J. Edward Jackson, A User's Guide to Principal Components % John Wiley & Sons, Inc. 1991 pp. 125. % B. Jones 31794 % Copyright 19932002 The MathWorks, Inc. % $Revision: 2.9 $ $Date: 2002/01/17 21:31:45 $ [m,n] = size(x); % get the lengh of the rows and columns of matrix x. r = min(m1,n); % max possible rank of X avg = mean(x); % the mean of every column of X centerx = (x  avg(ones(m,1),:)); % centers X by subtracting off column means [U,latent,pc] = svd(centerx./sqrt(m1),0); % "economy size" decomposition score = centerx*pc; % the representation of X in the principal component space if nargout < 3 return; end latent = diag(latent).^2; if (r latent = [latent(1:r); zeros(nr,1)]; score(:,r+1:end) = 0; end if nargout < 4 return; end tmp = sqrt(diag(1./latent(1:r)))*score(:,1:r)'; tsquare = sum(tmp.*tmp)';
From the above code, we should pay attention to the following aspects when comparing with SVD method:
First, Rows of [math]\,X[/math] correspond to observations, columns to variables. When using princomp on 2_3 data in assignment 1, note that we take the transpose of [math]\,X[/math].
>> load 2_3; >> [U, score] = princomp(X');
Second, princomp centers X by subtracting off column means.
The third, when [math]\,X=UdV'[/math], princomp uses [math]\,V[/math] as coefficients for principal components, rather than [math]\,U[/math].
The following is an example to perform PCA using princomp and SVD respectively to get the same results.
 SVD method
>> load 2_3 >> mn=mean(X,2); >> X1=Xrepmat(mn,1,400); >> [s d v]=svd(X1'); >> y=X1'*v;
 princomp
>>[U score]=princomp(X');
Then we can see that y=score, v=U.
useful resouces: LDA and QDA in Matlab[8],[9],[10]
Trick: Using LDA to do QDA  October 7, 2009
There is a trick that allows us to use the linear discriminant analysis (LDA) algorithm to generate as its output a quadratic function that can be used to classify data. This trick is similar to, but more primitive than, the Kernel trick that will be discussed later in the course.
Essentially, the trick involves adding one or more new features (i.e. new dimensions) that just contain our original data projected to that dimension. We then do LDA on our new higherdimensional data. The answer provided by LDA can then be collapsed onto a lower dimension, giving us a quadratic answer.
Motivation
Why would we want to use LDA over QDA? In situations where we have fewer data points, LDA turns out to be more robust.
If we look back at the equations for LDA and QDA, we see that in LDA we must estimate [math]\,\mu_1[/math], [math]\,\mu_2[/math] and [math]\,\Sigma[/math]. In QDA we must estimate all of those, plus another [math]\,\Sigma[/math]; the extra [math]\,\frac{d(d1)}{2}[/math] estimations make QDA less robust with fewer data points.
Theoretically
Suppose we can estimate some vector [math]\underline{w}^T[/math] such that
[math]y = \underline{w}^Tx[/math]
where [math]\underline{w}[/math] is a ddimensional column vector, and [math]x\ \epsilon\ \mathbb{R}^d[/math] (vector in d dimensions).
We also have a nonlinear function [math]g(x) = y = x^Tvx + \underline{w}^Tx[/math] that we cannot estimate.
Using our trick, we create two new vectors, [math]\,\underline{w}^*[/math] and [math]\,x^*[/math] such that:
[math]\underline{w}^{*T} = [w_1,w_2,...,w_d,v_1,v_2,...,v_d][/math]
and
[math]x^{*T} = [x_1,x_2,...,x_d,{x_1}^2,{x_2}^2,...,{x_d}^2][/math]
We can then estimate a new function, [math]g^*(x,x^2) = y^* = \underline{w}^{*T}x^*[/math].
Note that we can do this for any [math]x[/math] and in any dimension; we could extend a [math]D \times n[/math] matrix to a quadratic dimension by appending another [math]D \times n[/math] matrix with the original matrix squared, to a cubic dimension with the original matrix cubed, or even with a different function altogether, such as a [math]\,sin(x)[/math] dimension.
By Example
Let's use our trick to do a quadratic analysis of the 2_3 data using LDA.
>> load 2_3; >> [U, sample] = princomp(X'); >> sample = sample(:,1:2);
 We start off the same way, by using PCA to reduce the dimensionality of our data to 2.
>> X_star = zeros(400,4); >> X_star(:,1:2) = sample(:,:); >> for i=1:400 for j=1:2 X_star(i,j+2) = X_star(i,j)^2; end end
 This projects our sample into two more dimensions by squaring our initial two dimensional data set.
>> group = ones(400,1); >> group(201:400) = 2; >> [class, error, POSTERIOR, logp, coeff] = classify(X_star, X_star, group, 'linear'); >> sum (class==group) ans = 375
 We can now display our results.
>> k = coeff(1,2).const; >> l = coeff(1,2).linear; >> f = sprintf('0 = %g+%g*x+%g*y+%g*(x)^2+%g*(y)^2', k, l(1), l(2),l(3),l(4)); >> ezplot(f,[min(sample(:,1)), max(sample(:,1)), min(sample(:,2)), max(sample(:,2))]);
 Not only does LDA give us a better result than it did previously, it actually beats QDA, which only correctly classified 371 data points for this data set. Continuing this procedure by adding another two dimensions with [math]x^4[/math] (i.e. we set
X_star(i,j+2) = X_star(i,j)^4
) we can correctly classify 376 points.
Introduction to Fisher's Discriminant Analysis  October 7, 2009
Fisher's Discriminant Analysis (FDA), also known as Fisher's Linear Discriminant Analysis (LDA) in some sources, is a classical feature extraction technique. It was originally described in 1936 by Sir Ronald Aylmer Fisher, an English statistician and eugenicist who has been described as one of the founders of modern statistical science. His original paper describing FDA can be found here; a Wikipedia article summarizing the algorithm can be found here.
LDA is for classification and FDA is used for feature extraction.
Contrasting FDA with PCA
The goal of FDA is in contrast to our other main feature extraction technique, principal component analysis (PCA).
 In PCA, we map data to lower dimensions to maximize the variation in those dimensions.
 In FDA, we map data to lower dimensions to best separate data in different classes.
Because we are concerned with identifying which class data belongs to, FDA is often a better feature extraction algorithm for classification.
Another difference between PCA and FDA is that FDA is a supervised algorithm; that is, we know what class data belongs to, and we exploit that knowledge to find a good projection to lower dimensions.
Intuitive Description of FDA
An intuitive description of FDA can be given by visualizing two clouds of data, as shown above. Ideally, we would like to collapse all of the data points in each cloud onto one point on some projected line, then make those two points as far apart as possible. In doing so, we make it very easy to tell which class a data point belongs to. In practice, it is not possible to collapse all of the points in a cloud to one point, but we attempt to make all of the points in a cloud close to each other while simultaneously far from the points in the other cloud.
Example in R
>> X = matrix(nrow=400,ncol=2) >> X[1:200,] = mvrnorm(n=200,mu=c(1,1),Sigma=matrix(c(1,1.5,1.5,3),2)) >> X[201:400,] = mvrnorm(n=200,mu=c(5,3),Sigma=matrix(c(1,1.5,1.5,3),2)) >> Y = c(rep("red",200),rep("blue",200))
 Create 2 multivariate normal random variables with [math]\, \mu_1 = \left( \begin{array}{c}1 \\ 1 \end{array} \right), \mu_2 = \left( \begin{array}{c}5 \\ 3 \end{array} \right). ~\textrm{Cov} = \left( \begin{array}{cc} 1 & 1.5 \\ 1.5 & 3 \end{array} \right)[/math]. Create
Y
, an index indicating which class they belong to.
>> s < svd(X,nu=1,nv=1)
 Calculate the singular value decomposition of X. The most significant direction is in
s$v[,1]
, and is displayed as a black line.
>> s2 < lda(X,grouping=Y)
 The
lda
function, given the group for each item, uses Fischer's Linear Discriminant Analysis (FLDA) to find the most discriminant direction. This can be found ins2$scaling
.
Now that we've calculated the PCA and FLDA decompositions, we create a plot to demonstrate the differences between the two algorithms. FLDA is clearly better suited to discriminating between two classes whereas PCA is primarily good for reducing the number of dimensions when data is highdimensional.
>> plot(X,col=Y,main="PCA vs. FDA example")
 Plot the set of points, according to colours given in Y.
>> slope = s$v[2]/s$v[1] >> intercept = mean(X[,2])slope*mean(X[,1]) >> abline(a=intercept,b=slope)
 Plot the main PCA direction, drawn through the mean of the dataset. Only the direction is significant.
>> slope2 = s2$scaling[2]/s2$scaling[1] >> intercept2 = mean(X[,2])slope2*mean(X[,1]) >> abline(a=intercept2,b=slope2,col="red")
 Plot the FLDA direction, again through the mean.
>> legend(2,7,legend=c("PCA","FDA"),col=c("black","red"),lty=1)
 Labeling the lines directly on the graph makes it easier to interpret.
Distance Metric Learning VS FDA
In many fundamental machine learning problems, the Euclidean distances between data points do not represent the desired topology that we are trying to capture. Kernel methods address this problem by mapping the points into new spaces where Euclidean distances may be more useful. An alternative approach is to construct a Mahalanobis distance (quadratic Gaussian metric) over the input space and use it in place of Euclidean distances. This approach can be equivalently interpreted as a linear transformation of the original inputs,followed by Euclidean distance in the projected space. This approach has attracted a lot of recent interest.
Some of the proposed algorithms are iterative and computationally expensive. In the paper,"Distance Metric Learning VS FDA " written by our instructor, they propose a closedform solution to one algorithm that previously required expensive semidefinite optimization. They provide a new problem setup in which the algorithm performs better or as well as some standard methods, but without the computational complexity. Furthermore, they show a strong relationship between these methods and the Fisher Discriminant Analysis (FDA). They also extend the approach by kernelizing it, allowing for nonlinear transformations of the metric.
Fisher's Discriminant Analysis (FDA)  October 9, 2009
The goal of FDA is to reduce the dimensionality of data in order to have separable data points in a new space. We can consider two kinds of problems:
 2class problem
 multiclass problem
Twoclass problem
In the twoclass problem, we have the preknowledge that data points belong to two classes. Intuitively speaking points of each class form a cloud around the mean of the class, with each class having possibly different size. To be able to separate the two classes we must determine the class whose mean is closest to a given point while also accounting for the different size of each class, which is represented by the covariance of each class.
Assume [math]\underline{\mu_{1}}=\frac{1}{n_{1}}\displaystyle\sum_{i:y_{i}=1}\underline{x_{i}}[/math] and [math]\displaystyle\Sigma_{1}[/math], represent the mean and covariance of the 1st class, and [math]\underline{\mu_{2}}=\frac{1}{n_{2}}\displaystyle\sum_{i:y_{i}=2}\underline{x_{i}}[/math] and [math]\displaystyle\Sigma_{2}[/math] represent the mean and covariance of the 2nd class. We have to find a transformation which satisfies the following goals:
1.To make the means of these two classes as far apart as possible
 In other words, the goal is to maximize the distance after projection between class 1 and class 2. This can be done by maximizing the distance between the means of the classes after projection. When projecting the data points to a onedimensional space, all points will be projected to a single line; the line we seek is the one with the direction that achieves maximum separation of classes upon projetion. If the original points are [math]\underline{x_{i}} \in \mathbb{R}^{d}[/math]and the projected points are [math]\underline{w}^T \underline{x_{i}}[/math] then the mean of the projected points will be [math]\underline{w}^T \underline{\mu_{1}}[/math] and [math]\underline{w}^T \underline{\mu_{2}}[/math] for class 1 and class 2 respectively. The goal now becomes to maximize the Euclidean distance between projected means, [math](\underline{w}^T\underline{\mu_{1}}\underline{w}^T\underline{\mu_{2}})^T (\underline{w}^T\underline{\mu_{1}}\underline{w}^T\underline{\mu_{2}})[/math]. The steps of this maximization are given below.
2.We want to collapse all data points of each class to a single point, i.e., minimize the covariance within classes
 Notice that the variance of the projected classes 1 and 2 are given by [math]\underline{w}^T\Sigma_{1}\underline{w}[/math] and [math]\underline{w}^T\Sigma_{2}\underline{w}[/math]. The second goal is to minimize the sum of these two covariances.
As is demonstrated below, both of these goals can be accomplished simultaneously.
Original points are [math]\underline{x_{i}} \in \mathbb{R}^{d}[/math]
[math]\ \{ \underline x_1 \underline x_2 \cdot \cdot \cdot \underline x_n \} [/math]
Projected points are [math]\underline{z_{i}} \in \mathbb{R}^{1}[/math] with [math]\underline{z_{i}} = \underline{w}^T \cdot\underline{x_{i}}[/math] [math]\ z_i [/math] is a sclar
Between class covariance
In this particular case, we want to project all the data points in one dimensional space.
We want to maximize the Euclidean distance between projected means, which is
 [math] \begin{align} (\underline{w}^T \underline{\mu_{1}}  \underline{w}^T \underline{\mu_{2}})^T(\underline{w}^T \underline{\mu_{1}}  \underline{w}^T \underline{\mu_{2}}) &= (\underline{\mu_{1}}\underline{\mu_{2}})^T\underline{w} . \underline{w}^T(\underline{\mu_{1}}\underline{\mu_{2}})\\ &= \underline{w}^T(\underline{\mu_{1}}\underline{\mu_{2}})(\underline{\mu_{1}}\underline{\mu_{2}})^T\underline{w} \end{align} [/math] which is scalar
The quantity [math](\underline{\mu_{1}}\underline{\mu_{2}})(\underline{\mu_{1}}\underline{\mu_{2}})^T[/math] is called between class covariance or [math]\,S_{B}[/math].
The goal is to maximize : [math]\underline{w}^T S_{B} \underline{w}[/math]
Within class covariance
Covariance of class 1 is [math]\,\Sigma_{1}[/math] Covariance of class 2 is [math]\,\Sigma_{2}[/math] So covariance of projected points will be [math]\,\underline{w}^T \Sigma_{1} \underline{w}[/math] and [math]\underline{w}^T \Sigma_{2} \underline{w}[/math]
If we sum this two quantities we have
 [math] \begin{align} \underline{w}^T \Sigma_{1} \underline{w} + \underline{w}^T \Sigma_{2} \underline{w} &= \underline{w}^T(\Sigma_{1} + \Sigma_{2})\underline{w} \end{align} [/math]
The quantity [math]\,(\Sigma_{1} + \Sigma_{2})[/math] is called within class covariance or [math]\,S_{W}[/math]
The goal is to minimize [math]\underline{w}^T S_{W} \underline{w}[/math]
Objective Function
Instead of maximizing [math]\underline{w}^T S_{B} \underline{w}[/math] and minimizing [math]\underline{w}^T S_{W} \underline{w}[/math] we can define the following objective function:
 [math]\underset{\underline{w}}{max}\ \frac{\underline{w}^T S_{B} \underline{w}}{\underline{w}^T S_{W} \underline{w}}[/math]
This maximization problem is equivalent to [math]\underset{\underline{w}}{max}\ \underline{w}^T S_{B} \underline{w} \equiv \max(\underline w^T S_B \underline w)[/math] subject to constraint [math]\underline{w}^T S_{W} \underline{w} = 1[/math], where [math]\ \underline w^T S_B \underline w[/math] is no upper bound and [math]\ \underline w^T S_w \underline w[/math] is no lower bound.
We can use the Lagrange multiplier method to solve it:
 [math]L(\underline{w},\lambda) = \underline{w}^T S_{B} \underline{w}  \lambda(\underline{w}^T S_{W} \underline{w}  1)[/math] where [math]\ \lambda [/math] is the weight
With [math]\frac{\part L}{\part \underline{w}} = 0[/math] we get:
 [math] \begin{align} &\Rightarrow\ 2\ S_{B}\ \underline{w}\  2\lambda\ S_{W}\ \underline{w}\ = 0\\ &\Rightarrow\ S_{B}\ \underline{w}\ =\ \lambda\ S_{W}\ \underline{w} \\ &\Rightarrow\ S_{W}^{1}\ S_{B}\ \underline{w}\ =\ \lambda\ \underline{w} \end{align} [/math]
Note that [math]\, S_{W}=\Sigma_1+\Sigma_2[/math] is sum of two positive matrices and so it has an inverse.
Here [math]\underline{w}[/math] is the eigenvector of [math]S_{w}^{1}\ S_{B}[/math] corresponding to the largest eigenvalue [math]\ \lambda [/math].
In facts, this expression can be simplified even more.
 [math]\Rightarrow\ S_{w}^{1}\ S_{B}\ \underline{w}\ =\ \lambda\ \underline{w}[/math] with [math]S_{B}\ =\ (\underline{\mu_{1}}\underline{\mu_{2}})(\underline{\mu_{1}}\underline{\mu_{2}})^T[/math]
 [math]\Rightarrow\ S_{w}^{1}\ (\underline{\mu_{1}}\underline{\mu_{2}})(\underline{\mu_{1}}\underline{\mu_{2}})^T \underline{w}\ =\ \lambda\ \underline{w}[/math]
The quantity [math](\underline{\mu_{1}}\underline{\mu_{2}})^T \underline{w}[/math] and [math]\lambda[/math] are scalars.
So we can say the quantity [math]S_{w}^{1}\ (\underline{\mu_{1}}\underline{\mu_{2}})[/math] is proportional to [math]\underline{w}[/math]
FDA vs. PCA Example in Matlab
We can compare PCA and FDA through the figure produced by matlab.
The following are the code to produce the figure step by step and the explanation for steps.
>>X1=mvnrnd([1,1],[1 1.5;1.5 3],300); >>X2=mvnrnd([5,3],[1 1.5;1.5 3],300); >>X=[X1;X2];
 Create two multivariate normal random variables with [math]\, \mu_1 = \left( \begin{array}{c}1 \\ 1 \end{array} \right), \mu_2 = \left( \begin{array}{c}5 \\ 3 \end{array} \right). ~\textrm{Cov} = \left( \begin{array}{cc} 1 & 1.5 \\ 1.5 & 3 \end{array} \right)[/math].
>>plot(X(1:300,1),X(1:300,2),'.'); >>hold on >>p1=plot(X(301:600,1),X(301:600,2),'r.');
 Plot the the data of the two classes respectively.
>>[U Y]=princomp(X); >>plot([0 U(1,1)*10],[0 U(2,1)*10]);
 Using PCA to find the principal component and plot it.
>>sw=2*[1 1.5;1.5 3]; >>sb=([1; 1][5 ;3])*([1; 1][5; 3])'; >>g =inv(sw)*sb; >>[v w]=eigs(g); >>plot([v(1,1)*5 0],[v(2,1)*5 0],'r')
 Using FDA to find the principal component and plot it.
Now we can compare them through the figure.
From the graph: when we see using PCA, we have a huge overlap for two classes, so PCA is not good. However, there is no overlap for the two classes and they are seperated pretty. Thus, FDA is better than PCA here.
Practical example of 2_3
In this matlab example we explore FDA using our familiar data set 2_3 which consists of 200 handwritten "2" and 200 handwritten "3".
X is a matrix of size 64*400 and each column represents an 8*8 image of "2" or "3". Here X1 gets all "2" and X2 gets all "3".
>>load 2_3 >>X1 = X(:, 1:200); >>X2 = X(:, 201:400);
Next we calculate within class covariance and between class covariance as before.
>>mu1 = mean(X1, 2); >>mu2 = mean(X2, 2); >>sb = (mu1  mu2) * (mu1  mu2)'; >>sw = cov(X1') + cov(X2');
We use the first two eigenvectors to project the dato in a twodimensional space.
>>[v d] = eigs( inv(sw) * sb ); >>w = v(:, 1:2); >>X_hat = w'*X;
Finally we plot the data and visualize the effect of FDA.
>> scatter(ones(1,200),X_hat(1:200)) >> hold on >> scatter(ones(1,200),X_hat(201:400),'r')
Map the data into a linear line, and the two classes are seperated perfectly here.
An extension of Fisher's discriminant analysis for stochastic processes
A general notion of Fisher's linear discriminant analysis can extend the classical multivariate concept to situations that allow for functionvalued random elements. The development uses a bijective mapping that connects a second order process to the reproducing kernel Hilbert space generated by its within class covariance kernel. This approach provides a seamless transition between Fisher's original development and infinite dimensional settings that lends itself well to computation via smoothing and regularization.
Link for Algorithm introduction:[[11]]
FDA for Multiclass Problems  October 14, 2009
FDA method for Multiclass Problems
For the [math]k[/math]class problem, we need to find a projection from [math]d[/math]dimensional space to a [math](k1)[/math]dimensional space.
(It is more reasonable to have at least 2 directions)
Basically, the within class covariance matrix [math]\mathbf{S}_{W}[/math] is easily to obtain:
 [math] \begin{align} \mathbf{S}_{W} = \sum_{i=1}^{k} \mathbf{S}_{W,i} \end{align} [/math]
where [math]\mathbf{S}_{W,i} = \frac{1}{n_{i}}\sum_{j: y_{j}=i}(\mathbf{x}_{j}  \mathbf{\mu}_{i})(\mathbf{x}_{j}  \mathbf{\mu}_{i})^{T}[/math] and [math]\mathbf{\mu}_{i} = \frac{\sum_{j: y_{j}=i}\mathbf{x}_{j}}{n_{i}}[/math].
However, the between class covariance matrix [math]\mathbf{S}_{B}[/math] is not easy to obtain. One of the simplifications is that we may assume that the total covariance [math]\mathbf{S}_{T}[/math] of the data is constant, since [math]\mathbf{S}_{W}[/math] is easy to compute, we can get [math]\mathbf{S}_{B}[/math] using the following relationship:
 [math] \begin{align} \mathbf{S}_{B} = \mathbf{S}_{T}  \mathbf{S}_{W} \end{align} [/math]
Actually, there is another generation for [math]\mathbf{S}_{B}[/math]. Denote a total mean vector [math]\mathbf{\mu}[/math] by
 [math] \begin{align} \mathbf{\mu} = \frac{1}{n}\sum_{i}\mathbf{x_{i}} = \frac{1}{n}\sum_{j=1}^{k}n_{j}\mathbf{\mu}_{j} \end{align} [/math]
Thus the total covariance matrix [math]\mathbf{S}_{T}[/math] is
 [math] \begin{align} \mathbf{S}_{T} = \sum_{i}(\mathbf{x_{i}\mu})(\mathbf{x_{i}\mu})^{T} \end{align} [/math]
Thus we obtain
 [math] \begin{align} & \mathbf{S}_{T} = \sum_{i=1}^{k}\sum_{j: y_{j}=i}(\mathbf{x}_{j}  \mathbf{\mu}_{i} + \mathbf{\mu}_{i}  \mathbf{\mu})(\mathbf{x}_{j}  \mathbf{\mu}_{i} + \mathbf{\mu}_{i}  \mathbf{\mu})^{T} \\& = \sum_{i=1}^{k}\sum_{j: y_{j}=i}(\mathbf{x}_{j}\mathbf{\mu}_{i})(\mathbf{x}_{j}\mathbf{\mu}_{i})^{T}+ \sum_{i=1}^{k}\sum_{j: y_{j}=i}(\mathbf{\mu}_{i}\mathbf{\mu})(\mathbf{\mu}_{i}\mathbf{\mu})^{T} \\& = \mathbf{S}_{W} + \sum_{i=1}^{k} n_{i}(\mathbf{\mu}_{i}\mathbf{\mu})(\mathbf{\mu}_{i}\mathbf{\mu})^{T} \end{align} [/math]
Since the total covariance [math]\mathbf{S}_{T}[/math] is the sum of the within class covariance [math]\mathbf{S}_{W}[/math] and the between class covariance [math]\mathbf{S}_{B}[/math], we can denote the second term as the general between class covariance matrix [math]\mathbf{S}_{B}[/math], thus we obtain
 [math] \begin{align} \mathbf{S}_{B} = \sum_{i=1}^{k} n_{i}(\mathbf{\mu}_{i}\mathbf{\mu})(\mathbf{\mu}_{i}\mathbf{\mu})^{T} \end{align} [/math]
Therefore,
 [math] \begin{align} \mathbf{S}_{T} = \mathbf{S}_{W} + \mathbf{S}_{B} \end{align} [/math]
Recall that in the two class case problem, we have
 [math] \begin{align} & \mathbf{S}_{B^{\ast}} = (\mathbf{\mu}_{1}\mathbf{\mu}_{2})(\mathbf{\mu}_{1}\mathbf{\mu}_{2})^{T} \\ & = (\mathbf{\mu}_{1}\mathbf{\mu}+\mathbf{\mu}\mathbf{\mu}_{2})(\mathbf{\mu}_{1}\mathbf{\mu}+\mathbf{\mu}\mathbf{\mu}_{2})^{T} \\ & = ((\mathbf{\mu}_{1}\mathbf{\mu})(\mathbf{\mu}_{2}\mathbf{\mu}))((\mathbf{\mu}_{1}\mathbf{\mu})(\mathbf{\mu}_{2}\mathbf{\mu}))^{T} \\ & = (\mathbf{\mu}_{1}\mathbf{\mu})(\mathbf{\mu}_{1}\mathbf{\mu})^{T}+(\mathbf{\mu}_{2}\mathbf{\mu})(\mathbf{\mu}_{2}\mathbf{\mu})^{T} \end{align} [/math]
From the general form,
 [math] \begin{align} & \mathbf{S}_{B} = n_{1}(\mathbf{\mu}_{1}\mathbf{\mu})(\mathbf{\mu}_{1}\mathbf{\mu})^{T} + n_{2}(\mathbf{\mu}_{2}\mathbf{\mu})(\mathbf{\mu}_{2}\mathbf{\mu})^{T} \end{align} [/math]
Apparently, they are very similar.
Now, we are trying to find the optimal transformation. Basically, we have
 [math] \begin{align} \mathbf{z}_{i} = \mathbf{W}^{T}\mathbf{x}_{i}, i=1,2,...,k1 \end{align} [/math]
where [math]\mathbf{z}_{i}[/math] is a [math](k1)\times 1[/math] vector, [math]\mathbf{W}[/math] is a [math]d\times (k1)[/math] transformation matrix, i.e. [math]\mathbf{W} = [\mathbf{w}_{1}, \mathbf{w}_{2},..., \mathbf{w}_{k1}][/math], and [math]\mathbf{x}_{i}[/math] is a [math]d\times 1[/math] column vector.
Thus we obtain
 [math] \begin{align} & \mathbf{S}_{W}^{\ast} = \sum_{i=1}^{k}\sum_{j: y_{j}=i}(\mathbf{W}^{T}\mathbf{x}_{j}\mathbf{W}^{T}\mathbf{\mu}_{i})(\mathbf{W}^{T}\mathbf{x}_{j}\mathbf{W}^{T}\mathbf{\mu}_{i})^{T} \\ & = \sum_{i=1}^{k}\sum_{j: y_{j}=i}\mathbf{W}^{T}(\mathbf{x}_{j}\mathbf{\mu}_{i})(\mathbf{x}_{j}\mathbf{\mu}_{i})\mathbf{W} \\ & = \mathbf{W}^{T}\left[\sum_{i=1}^{k}\sum_{j: y_{j}=i}(\mathbf{x}_{j}\mathbf{\mu}_{i})(\mathbf{x}_{j}\mathbf{\mu}_{i})\right]\mathbf{W} \\ & = \mathbf{W}^{T}\mathbf{S}_{W}\mathbf{W} \end{align} [/math]
Similarly, we obtain
 [math] \begin{align} & \mathbf{S}_{B}^{\ast} = \sum_{i=1}^{k}n_{i}(\mathbf{W}^{T}\mathbf{\mu}_{i}\mathbf{W}^{T}\mathbf{\mu})(\mathbf{W}^{T}\mathbf{\mu}_{i}\mathbf{W}^{T}\mathbf{\mu})^{T} \\ & = \sum_{i=1}^{k}n_{i}\mathbf{W}^{T}(\mathbf{\mu}_{i}\mathbf{\mu})(\mathbf{\mu}_{i}\mathbf{\mu})^{T}\mathbf{W} \\ & = \mathbf{W}^{T}\left[ \sum_{i=1}^{k}n_{i}(\mathbf{\mu}_{i}\mathbf{\mu})(\mathbf{\mu}_{i}\mathbf{\mu})^{T}\right]\mathbf{W} \\ & = \mathbf{W}^{T}\mathbf{S}_{B}\mathbf{W} \end{align} [/math]
Now, we use the determinant of the matrix, i.e. the product of the eigenvalues of the matrix, as our measure.
 [math] \begin{align} \phi(\mathbf{W}) = \frac{\mathbf{S}_{B}^{\ast}}{\mathbf{S}_{W}^{\ast}} = \frac{\mathbf{W}^{T}\mathbf{S}_{B}\mathbf{W}}{\mathbf{W}^{T}\mathbf{S}_{W}\mathbf{W}} \end{align} [/math]
The solution for this question is that the columns of the transformation matrix [math]\mathbf{W}[/math] are exactly the eigenvectors that correspond to largest [math]k1[/math] eigenvalues with respect to
 [math] \begin{align} \mathbf{S}_{W}^{1}\mathbf{S}_{B}\mathbf{w}_{i} = \lambda_{i}\mathbf{w}_{i} \end{align} [/math]
Also, note that we can use
 [math] \begin{align} \sum_{i=1}^{k}n_{i}\(\mathbf{W}^{T}\mathbf{\mu}_{i}\mathbf{W}^{T}\mathbf{\mu})^{T}\^{2} \end{align} [/math]
as our measure.
Recall that
 [math] \begin{align} \\mathbf{X}\^2 = Tr(\mathbf{X}^{T}\mathbf{X}) \end{align} [/math]
Thus we obtain that
 [math] \begin{align} & \sum_{i=1}^{k}n_{i}\(\mathbf{W}^{T}\mathbf{\mu}_{i}\mathbf{W}^{T}\mathbf{\mu})^{T}\^{2} \\ & = \sum_{i=1}^{k}n_{i}Tr[(\mathbf{W}^{T}\mathbf{\mu}_{i}\mathbf{W}^{T}\mathbf{\mu})(\mathbf{W}^{T}\mathbf{\mu}_{i}\mathbf{W}^{T}\mathbf{\mu})^{T}] \\ & = Tr[\sum_{i=1}^{k}n_{i}(\mathbf{W}^{T}\mathbf{\mu}_{i}\mathbf{W}^{T}\mathbf{\mu})(\mathbf{W}^{T}\mathbf{\mu}_{i}\mathbf{W}^{T}\mathbf{\mu})^{T}] \\ & = Tr[\sum_{i=1}^{k}n_{i}\mathbf{W}^{T}(\mathbf{\mu}_{i}\mathbf{\mu})(\mathbf{\mu}_{i}\mathbf{\mu})^{T}\mathbf{W}] \\ & = Tr[\mathbf{W}^{T}\sum_{i=1}^{k}n_{i}(\mathbf{\mu}_{i}\mathbf{\mu})(\mathbf{\mu}_{i}\mathbf{\mu})^{T}\mathbf{W}] \\ & = Tr[\mathbf{W}^{T}\mathbf{S}_{B}\mathbf{W}] \end{align} [/math]
Similarly, we can get [math]Tr[\mathbf{W}^{T}\mathbf{S}_{W}\mathbf{W}][/math]. Thus we have following criterion function
 [math] \begin{align} \phi(\mathbf{W}) = \frac{Tr[\mathbf{W}^{T}\mathbf{S}_{B}\mathbf{W}]}{Tr[\mathbf{W}^{T}\mathbf{S}_{W}\mathbf{W}]} \end{align} [/math]
Similar to the two class case problem, we have:
max [math]Tr[\mathbf{W}^{T}\mathbf{S}_{B}\mathbf{W}][/math] subject to [math]Tr[\mathbf{W}^{T}\mathbf{S}_{W}\mathbf{W}]=1[/math]
To solve this optimization problem a Lagrange multiplier [math]\Lambda[/math], which actually is a [math]d \times d[/math] diagonal matrix, is introduced:
 [math] \begin{align} L(\mathbf{W},\Lambda) = Tr[\mathbf{W}^{T}\mathbf{S}_{W}\mathbf{B}]  \Lambda\left\{ Tr[\mathbf{W}^{T}\mathbf{S}_{W}\mathbf{W}]  1 \right\} \end{align} [/math]
Differentiating with respect to [math]\mathbf{W}[/math] we obtain:
 [math] \begin{align} \frac{\partial L}{\partial \mathbf{W}} = (\mathbf{S}_{B} + \mathbf{S}_{B}^{T})\mathbf{W}  \Lambda (\mathbf{S}_{W} + \mathbf{S}_{W}^{T})\mathbf{W} \end{align} [/math]
Note that the [math]\mathbf{S}_{B}[/math] and [math]\mathbf{S}_{W}[/math] are both symmetric matrices, thus set the first derivative to zero, we obtain:
 [math] \begin{align} \mathbf{S}_{B}\mathbf{W}  \Lambda\mathbf{S}_{W}\mathbf{W}=0 \end{align} [/math]
Thus,
 [math] \begin{align} \mathbf{S}_{B}\mathbf{W} = \Lambda\mathbf{S}_{W}\mathbf{W} \end{align} [/math]
where
 [math] \mathbf{\Lambda} = \begin{pmatrix} \lambda_{1} & & 0\\ &\ddots&\\ 0 & &\lambda_{d} \end{pmatrix} [/math]
and [math]\mathbf{W} = [\mathbf{w}_{1}, \mathbf{w}_{2},..., \mathbf{w}_{k1}][/math].
As a matter of fact, [math]\mathbf{\Lambda}[/math] must have [math]\mathbf{k1}[/math] nonzero eigenvalues, because [math]rank({S}_{W}^{1}\mathbf{S}_{B})=k1[/math].
Therefore, the solution for this question is as same as the previous case. The columns of the transformation matrix [math]\mathbf{W}[/math] are exactly the eigenvectors that correspond to largest [math]k1[/math] eigenvalues with respect to
 [math] \begin{align} \mathbf{S}_{W}^{1}\mathbf{S}_{B}\mathbf{w}_{i} = \lambda_{i}\mathbf{w}_{i} \end{align} [/math]
Generalization of Fisher's Linear Discriminant
Fisher's linear discriminant (Fisher, 1936) is very popular among users of discriminant analysis.Some of the reasons for this are its simplicity and unnecessity of strict assumptions. However it has optimality properties only if the underlying distributions of the groups are multivariate normal. It is also easy to verify that the discriminant rule obtained can be very harmed by only a small number of outlying observations. Outliers are very hard to detect in multivariate data sets and even when they are detected simply discarding them is not the most effcient way of handling the situation. Therefore the need for robust procedures that can accommodate the outliers and are not strongly affected by them. Then, a generalization of Fisher's linear discriminant algorithm [[12]]is developed to lead easily to a very robust procedure.
Linear Regression Models  October 14, 2009
Regression analysis is a general statistical technique for modelling and analyzing how a dependent variable changes according to changes in independent variables. In classification, we are interested in how a label, [math]\,y[/math], changes according to changes in [math]\,X[/math].
We will start by considering a very simple regression model, the linear regression model.
General information on linear regression can be found at the University of South Florida and this MIT lecture.
For the purpose of classification, the linear regression model assumes that the regression function [math]\,E(YX)[/math] is linear in the inputs [math]\,\mathbf{x}_{1}, ..., \mathbf{x}_{p}[/math].
The simple linear regression model has the general form:
 [math] \begin{align} f(x) = \beta^{T}\mathbf{x}_{i}+\beta_{0} \end{align} [/math]
where [math]\,\beta[/math] is a [math]1 \times d[/math] vector and [math]\ x_i [/math] is a [math]d \times 1[/math] vector .
Given input data [math]\,\mathbf{x}_{1}, ..., \mathbf{x}_{p}[/math] and [math]\,y_{1}, ..., y_{p}[/math] our goal is to find [math]\,\beta[/math] and [math]\,\beta_0[/math] such that the linear model fits the data while minimizing sum of squared errors using the Least Squares method.
Note that vectors [math]\mathbf{x}_{i}[/math] could be numerical inputs, transformations of the original data, i.e. [math]\log \mathbf{x}_{i}[/math] or [math]\sin \mathbf{x}_{i}[/math], or basis expansions, i.e. [math]\mathbf{x}_{i}^{2}[/math] or [math]\mathbf{x}_{i}\times \mathbf{x}_{j}[/math].
Denote [math]\mathbf{X}[/math] as a [math]n\times(d+1)[/math] matrix with each row an input vector (with 1 in the first position), [math]\,\beta = (\beta_0, \beta_1,..., \beta_{d})^{T}[/math] and [math]\mathbf{y}[/math] as a [math]n \times 1[/math] vector of outputs. We then try to minimize the residual sumofsquares
 [math] \begin{align} \mathrm{RSS}(\beta)=(\mathbf{y}\mathbf{X}\beta)(\mathbf{y}\mathbf{X}\beta)^{T} \end{align} [/math]
This is a quadratic function in the [math]\,d+1[/math] parameters. Differentiating with respect to [math]\,\beta[/math] we obtain
 [math] \begin{align} \frac{\partial \mathrm{RSS}}{\partial \beta} = 2\mathbf{X}^{T}(\mathbf{y}\mathbf{X}\beta) \end{align} [/math]
 [math] \begin{align} \frac{\partial^{2}\mathrm{RSS}}{\partial \beta \partial \beta^{T}}=2\mathbf{X}^{T}\mathbf{X} \end{align} [/math]
Set the first derivative to zero
 [math] \begin{align} \mathbf{X}^{T}(\mathbf{y}\mathbf{X}\beta)=0 \end{align} [/math]
we obtain the solution
 [math] \begin{align} \hat \beta = (\mathbf{X}^{T}\mathbf{X})^{1}\mathbf{X}^{T}\mathbf{y} \end{align} [/math]
Thus the fitted values at the inputs are
 [math] \begin{align} \mathbf{\hat y} = \mathbf{X}\hat\beta = \mathbf{X} (\mathbf{X}^{T}\mathbf{X})^{1}\mathbf{X}^{T}\mathbf{y} \end{align} [/math]
where [math]\mathbf{H} = \mathbf{X} (\mathbf{X}^{T}\mathbf{X})^{1}\mathbf{X}^{T}[/math] is called the hat matrix.
 Note For classification purposes, this is not a correct model. Recall the following application of Bayes classifier:
[math]r(x)= P( Y=k  X=x )= \frac{f_{k}(x)\pi_{k}}{\Sigma_{k}f_{k}(x)\pi_{k}}[/math]
It is clear that to make sense mathematically, [math]\displaystyle r(x)[/math] must be a value between 0 and 1. If this is estimated with the
regression function [math]\displaystyle r(x)=E(YX=x)[/math] and [math]\mathbf{\hat\beta} [/math] is learned as above, then there is nothing that would restrict [math]\displaystyle r(x)[/math] to taking values between 0 and 1. This is more direct approach to classification since it do not need to estimate [math]\ f_k(x) [/math] and [math]\ \pi_k [/math].
[math]\ 1 \times P(Y=1X=x)+0 \times P(Y=0X=x)=E(YX) [/math]
This model does not classify Y between 0 and 1, so it is not good and sometimes it can lead to a decent classifier. [math]\ y_i=\frac{1}{n_1} [/math] [math]\ \frac{1}{n_2} [/math]
A linear regression example in Matlab
We can see how linear regression works through the following example in Matlab. The following is the code and the explanation for each step.
Again, we use the data in 2_3.m.
>>load 2_3; >>[U, sample] = princomp(X'); >>sample = sample(:,1:2);
We carry out Principal Component Analysis (PCA) to reduce the dimensionality from 64 to 2.
>>y = zeros(400,1); >>y(201:400) = 1;
We let y represent the set of labels coded as 0 and 1.
>>x=[sample;ones(1,400)];
Construct x by adding a row of vector 1 to data.
>>b=inv(x*x')*x*y;
Calculate b, which represents [math]\beta[/math] in the linear regression model.
>>x1=x'; >>for i=1:400 if x1(i,:)*b>0.5 plot(x1(i,1),x1(i,2),'.') hold on elseif x1(i,:)*b < 0.5 plot(x1(i,1),x1(i,2),'r.') end end
Plot the fitted y values.
Comments about Linear regression model
Linear regression model is almost the easiest and most popular way to analyze the relationship of different data sets. However, it has some disadvantages as well as its advantages. We should be clear about them before we apply the model.
Advantages: Linear least squares regression has earned its place as the primary tool for process modeling because of its effectiveness and completeness. Though there are types of data that are better described by functions that are nonlinear in the parameters, many processes in science and engineering are welldescribed by linear models. This is because either the processes are inherently linear or because, over short ranges, any process can be wellapproximated by a linear model. The estimates of the unknown parameters obtained from linear least squares regression are the optimal estimates from a broad class of possible parameter estimates under the usual assumptions used for process modeling. Practically speaking, linear least squares regression makes very efficient use of the data. Good results can be obtained with relatively small data sets. Finally, the theory associated with linear regression is wellunderstood and allows for construction of different types of easilyinterpretable statistical intervals for predictions, calibrations, and optimizations. These statistical intervals can then be used to give clear answers to scientific and engineering questions.
Disadvantages: The main disadvantages of linear least squares are limitations in the shapes that linear models can assume over long ranges, possibly poor extrapolation properties, and sensitivity to outliers. Linear models with nonlinear terms in the predictor variables curve relatively slowly, so for inherently nonlinear processes it becomes increasingly difficult to find a linear model that fits the data well as the range of the data increases. As the explanatory variables become extreme, the output of the linear model will also always more extreme. This means that linear models may not be effective for extrapolating the results of a process for which data cannot be collected in the region of interest. Of course extrapolation is potentially dangerous regardless of the model type. Finally, while the method of least squares often gives optimal estimates of the unknown parameters, it is very sensitive to the presence of unusual data points in the data used to fit a model. One or two outliers can sometimes seriously skew the results of a least squares analysis. This makes model validation, especially with respect to outliers, critical to obtaining sound answers to the questions motivating the construction of the model.
Logistic Regression October 16, 2009
The logistic regression model arises from the desire to model the posterior probabilities of the [math]\displaystyle K[/math] classes via linear functions in [math]\displaystyle x[/math], while at the same time ensuring that they sum to one and remain in [0,1].Logistic regression models are usually fit by maximum likelihood, using the conditional likelihood ,using [math]\displaystyle Pr(YX)[/math]. Since [math]\displaystyle Pr(YX)[/math] completely specifies the conditional distribution, the multinomial distribution is appropriate. This model is widely used in biostatistical applications for two classes. For instance: people survive or die, have a disease or not, have a risk factor or not.
logistic function
A logistic function or logistic curve is the most common sigmoid curve.
 [math]y = \frac{1}{1+e^{x}}[/math]
1. [math]\frac{dy}{dx} = y(1y)=\frac{e^{x}}{(1+e^{x})^{2}}[/math]
2. [math]y(0) = \frac{1}{2}[/math]
3. [math] \int y dx = ln(1 + e^{x})[/math]
4. [math] y(x) = \frac{1}{2} + \frac{1}{4}x  \frac{1}{48}x^{3} + \frac{1}{48}x^{5} \cdots [/math]
5. The logistic curve shows early exponential growth for negative t, which slows to linear growth of slope 1/4 near t = 0, then approaches y = 1 with an exponentially decaying gap.
Intuition behind Logistic Regression
Recall that, for classification purposes, the linear regression model presented in the above section is not correct because it does not force [math]\,r(x)[/math] to be between 0 and 1 and sum to 1. Consider the following log odds model (for two classes):
 [math]\log\left(\frac{P(Y=1X=x)}{P(Y=0X=x)}\right)=\beta^Tx[/math]
Calculating [math]\,P(Y=1X=x)[/math] leads us to the logistic regression model, which as opposed to the linear regression model, allows the modelling of the posterior probabilities of the classes through linear methods and at the same time ensures that they sum to one and are between 0 and 1. It is a type of Generalized Linear Model (GLM).
The Logistic Regression Model
The logistic regression model for the two class case is defined as
Class 1
 [math]P(Y=1  X=x) =\frac{\exp(\underline{\beta}^T \underline{x})}{1+\exp(\underline{\beta}^T \underline{x})}=P(x;\underline{\beta})[/math]
Then we have that
Class 0
 [math]P(Y=0  X=x) = 1P(Y=1  X=x)=1\frac{\exp(\underline{\beta}^T \underline{x})}{1+\exp(\underline{\beta}^T \underline{x})}=\frac{1}{1+\exp(\underline{\beta}^T \underline{x})}[/math]
Fitting a Logistic Regression
Logistic regression tries to fit a distribution. The fitting of logistic regression models is usually accomplished by maximum likelihood, using Pr(YX). The maximum likelihood of [math]\underline\beta[/math] maximizes the probability of obtaining the data [math]\displaystyle{x_{1},...,x_{n}}[/math] from the known distribution. Combining [math]\displaystyle P(Y=1  X=x)[/math] and [math]\displaystyle P(Y=0  X=x)[/math] as follows, we can consider the two classes at the same time:
 [math]p(\underline{x_{i}};\underline{\beta}) = \left(\frac{\exp(\underline{\beta}^T \underline{x_i})}{1+\exp(\underline{\beta}^T \underline{x_i})}\right)^{y_i} \left(\frac{1}{1+\exp(\underline{\beta}^T \underline{x_i})}\right)^{1y_i}[/math]
Assuming the data [math]\displaystyle {x_{1},...,x_{n}}[/math] is drawn independently, the likelihood function is
 [math] \begin{align} \mathcal{L}(\theta)&=p({x_{1},...,x_{n}};\theta)\\ &=\displaystyle p(x_{1};\theta) p(x_{2};\theta)... p(x_{n};\theta) \quad \mbox{(by independence)}\\ &= \prod_{i=1}^n p(x_{i};\theta) \end{align} [/math]
Since it is more convenient to work with the loglikelihood function, take the log of both sides, we get
 [math]\displaystyle l(\theta)=\displaystyle \sum_{i=1}^n \log p(x_{i};\theta)[/math]
So,
 [math] \begin{align} l(\underline\beta)&=\displaystyle\sum_{i=1}^n y_{i}\log\left(\frac{\exp(\underline{\beta}^T \underline{x_i})}{1+\exp(\underline{\beta}^T \underline{x_i})}\right)+(1y_{i})\log\left(\frac{1}{1+\exp(\underline{\beta}^T\underline{x_i})}\right)\\ &= \displaystyle\sum_{i=1}^n y_{i}(\underline{\beta}^T\underline{x_i}\log(1+\exp(\underline{\beta}^T\underline{x_i}))+(1y_{i})(\log(1+\exp(\underline{\beta}^T\underline{x_i}))\\ &= \displaystyle\sum_{i=1}^n y_{i}\underline{\beta}^T\underline{x_i}y_{i} \log(1+\exp(\underline{\beta}^T\underline{x_i})) \log(1+\exp(\underline{\beta}^T\underline{x_i}))+y_{i} \log(1+\exp(\underline{\beta}^T\underline{x_i}))\\ &=\displaystyle\sum_{i=1}^n y_{i}\underline{\beta}^T\underline{x_i} \log(1+\exp(\underline{\beta}^T\underline{x_i}))\\ \end{align} [/math]
To maximize the loglikelihood, set its derivative to 0.
 [math] \begin{align} \frac{\partial l}{\partial \underline{\beta}} &= \sum_{i=1}^n \left[{y_i} \underline{x}_i \frac{\exp(\underline{\beta}^T \underline{x_i})}{1+\exp(\underline{\beta}^T \underline{x}_i)}\underline{x}_i\right]\\ &=\sum_{i=1}^n \left[{y_i} \underline{x}_i  p(\underline{x}_i;\underline{\beta})\underline{x}_i\right] \end{align} [/math]
There are n+1 nonlinear equations in [math]/ \beta [/math]. The first column is vector 1, then [math]\ \sum_{i=1}^n {y_i} =\sum_{i=1}^n p(\underline{x}_i;\underline{\beta}) [/math] i.e. the expected number of class ones matches the observed number.
To solve this equation, the NewtonRaphson algorithm is used which requires the second derivative in addition to the first derivative. This is demonstrated in the next section.
Advantages and Disadvantages
Logistic regression has several advantages over discriminant analysis:
 it is more robust: the independent variables don't have to be normally distributed, or have equal variance in each group
 It does not assume a linear relationship between the IV and DV
 It may handle nonlinear effects
 You can add explicit interaction and power terms
 The DV need not be normally distributed.
 There is no homogeneity of variance assumption.
 Normally distributed error terms are not assumed.
 It does not require that the independents be interval.
 It does not require that the independents be unbounded.
With all this flexibility, you might wonder why anyone would ever use discriminant analysis or any other method of analysis. Unfortunately, the advantages of logistic regression come at a cost: it requires much more data to achieve stable, meaningful results. With standard regression, and DA, typically 20 data points per predictor is considered the lower bound. For logistic regression, at least 50 data points per predictor is necessary to achieve stable results.
Extension
 When we are dealing with a problem with more than two classes, we need to generalize our logistic regression to a Multinomial Logit model.
 Limitations of Logistic Regression:
 1. We know that there is no assumptions are made about the distributions of the features of the data (i.e. the explanatory variables). However, the features should not be highly correlated with one another because this could cause problems with estimation.
 2. Large number of data points (i.e.the sample sizes) are required for logistic regression to provide sufficient numbers in both classes. The more number of features/dimensions of the data, the larger the sample size required.
Logistic Regression(2)  October 19, 2009
Logistic Regression Model
Recall that in the last lecture, we learned the logistic regression model.
 [math]P(Y=1  X=x)=P(\underline{x}_i;\underline{\beta})=\frac{exp(\underline{\beta}^T \underline{x})}{1+exp(\underline{\beta}^T \underline{x})}[/math]
 [math]P(Y=0  X=x)=1P(\underline{x}_i;\underline{\beta})=\frac{1}{1+exp(\underline{\beta}^T \underline{x})}[/math]
Find [math]\underline{\beta}[/math]
Criteria: find a [math]\underline{\beta}[/math] that maximizes the conditional likelihood of Y given X using the training data.
From above, we have the first derivative of the loglikelihood:
[math]\frac{\partial l}{\partial \underline{\beta}} = \sum_{i=1}^n \left[{y_i} \underline{x}_i \frac{exp(\underline{\beta}^T \underline{x_i})}{1+exp(\underline{\beta}^T \underline{x}_i)}\underline{x}_i\right] [/math] [math]=\sum_{i=1}^n \left[{y_i} \underline{x}_i  P(\underline{x}_i;\underline{\beta})\underline{x}_i\right][/math]
NewtonRaphson algorithm:
If we want to find [math]\ x^* [/math] such that [math]\ f(x^*)=0[/math]
[math]\ X^{new} \leftarrow x^{old}\frac {f(x^{old})}{\partial f(x^{old})} [/math]
[math]\ x^{new} \rightarrow x^* [/math]
If we want to maximize or minimize [math]\ f(x) [/math], then solve for [math]\ \partial f(x)=0 [/math]
[math]\ X^{new} \leftarrow x^{old}\frac {\partial f(x^{old})}{\partial^2 f(x^{old})} [/math]
The NewtonRaphson algorithm requires the secondderivative or Hessian matrix.
[math]\frac{\partial^{2} l}{\partial \underline{\beta} \partial \underline{\beta}^T }= \sum_{i=1}^n  \underline{x_i} \frac{(exp(\underline{\beta}^T\underline{x}_i) \underline{x}_i^T)(1+exp(\underline{\beta}^T \underline{x}_i))\underline{x}_i exp(\underline{\beta}^T\underline{x}_i)exp(\underline{\beta}^T\underline{x}_i)}{(1+exp(\underline{\beta}^T \underline{x}))^2}[/math]
(note: [math]\frac{\partial\underline{\beta}^T\underline{x}_i}{\partial \underline{\beta}^T}=\underline{x}_i^T[/math] you can check it here, it's a very useful website including a Matrix Reference Manual that you can find information about linear algebra and the properties of real and complex matrices.)
 [math]=\sum_{i=1}^n  \underline{x}_i \frac{(\underline{x}_i exp(\underline{\beta}^T\underline{x}_i) \underline{x}_i^T)}{(1+exp(\underline{\beta}^T \underline{x}))(1+exp(\underline{\beta}^T \underline{x}))}[/math] (by cancellation)
 [math]=\sum_{i=1}^n  \underline{x}_i \underline{x}_i^T P(\underline{x}_i;\underline{\beta}）)[1P(\underline{x}_i;\underline{\beta})]）[/math](since [math]P(\underline{x}_i;\underline{\beta})=\frac{exp(\underline{\beta}^T \underline{x})}{1+exp(\underline{\beta}^T \underline{x})}[/math] and [math]1P(\underline{x}_i;\underline{\beta})=\frac{1}{1+exp(\underline{\beta}^T \underline{x})}[/math])
The same second derivative can be achieved if we reduce the occurrences of beta to 1 by the identity[math]\frac{a}{1+a}=1\frac{1}{1+a}[/math]
And solving [math]\frac{\partial}{\partial \underline{\beta}^T}\sum_{i=1}^n \left[{y_i} \underline{x}_i\left[1\frac{1}{1+exp(\underline{\beta}^T \underline{x}_i)}\right]\underline{x}_i\right] [/math]
Starting with [math]\,\underline{\beta}^{old}[/math], the NewtonRaphson update is
[math]\,\underline{\beta}^{new}\leftarrow \,\underline{\beta}^{old} (\frac{\partial ^2 l}{\partial \underline{\beta}\partial \underline{\beta}^T})^{1}(\frac{\partial l}{\partial \underline{\beta}})[/math] where the derivatives are evaluated at [math]\,\underline{\beta}^{old}[/math]
The iteration will terminate when [math]\underline{\beta}^{new}[/math] is very close to [math]\underline{\beta}^{old}[/math].
The iteration can be described in matrix form.
 Let [math]\,\underline{Y}[/math] be the column vector of [math]\,y_i[/math]. ([math]n\times1[/math])
 Let [math]\,X[/math] be the [math]{d}\times{n}[/math] input matrix.
 Let [math]\,\underline{P}[/math] be the [math]{n}\times{1}[/math] vector with [math]i[/math]th element [math]P(\underline{x}_i;\underline{\beta}^{old})[/math].
 Let [math]\,W[/math] be an [math]{n}\times{n}[/math] diagonal matrix with [math]i[/math]th element [math]P(\underline{x}_i;\underline{\beta}^{old})[1P(\underline{x}_i;\underline{\beta}^{old})][/math]
then
[math]\frac{\partial l}{\partial \underline{\beta}} = X(\underline{Y}\underline{P})[/math]
[math]\frac{\partial ^2 l}{\partial \underline{\beta}\partial \underline{\beta}^T} = XWX^T[/math]
The NewtonRaphson step is
[math]\underline{\beta}^{new} \leftarrow \underline{\beta}^{old}+(XWX^T)^{1}X(\underline{Y}\underline{P})[/math]
This equation is sufficient for computation of the logistic regression model. However, we can simplify further to uncover an interesting feature of this equation.
[math] \begin{align} \underline{\beta}^{new} &= (XWX^T)^{1}(XWX^T)\underline{\beta}^{old}+(XWX^T)^{1}XWW^{1}(\underline{Y}\underline{P})\\ &=(XWX^T)^{1}XW[X^T\underline{\beta}^{old}+W^{1}(\underline{Y}\underline{P})]\\ &=(XWX^T)^{1}XWZ \end{align}[/math]
where [math]Z=X\underline{\beta}^{old}+W^{1}(\underline{Y}\underline{P})[/math]
This is a adjusted response and it is solved repeatedly when [math]\ p [/math], [math]\ W [/math], and [math]\ z [/math] changes. This algorithm is called iteratively reweighted least squares because it solves the weighted least squares problem repeatedly.
Recall that linear regression by least square finds the following minimum: [math]\min_{\underline{\beta}}(\underline{y}\underline{\beta}^T X)^T(\underline{y}\underline{\beta}^TX)[/math]
we have [math]\underline\hat{\beta}=(XX^T)^{1}X\underline{y}[/math]
Similarly, we can say that [math]\underline{\beta}^{new}[/math] is the solution of a weighted least square problem:
[math]\underline{\beta}^{new} \leftarrow arg \min_{\underline{\beta}}(ZX\underline{\beta}^T)W(ZX\underline{\beta})[/math]
WLS
Actually, the weighted least squares estimator minimizes the weighted sum of squared errors [math] S(\beta) = \sum_{i=1}^{n}w_{i}[y_{i}\mathbf{x}_{i}^{T}\beta]^{2} [/math] where [math]\displaystyle w_{i}\gt 0[/math]. Hence the WLS estimator is given by [math] \hat\beta^{WLS}=\left[\sum_{i=1}^{n}w_{i}\mathbf{x}_{i}\mathbf{x}_{i}^{T} \right]^{1}\left[ \sum_{i=1}^{n}w_{i}\mathbf{x}_{i}y_{i}\right] [/math]
A weighted linear regression of the iteratively computed response [math] \mathbf{z}=\mathbf{X}^{T}\beta^{old}+\mathbf{W}^{1}(\mathbf{y}\mathbf{p}) [/math]
Therefore, we obtain
 [math] \begin{align} & \hat\beta^{WLS}=\left[\sum_{i=1}^{n}w_{i}\mathbf{x}_{i}\mathbf{x}_{i}^{T} \right]^{1}\left[ \sum_{i=1}^{n}w_{i}\mathbf{x}_{i}z_{i}\right] \\& = \left[ \mathbf{XWX}^{T}\right]^{1}\left[ \mathbf{XWz}\right] \\& = \left[ \mathbf{XWX}^{T}\right]^{1}\mathbf{XW}(\mathbf{X}^{T}\beta^{old}+\mathbf{W}^{1}(\mathbf{y}\mathbf{p})) \\& = \beta^{old}+ \left[ \mathbf{XWX}^{T}\right]^{1}\mathbf{X}(\mathbf{y}\mathbf{p}) \end{align} [/math]
note:Here we obtain [math]\underline{\beta}[/math], which is a [math]d\times{1}[/math] vector, because we construct the model like [math]\underline{\beta}^T\underline{x}[/math]. If we construct the model like [math]\underline{\beta}_0+ \underline{\beta}^T\underline{x}[/math], then similar to linear regression, [math]\underline{\beta}[/math] will be a [math](d+1)\times{1}[/math] vector.
 Choosing [math]\displaystyle\beta=0[/math] seems to be a suitable starting value for the NewtonRaphson iteration procedure in this case. However, this does not guarantee convergence. The procedure will usually converge since the loglikelihood function is concave(or convex), but overshooting can occur. In the rare cases that the loglikelihood decreases, cut step
size by half, then we can always have convergence. In the case that it does not, we can just prove the local convergence of the method, which means the iteration would converge only if the initial point is closed enough to the exact solution. However, in practice, choosing an appropriate initial value is really trivial, namely, it is not often to find a initial too far from the exact solution to make the iteration invalid. <ref>C. T. Kelley, Iterative Methods for Linear and Nonlinear Equations, chapter 5 </ref> Besides, stepsize halving will solve this problem. <ref>H. Trevor, R. Tibshirani, J. Friedman, The Elements of Statistical Learning (Springer 2009),121.</ref>
For multiclass cases: the Newton algorithm can also be expressed as an iteratively reweighted least squares algorithm, but with a vector of [math]\ k1 [/math] response and a nondiagonal weight matrix per observation. And we can use coordinatedescent method to maximize the loglikelihood efficiently.
Pseudo Code
 [math]\underline{\beta} \leftarrow 0[/math]
 Set [math]\,\underline{Y}[/math], the label associated with each observation [math]\,i=1...n[/math].
 Compute [math]\,\underline{P}[/math] according to the equation [math]P(\underline{x}_i;\underline{\beta})=\frac{exp(\underline{\beta}^T \underline{x})}{1+exp(\underline{\beta}^T \underline{x})}[/math] for all [math]\,i=1...n[/math].
 Compute the diagonal matrix [math]\,W[/math] by setting [math]\,w_i,i[/math] to [math]P(\underline{x}_i;\underline{\beta}）)[1P(\underline{x}_i;\underline{\beta})][/math] for all [math]\,i=1...n[/math].
 [math]Z \leftarrow X^T\underline{\beta}+W^{1}(\underline{Y}\underline{P})[/math].
 [math]\underline{\beta} \leftarrow (XWX^T)^{1}XWZ[/math].
 If the new [math]\underline{\beta}[/math] value is sufficiently close to the old value, stop; otherwise go back to step 3.
Comparison with Linear Regression
 Similarities
 They are both to attempt to estimate [math]\,P(Y=kX=x)[/math] (For logistic regression, we just mentioned about the case that [math]\,k=0[/math] or [math]\,k=1[/math] now).
 They are both have linear boundaris.
 note:For linear regression, we assume the model is linear. The boundary is [math]P(Y=kX=x)=\underline{\beta}^T\underline{x}_i+\underline{\beta}_0=0.5[/math] (linear)
 For logistic regression, the boundary is [math]P(Y=kX=x)=\frac{exp(\underline{\beta}^T \underline{x})}{1+exp(\underline{\beta}^T \underline{x})}=0.5 \Rightarrow exp(\underline{\beta}^T \underline{x})=1\Rightarrow \underline{\beta}^T \underline{x}=0[/math] (linear)
 Differences
 Linear regression: [math]\,P(Y=kX=x)[/math] is linear function of [math]\,x[/math], [math]\,P(Y=kX=x)[/math] is not guaranteed to fall between 0 and 1 and to sum up to 1.
 Logistic regression: [math]\,P(Y=kX=x)[/math] is a nonlinear function of [math]\,x[/math], and it is guaranteed to range from 0 to 1 and to sum up to 1.
Comparison with LDA
 The linear logistic model only consider the conditional distribution [math]\,P(Y=kX=x)[/math]. No assumption is made about [math]\,P(X=x)[/math].
 The LDA model specifies the joint distribution of [math]\,X[/math] and [math]\,Y[/math].
 Logistic regression maximizes the conditional likelihood of [math]\,Y[/math] given [math]\,X[/math]: [math]\,P(Y=kX=x)[/math]
 LDA maximizes the joint likelihood of [math]\,Y[/math] and [math]\,X[/math]: [math]\,P(Y=k,X=x)[/math].
 If [math]\,\underline{x}[/math] is ddimensional,the number of adjustable parameter in logistic regression is [math]\,d[/math]. The number of parameters grows linearly w.r.t dimension.
 If [math]\,\underline{x}[/math] is ddimensional,the number of adjustable parameter in LDA is [math]\,(2d)+d(d+1)/2+2=(d^2+5d+4)/2[/math]. The number of parameters grows quardratically w.r.t dimension.
 LDA estimate parameters more efficiently by using more information about data and samples without class labels can be also used in LDA.
 As logistic regression relies on fewer assumptions, it seems to be more robust.
 In practice, Logistic regression and LDA often give the similar results.
By example
Now we compare LDA and Logistic regression by an example. Again, we use them on the 2_3 data.
>>load 2_3; >>[U, sample] = princomp(X'); >>sample = sample(:,1:2); >>plot (sample(1:200,1), sample(1:200,2), '.'); >>hold on; >>plot (sample(201:400,1), sample(201:400,2), 'r.');
 First, we do PCA on the data and plot the data points that represent 2 or 3 in different colors. See the previous example for more details.
>>group = ones(400,1); >>group(201:400) = 2;
 Group the data points.
>>[B,dev,stats] = mnrfit(sample,group); >>x=[ones(1,400); sample'];
 Now we use mnrfit to use logistic regression to classfy the data. This function can return B which is a [math](d+1)\times{(k–1)}[/math] matrix of estimates, where each column corresponds to the estimated intercept term and predictor coefficients. In this case, B is a [math]3\times{1}[/math] matrix.
>> B B =0.1861 5.5917 3.0547
 This is our [math]\underline{\beta}[/math]. So the posterior probabilities are:
 [math]P(Y=1  X=x)=\frac{exp(0.18615.5917X_13.0547X_2)}{1+exp(0.18615.5917X_13.0547X_2)}[/math].
 [math]P(Y=2  X=x)=\frac{1}{1+exp(0.18615.5917X_13.0547X_2)}[/math]
 The classification rule is:
 [math]\hat Y = 1[/math], if [math]\,0.18615.5917X_13.0547X_2\gt =0[/math]
 [math]\hat Y = 2[/math], if [math]\,0.18615.5917X_13.0547X_2\lt 0[/math]
>>f = sprintf('0 = %g+%g*x+%g*y', B(1), B(2), B(3)); >>ezplot(f,[min(sample(:,1)), max(sample(:,1)), min(sample(:,2)), max(sample(:,2))])
 Plot the decision boundary by logistic regression.
>>[class, error, POSTERIOR, logp, coeff] = classify(sample, sample, group, 'linear'); >>k = coeff(1,2).const; >>l = coeff(1,2).linear; >>f = sprintf('0 = %g+%g*x+%g*y', k, l(1), l(2)); >>h=ezplot(f, [min(sample(:,1)), max(sample(:,1)), min(sample(:,2)), max(sample(:,2))]);
 Plot the decision boundary by LDA. See the previous example for more information about LDA in matlab.
2009.10.21
MultiClass Logistic Regression
Our earlier goal with logistic regression was to model the posteriors for a 2 class classification problem with a linear function bounded by the interval [0,1]. In that case our model was,
[math]\log\left(\frac{P(Y=1X=x)}{P(Y=0X=x)}\right)= \log\left(\frac{\frac{\exp(\beta^T x)}{1+\exp(\beta^T x)}}{\frac{1}{1+\exp(\beta^T x)}}\right) =\beta^Tx[/math]
We can extend this idea to the more general case with Kclasses. This model is specified with K  1 terms where the Kth class in the denominator can be chosen arbitrarily.
[math]\log\left(\frac{P(Y=iX=x)}{P(Y=KX=x)}\right)=\beta_i^Tx,\quad i \in \{1,\dots,K1\} [/math]
The posteriors for each class are given by,
[math]P(Y=iX=x) = \frac{\exp(\beta_i^T x)}{1+\sum_{k=1}^{K1}\exp(\beta_k^T x)}, \quad i \in \{1,\dots,K1\}[/math]
[math]P(Y=KX=x) = \frac{1}{1+\sum_{k=1}^{K1}\exp(\beta_k^T x)}[/math]
Seeing these equations as a weighted least squares problem makes them easier to derivate.
Note that we still retain the property that the sum of the posteriors is 1. In general the posteriors are no longer complements of each other as in true in the 2 class problem where we could express [math]\displaystyle P(Y=1X=x)=1P(Y=0X=x)[/math]. Fitting a Logistic model for the K>2 class problem isn't as 'nice' as in the 2 class problem since we don't have the same simplification.
Multiclass kernel logistic regression
Logistic regression (LR) and kernel logistic regression (KLR) have already proven their value in the statistical and machine learning community. Opposed to an empirically risk minimization approach such as employed by Support Vector Machines (SVMs), LR and KLR yield probabilistic outcomes based on a maximum likelihood argument. It seems that this framework provides a natural extension to multiclass classification tasks, which must be contrasted to the commonly used coding approach.
A paper uses the LSSVM framework to solve the KLR problem. In that paper,they see that the minimization of the negative penalized log likelihood criterium is equivalent to solving in each iteration a weighted version of least squares support vector machines (wLSSVMs). In the derivation it turns out that the global regularization term is reflected as usual in each step. In a similar iterative weighting of wLSSVMs, with different weighting factors is reported to converge to an SVM solution.
Unlike SVMs, KLR by its nature is not sparse and needs all training samples in its final model. Different adaptations to the original algorithm were proposed to obtain sparseness. The second one uses a sequential minimization optimization (SMO) approach and in the last case, the binary KLR problem is reformulated into a geometric programming system which can be efficiently solved by an interiorpoint algorithm. In the LSSVM framework, fixedsize LSSVM has shown its value on large data sets. It approximates the feature map using a spectral decomposition, which leads to a sparse representation of the model when estimating in the primal space. They use this technique as a practical implementation of KLR with estimation in the primal space. To reduce the size of the Hessian, an alternating descent version of Newton’s method is used which has the extra advantage that it can be easily used in a distributed computing environment. The proposed algorithm is compared to existing algorithms using small size to large scale benchmark data sets.
Paper's Link: [[17]]
Perceptron (Foundation of Neural Network)
Separating Hyperplane Classifiers
Separating hyperplane trys to separate the data using linear decision boundaries. When the classes overlap, it can be generalized to support vector machine, which constructs nonlinear boundaries by constructing a linear boundary in an enlarged and transformed feature space.
Perceptron
Recall the use of Least Squares regression as a classifier, shown to be identical to LDA. To classify points with least squares we take the sign of a linear combination of data points and assign a label equivalent to +1 or 1.
Least Squares returns the sign of a linear combination of data points as the class label
[math]sign(\underline{\beta}^T \underline{x} + {\beta}_0) = sign(\beta_{0}+\beta_{1}x_{1}+\beta_{2}x_{2})[/math]
In the 1950s Frank Rosenblatt developed an iterative linear classifier while at Cornell University known as the Perceptron. The concept of a perceptron was fundamental to the later development of the Artificial Neural Network models. The perceptron is a simple type of neural network which models the electrical signals of biological neurons. In fact, it was the first neural network to be algorithmically described. <ref>Simon S. Haykin, Neural Networks and Learning Machines, (Prentice Hall 2008). </ref>
As in other linear classification methods like Least Squares, Rosenblatt's classifier determines a hyperplane for the decision boundary. Linear methods all determine slightly different decision boundaries, Rosenblatt's algorithm seeks to minimize the distance between the decision boundary and the misclassified points <ref>H. Trevor, R. Tibshirani, J. Friedman, The Elements of Statistical Learning (Springer 2009),156.</ref>.
Particular to the iterative nature of the solution, the problem has no global mean (not convex). It does not converge to give a unique hyperplane, and the solutions depend on the size of the gap between classes. If the classes are separable then the algorithm is shown to converge to a local mean. The proof of this convergence is known as the perceptron convergence theorem. However, for overlapping classes convergence to a local mean cannot be guaranteed.
If we find a hyperplane that is not unique between 2 classes, there will be infinitely many solutions obtained from the perceptron algorithm.
As seen in Figure 1, after training, the perceptron determines the label of the data by computing the sign of a linear combination of components.
A Perceptron Example
The perceptron network can figure out the decision boundray line even if we dont know how to draw the line. We just have to give it some examples first. For example:
Features:x1, x2, x3  Answer 

1,0,0  +1 
1,0,1  +1 
1,1,0  +1 
0,0,1  1 
0,1,1  1 
1,1,1  1 
Then the perceptron starts out not knowing how to separate the answers so it guesses. For example we input 1,0,0 and it guesses 1. But the right answer is +1. So the perceptron adjusts its line and we try the next example. Eventually the perceptron will have all the answers right.
y=[1;1;1;1;1;1]; x=[1,0,0;1,0,1;,1,1,0;0,0,1;0,1,1;1,1,1]'; b_0=0; b=[1;1;1]; rho=.5; for j=1:100; changed=0; for i=1:6 d=(b'*x(:,i)+b_0)*y(i); if d<0 b=b+rho*x(:,i)*y(i); b_0=b_0+rho*y(i); changed=1; end end if changed==0 break; end end
The Perceptron (Lecture October 23, 2009)
A Perceptron can be modeled as shown in Figure 1 of the previous lecture where[math]\,x_0[/math] is the model intercept and [math]x_{1},\ldots,x_{d}[/math] represent the feature data, [math]\sum_{i=0}^d \beta_{j}x_{j}[/math] is a linear combination of some weights of these inputs, and [math]I(\sum_{i=1}^d \beta_{j}x_{j})[/math], where [math]\,I[/math] indicates the sign of the expression and returns the label of the data point.
The Perceptron algorithm seeks a linear boundary between two classes. A linear decision boundary can be represented by[math] \underline{\beta}^T\underline{x}+\beta_{0}. [/math] The algorithm begins with an arbitrary hyperplane [math]\underline{\beta}^T\underline{x}+\beta_{0} [/math] (initial guess). Its goal is to minimize the distance between the decision boundary and the misclassified data points. This is illustrated in Figure 2. It attempts to find the optimal [math]\underline\beta[/math] by iteratively adjusting the decision boundary until all points are on the correct side of the boundary. It terminates when there are no misclassified points.
Derivation: The distance between the decision boundary and misclassified points.
If [math]\underline{x_{1}}[/math] and [math]\underline{x_{2}}[/math]both lie on the decision boundary then,
 [math] \begin{align} \underline{\beta}^T\underline{x_{1}}+\beta_{0} &= \underline{\beta}^T\underline{x_{2}}+\beta_{0} \\ \underline{\beta}^T (x_{1}x_{2})&=0 \end{align} [/math]
[math]\underline{\beta}^T (x_{1}x_{2})[/math] denotes an inner product. Since the inner product is 0 and [math](\underline{x_{1}}\underline{x_{2}})[/math] is a vector lying on the decision boundary, [math]\underline{\beta}[/math] is orthogonal to the decision boundary.
Let [math]\underline{x_{i}}[/math] be a misclassified point.
Then the projection of the vector [math] \underline{x_{i}}[/math] on the direction that is orthogonal to the decision boundary is [math]\underline{\beta}^T\underline{x_{i}}[/math].
Now, if [math]\underline{x_{0}}[/math] is also on the decision boundary, then [math]\underline{\beta}^T\underline{x_{0}}+\beta_{0}=0[/math] and so [math]\underline{\beta}^T\underline{x_{0}}= \beta_{0}[/math]. Looking at Figure 3, it can be seen that the distance between [math]\underline{x_{i}}[/math] and the decision boundary is the absolute value of [math]\underline{\beta}^T\underline{x_{i}}+\beta_{0}. [/math]
Consider [math]y_{i}(\underline{\beta}^T\underline{x_{i}}+\beta_{0}).[/math]
 Notice that if [math]\underline{x_{i}}[/math] is classified correctly then this product is positive. This is because if it is classified correctly, then either both ([math]\underline{\beta}^T\underline{x_{i}}+\beta_{0})[/math] and[math]\displaystyle y_{i}[/math] are positive or they are both negative. However, if [math]\underline{x_{i}}[/math] is classified incorrectly then one of [math](\underline{\beta}^T\underline{x_{i}}+\beta_{0})[/math] and [math]\displaystyle y_{i}[/math] is positive and the other is negative. The result is that the above product is negative for a point that is misclassified.
For the algorithm, we need only consider the distance between the misclassified points and the decision boundary.
 Consider [math]\phi(\underline{\beta},\beta_{0})= \displaystyle\sum_{i\in M} –y_{i}(\underline{\beta}^T\underline{x_{i}}+\beta_{0}) [/math]
which is a summation of positive numbers and where [math]\displaystyle M[/math] is the set of all misclassified points.
The goal now becomes to [math]\min_{\underline{\beta},\beta_{0}} \phi(\underline{\beta},\beta_{0}). [/math]
This can be done using a gradient descent approach, which is a numerical method that takes one predetermined step in the direction of the gradient, getting closer to a minimum at each step, until the gradient is zero. A problem with this algorithm is the possibility of getting stuck in a local minimum. To continue, the following derivatives are needed:
 [math]\frac{\partial \phi}{\partial \underline{\beta}}= \displaystyle\sum_{i \in M}y_{i}\underline{x_{i}} \ \ \ \ \ \ \ \ \ \ \ \frac{\partial \phi}{\partial \beta_{0}}= \displaystyle\sum_{i \in M}y_{i}[/math]
Then the gradient descent type algorithm (Perceptron Algorithm) is
 [math] \begin{pmatrix} \underline{\beta}^{\mathrm{new}}\\ \underline{\beta_0}^{\mathrm{new}} \end{pmatrix} = \begin{pmatrix} \underline{\beta}^{\mathrm{old}}\\ \underline{\beta_0}^{\mathrm{old}} \end{pmatrix} +\rho \begin{pmatrix} y_i \underline{x_i}\\ y_i \end{pmatrix} [/math]
where [math]\displaystyle\rho[/math] is the magnitude of each step called the "learning rate" or the "convergence rate". The algorithm continues until [math]
\begin{pmatrix}
\underline{\beta}^{\mathrm{new}}\\
\underline{\beta_0}^{\mathrm{new}}
\end{pmatrix}
=
\begin{pmatrix}
\underline{\beta}^{\mathrm{old}}\\
\underline{\beta_0}^{\mathrm{old}}
\end{pmatrix} [/math]
or until it has iterated a specified number of times. If the algorithm converges, it has found a linear classifier, ie., there are no misclassified points.
Problems with the Algorithm and Issues Affecting Convergence
 The output values of a perceptron can take on only one of two values (+1 or 1); that is, it only can be used for twoclass classification.
 If the data is not separable, then the Perceptron algorithm will not converge since it cannot find a linear classifier that classifies all of the points correctly.
 Convergence rates depend on the size of the gap between classes. If the gap is large, then the algorithm converges quickly. However, if the gap is small, the algorithm converges slowly. This problem can be eliminated by using basis expansions technique. To be specific, we try to find a hyperplane not in the original space, but in the enlarged space obtained by using some basis functions.
 If the classes are separable, there exists infinitely many solutions to Perceptron, all of which are hyperplanes.
 The speed of convergence of the algorithm is also dependent on the value of [math]\displaystyle\rho[/math], the learning rate. A larger value of [math]\displaystyle\rho[/math] could yield quicker convergence, but if this value is too large, it may also result in “skipping over” the minimum that the algorithm is trying to find and possibly oscillating forever between the last two points, before and after the min.
 A perfect separation is not always available even desirable. If observations comes from different classes sharing the same imput, the classification model seems to be overfitting and will generally have poor predictive performance.
 The perceptron convergence theorem states that if there exists an exact solution (in other words, if the training data set is linearly separable), then the perceptron learning algorithm is guaranteed to ﬁnd an exact solution in a ﬁnite number of steps. Proofs of this theorem can be found for example in Rosenblatt (1962), Block (1962), Nilsson (1965), Minsky and Papert (1969), Hertz et al. (1991), and Bishop (1995a). Note, however, that the number of steps required to achieve convergence could still be substantial, and in practice, until convergence is achieved we will not be able to distinguish between a nonseparable problem and one that is simply slow to converge<ref>
Pattern Recognition and Machine Learning,Christopher M. Bishop,194
</ref>.
Comment on gradient descent algorithm
Consider yourself on the peak and you want to get to the land as fast as possible. So which direction should you step? Intuitively it should be the direction in which the height decreases fastest, which is given by the gradient. However, if the mountain has a saddle shape and you initially stand in the middle, then you will finally arrive at the saddle point (local minimum) and get stuck there.
In addition, note that in the final form of our gradient descent algorithm, we get rid of the summation over [math]\,i[/math] (all data points). Actually, this is an alternative of the original gradient descent algorithm (sometimes called batch gradient descent) known as Stochastic gradient descent, where we approximate the true gradient by only evaluating on a single training example. This means that [math]\,{\beta}[/math] gets improved by computation of only one sample. When there is a large data set, say, population database, it's very timeconsuming to do summation over millions of samples. By Stochastic gradient descent, we can treat the problem sample by sample and still get decent result in practice.
 A Perceptron applet can be found at http://isl.ira.uka.de/neuralNetCourse/2004/VL_11_5/Perceptron.html .
Neural Networks (NN)  October 28, 2009
Introduction
A neural network is a two stage regression for classification model. It can be represented by a network diagram. It is a parallel, distributed information processing structure consisting of processing elements interconnected together with signal channels called connections. Each processing element has a single output connection with branches that "fan out" onto as many connections as desired, each carrying the same signal  the processing element output signal.
<ref> Haykin, Simon (2009). Neural Networks and Learning Machines. Pearson Education, Inc. </ref> A neural network resembles the brain in two respects:
1. Knowledge is acquired by the network from its environment through a learning process.
2. Interneuron connection strengths, known as synaptic weights, are used to store the acquired knowledge.
<ref> Theory of the Backpropagation Neural Network, R. NechtNielsen </ref> It is a multistage regression or classification model represented by a network. Figure 1 is an example of a typical neural network but it can have many different forms. It network applies both to regression or classification.
 A regression problem typically has only one unit [math]\ y_1 [/math] in the output layer, but these networks can handle multiple quantitative responses in a seamless fashion.
 In a kclass classification problem, there are usually k target measurements units [math]\ y_1,...,y_k [/math] in the output layer that each represent the probability of class k and each [math]\displaystyle y_k[/math] is coded (0,1).
Activation Function
Activation Function is a term that is frequently used in classification by NN.
In perceptron, we have a "sign" function that takes the sign of a weighted sum of input features.
File:signfuncperceptron.png
The sign function is of the form File:signfunc1.png ; it is not continuous at 0 and we cannot take derivative of it. Thus, we replace it by a smooth function [math]\displaystyle \sigma [/math] of the form File:signfunc2.png and call it the activation function.
The choice of this function [math]\displaystyle \sigma [/math] is determined by the properties of the data and the assumed distribution of target variables, but for multiple binary classification problems the logistic function, also known as inverselogit (sigmoid function), is often used:
[math]\sigma(a)=\frac {1}{1+e^{a}}[/math]
There are some important properties for the activation function.
 Activation function is nonlinear. It can be shown that if the activation function of the hidden units is linear, a threelayer neural network is equivalent to a two layer one.
 Activation function saturate, which means there are maximum and minimum output value. This property ensures that the weights are bounded and therefore the searching time is limited.
 Activation function is continuous and smooth.
 Activation function is monotonic. This property is not necessary, since we know that RBF networks is also a kind of power model.
Note: A key difference between a perceptron and a neural network is that a neural network uses continuous nonlinearities in the units, for the purpose of differentiation, whereas the perceptron often uses a nondifferentiable activation function. The neural network function is differentiable with respect to the network parameters so that a gradient descent method can be used in training. Moreover, a perceptron is a linear classifier, whereas a neural network,by introducting the nonlinear transformation [math]\ \sigma [/math],it greatly enlarges the class of linear models and by combining layers of perceptrons, neural network is able to classify nonlinear problems through proper training.
By assigning some weights to the connectors in the neural network (see diagram above) we weigh the input that comes into the perceptron, to get an output that in turn acts as an input to the next layer of perceptrons, and so on for each layer(There are no crossconnections between units in the same layer and no backward connections from layers downstream. Typically, units in layer k provide input only to units in layer k +1). This type of neural network is called FeedForward Neural Network. Applications to FeedForward Neural Networks include data reduction, speech recognition, sensor signal processing, and ECG abnormality detection, to name a few. <ref>J. Annema, FeedForward Neural Networks, (Springer 1995), pp. 9 </ref>
Backpropagation
Introduction:
For a while, the Neural Network model was just an idea, since there were no algorithms for training the model until 1986, when Geoffrey Hinton <ref> http://www.cs.toronto.edu/~hinton/backprop.html </ref> devised an algorithm called backpropagation [18]. After that, a number of other training algorithms and various configurations of neural networks were implemented. Work procedure: Each neuron receives a signal from previous neurons, and each of their signals is multiplied by a different weight value. Sum up these weighted inputs and passed through the activation function which scales the output to a fixed range of values. The output of the limiter is then broadcast to all of the neurons in the next layer i.e. we apply the input values to the inputs of the first layer, allow the signals to propagate through the network, and read the output values.
When we were talking about perceptrons, we applied a gradient descent algorithm for optimizing weights. Backpropagation uses this idea of gradient descent to train a neural network based on the chain rule in calculus.
Assume that the last output layer has only one unit, so we are working with a regression problem. Later we will see how this can be extended to more output layers and thus turn into a classificaiton problem.
For simplicity, there is only 1 unit at the end and assume for the moment we are doing regression.
Note that we make a distinction between the input weights [math]\displaystyle (w_i)[/math] and hidden weights [math]\displaystyle (u_i)[/math].
Within each unit we have a function [math]\displaystyle z_i=\sigma(a_i)[/math] that takes input [math]\displaystyle a_i[/math] (linear sum of previous level) and outputs [math]\displaystyle z_i's[/math]. The [math]\displaystyle z_i's[/math] are the inputs into the final output of the model [math]\Rightarrow \hat y_i=\sum_{i=1}^p w_i z_i[/math]
We can find the error of the neural network output by evaluating the squared difference between the true classification and the resulting classification output [math]\Rightarrow \displaystyle error=y\hat y ^2 [/math]
First find derivative of the model error with respect to output weights [math]\displaystyle w_i[/math]
[math]\frac{\partial err}{\partial w_i}=\frac{\partial err}{\partial \hat y} \cdot \frac{\partial \hat y}{\partial w_i}[/math]
[math]\frac{\partial err}{\partial w_i}=2(y\hat y) \cdot z_i[/math]
Now we need to find the derivative of the model error with respect to hidden weights [math]\displaystyle u_i's[/math]
Consider the following diagram that opens up the hidden layers of the neural network:
i j are reversed!
Notice that the weighted sum on the output of the perceptrons at layer [math]\displaystyle l[/math] are the inputs into the perceptrons at layer [math]\displaystyle j[/math] and so on for all hidden layers.
So, using the chain rule
[math]\frac{\partial err}{\partial u_{jl}}=\frac{\partial err}{\partial a_j} \cdot \frac{\partial a_j}{\partial u_{jl}}[/math]
[math]\frac{\partial err}{\partial u_{jl}}=\delta_j \cdot z_l[/math]
Note that a change in [math]\,a_j[/math] causes changes in all [math]\,a_i[/math] in the next layer on which the error is based, so we need to sum over i in the chain:
[math]\delta_j = \frac{\partial err}{\partial a_j} = \sum_i \frac{\partial err}{\partial a_i} \cdot \frac{\partial a_i}{\partial a_j} =\sum_i \delta_i \cdot \frac{\partial a_i}{\partial a_j}[/math]
[math]\,\frac{\partial a_i}{\partial a_j}=\frac{\partial a_i}{\partial z_j} \cdot \frac{\partial z_j}{\partial a_j}=u_{ij} \cdot \sigma'(a_j)[/math] Using the activation function [math]\,\sigma(\cdot)[/math]
So [math]\delta_j = \sum_i \delta_i \cdot u_{ij} \cdot \sigma'(a_j)[/math]
[math]\delta_j = \sigma'(a_j)\sum_i \delta_i \cdot u_{ij}[/math]
We can propagate the error calculated in the output back through the previous layers and adjust weights to minimize error.
BackPropagation neural network is a good method for following situations:
 The problem is very complex and the number of input or output data points is very large, and have no idea to relate the input to the output.
 The solution varies over time within the bounds of the given input and output data, or the output is not easy to measure.
Neural Networks (NN)  October 30, 2009
[math]\sum_{i=1}^k a_i\cdot\cfrac{\partial u_j}{\partial a_i}\lt \math\gt === Backpropagation === The idea is that we first feed an input (we can normalize the data before feeding) from the training set to the Neural Network, then find the error rate at the output and then we propagate the error to previous layers and for each edge of weight \lt math\gt \,u_{ij}[/math] we find [math]\frac{\partial \mathrm{err}}{\partial u_{ij}}[/math]. Having the error rates at hand we adjust the weight of each edge by taking steps proportional to the negative of the gradient to decrease the error at output. The next step is to apply the next input from the training set and go through the described adjustment procedure. The overview of Backpropagation algorithm:
 Feed a point [math]\,x[/math] in the training set to the network, and find the output of all the nodes.
 Evaluate [math]\,\delta_k=y_k\hat{y_k}[/math] for all output units, where [math]y_k[/math] is the expected output and [math]\hat{y_k}[/math] is the real output.
 By propagating to the previous layers evaluate all [math]\,\delta_j[/math]s for hidden units: [math]\,\delta_j=\sigma'(a_j)\sum_i \delta_i u_{ij}[/math] where [math]i[/math] is associated to the previous layer.
 Using [math]\frac{\partial \mathrm{err}}{\partial u_{jl}} = \delta_j\cdot z_l[/math] find all the derivatives.
 Adjust each weight by taking steps proportional to the negative of the gradient: [math]u_{jl}^{\mathrm{new}} \leftarrow u_{jl}^{\mathrm{old}} \rho \frac{\partial \mathrm{err}}{u_{ij}}[/math]
 Feed the next point in the training set and repeat the above steps.
Advantage of Back propageation:
 Reduce the cost of computing derivatives by a factor of the number of derivatives to be calculated when minimizing the error.
 Allow higher degrees of nonlinearity and precision to be applied to problems.
How to initialize the weights
This still leaves the question of how to initialize the weights [math]\,u_{ij}, w_i[/math]. The method of choosing weights mentioned in class was to randomize the weights before the first step. This is not likely to be near the optimal solution in every case, but is simple to implement. To be more specific, random values near zero will be a good choice for the initial weights(usually from [1,1]). In this case, the model evolves from a nearly linear one to a nonlinear one as we desired. An alternative is to use an orthogonal least squares method to find the initial weights <ref>http://www.mitpressjournals.org/doi/abs/10.1162/neco.1995.7.5.982</ref>. Regression is performed on the weights and output by using a linear approximation of [math]\,\sigma(a_i)[/math], and finds optimal weights in the linear model. Back propagation is used afterward to find the optimal solution, since the NN is nonlinear.
Why all initial weights should be randomized and small?
 Since the error back propagated through the network is proportional to the value of the weights. If all the weights are the same, then the back propagated errors will be the same as well and causing all of the weights will be updated by the same amount. Thus, same initial weights will prevent the network from learning.
 Since the weights updates in the Back Prop algorithm are proportional to the derivative of activation function, it is important to consider how the net input affects its value. The derivative is a maximum when the activation function is equal to 0.5 and approaches its minimum as the activation function approaches 0 or 1, then its associated weights will vary very little. Thus, if we choose small initial weights, we will have the activation function close to the maximal weight change.
How to set learning rates
The learning rate [math]\,\rho[/math] is usually a constant.
If we use Online learning, as a form of stochastic approximation process, [math]\,\rho[/math] should decrease as the iteration increase.
In typical feedforwad NNs with hidden units, the objective function has many local and global optimal values, so the optimal learning rate often changes dramatically during the training process. The larger the learning rate the larger the the weight changes on each epoch, and the quicker the network learns.However, the size of the learning rate can also influence whether the network achieves a stable solution. Choosing too large learning rate may cause the unstability of the system and make the weights and objective function diverge, while the too small learning rate may lead to a very slow convergence rate(very long time in learning phase). However, the advantage of small learning rate is that it can guarantee the convergence. Thus, generally, it is better to choose a relatively small learning rate to ensure the stability. Usually, choose [math]\,\rho[/math] between 0.01 and 0.7.
If the learning rate is appropriate, the algorithm is guaranteed to converge to a local minimum, but not a global minimum which is better. Furthermore, there can exist many local minimum values.
Here we will mainly discuss how to estimate the number of hidden units at very beginning. Obviously, we should adjust it to be more precise using CV, LOO or other complexity control methods.
Basically, if the patterns are well separated, few hidden units are fairly enough. If the patterns are drawn from some highly complicated mixture models, more hidden units are really needed.
Actually, the number of hidden units determines the size of the model, and therefore the total number of the weights in the model. Typically speaking, the number of weights should not be larger than the number of training data, say N. Thus, sometimes, N/10 is a good choice. However, in pratice, many well performed models will use more hidden units.
Dimensionality reduction application
One possible application of Neural Networks is to perform dimensionality reduction, like other techniques, e.g., PCA, MDS, LLE and Isomap.
Consider the following configuration as shown in figure 1: As we go forward in layers of this Neural Network, the number of nodes is reduced, until we reach a layer with the number of nodes representing the desired dimensionality. However, note that at the very first few layers the number of nodes may not be strictly decreasing, as long as finally it can reach a layer with less nodes. From now on in the Neural Network the previous layers are mirrored. So at the output layer we have the same number of states as we have in the input layer. Now note that if we feed the network with each point and get an output approximately equal to the fed input, that means at the output the same input is reconstructed from the middle layer units. So the output of the middle layer units can represent the input with less dimensions.
To train this Neural Network, we feed the network with a training point and through back propagation we adjust the network weights based on the error between the input layer and the reconstruction at the output layer. Our low dimensional mapping will be the observed output from the middle layer. Data reconstruction consists of putting the low dimensional data through the second half of the network.
Deep Neural Network
Backpropagation in practice may not work well when there are too many hidden layers, since the [math]\,\delta[/math] may become negligible and the errors vanish. This is a numerical problem, where it is difficult to estimate the errors. So in practice configuring a Neural Network with Backpropagation faces some subtleties. Deep Neural Networks became popular two or three years ago, when introduced by Bradford Nill in his PhD thesis. Deep Neural Network training algorithm deals with the training of a Neural Network with a large number of layers.
The approach of training the deep network is to assume the network has only two layers first and train these two layers. After that we train the next two layers, so on and so forth.
Although we know the input and we expect a particular output, we do not know the correct output of the hidden layers, and this will be the issue that the algorithm mainly deals with. There are two major techniques to resolve this problem: using Boltzman machine to minimize the energy function, which is inspired from the theory in atom physics concerning the most stable condition; or somehow finding out what output of the second layer is most likely to lead us to the expected output at the output layer.
Difficulties of training deep architecture <ref>Template:Cite journal</ref>
Given a particular task, a natural way to train a deep network is to frame it as an optimization problem by specifying a supervised cost function on the output layer with respect to the desired target and use a gradientbased optimization algorithm in order to adjust the weights and biases of the network so that its output has low cost on samples in the training set. Unfortunately, deep networks trained in that manner have generally been found to perform worse than neural networks with one or two hidden layers.
We discuss two hypotheses that may explain this difficulty. The first one is that gradient descent can easily get stuck in poor local minima (Auer et al., 1996) or plateaus of the nonconvex training criterion. The number and quality of these local minima and plateaus (Fukumizu and Amari, 2000) clearly also influence the chances for random initialization to be in the basin of attraction (via gradient descent) of a poor solution. It may be that with more layers, the number or the width of such poor basins increases. To reduce the difficulty, it has been suggested to train a neural network in a constructive manner in order to divide the hard optimization problem into several greedy but simpler ones, either by adding one neuron (e.g., see Fahlman and Lebiere, 1990) or one layer (e.g., see Lengell´e and Denoeux, 1996) at a time. These two approaches have demonstrated to be very effective for learning particularly complex functions, such as a very nonlinear classification problem in 2 dimensions. However, these are exceptionally hard problems, and for learning tasks usually found in practice, this approach commonly overfits.
This observation leads to a second hypothesis. For high capacity and highly flexible deep networks, there actually exists many basins of attraction in its parameter space (i.e., yielding different solutions with gradient descent) that can give low training error but that can have very different generalization errors. So even when gradient descent is able to find a (possibly local) good minimum in terms of training error, there are no guarantees that the associated parameter configuration will provide good generalization. Of course, model selection (e.g., by crossvalidation) will partly correct this issue, but if the number of good generalization configurations is very small in comparison to good training configurations, as seems to be the case in practice, then it is likely that the training procedure will not find any of them. But, as we will see in this paper, it appears that the type of unsupervised initialization discussed here can help to select basins of attraction (for the supervised finetuning optimization phase) from which learning good solutions is easier both from the point of view of the training set and of a test set.
Neural Networks in Practice
Now that we know so much about Neural Networks, what are suitable real world applications? Neural Networks have already been successfully applied in many industries.
Since neural networks are good at identifying patterns or trends in data, they are well suited for prediction or forecasting needs, such as customer research, sales forecasting, risk management and so on.
Take a specific marketing case for example. A feedforward neural network was trained using backpropagation to assist the marketing control of airline seat allocations. The neural approach was adaptive to the rule. The system is used to monitor and recommend booking advice for each departure.
Issues with Neural Network
When Neural Networks was first introduced they were thought to be modeling human brains, hence they were given the fancy name "Neural Network". But now we know that they are just logistic regression layers on top of each other but have nothing to do with the real function principle in the brain.
We do not know why deep networks turn out to work quite well in practice. Some people claim that they mimic the human brains, but this is unfounded. As a result of these kinds of claims it is important to keep the right perspective on what this field of study is trying to accomplish. For example, the goal of machine learning may be to mimic the 'learning' function of the brain, but necessarily the processes the brain uses to learn.
As for the algorithm, since it does not have a convex form, we still face the problem of local minimum, although people have devised other techniques to avoid this dilemma.
In sum, Neural Network lacks a strong learning theory to back up its "success", thus it's hard for people to wisely apply and adjust it. Having said that, it is not an active research area in machine learning. NN still has wide applications in the engineering field such as in control.
BUSINESS APPLICATIONS OF NEURAL NETWORKS
Neural networks are increasingly being used in realworld business applications and, in some cases, such as fraud detection, they have already become the method of choice. Their use for risk assessment is also growing and they have been employed to visualize complex databases for marketing segmentation. This method covers a wide range of business interests — from finance management, through forecasting, to production. The combination of statistical, neural and fuzzy methods now enables direct quantitative studies to be carried out without the need for rocketscience expertise.
 On the Use of Neural Networks for Analysis Travel Preference Data
 Extracting Rules Concerning Market Segmentation from Artificial Neural Networks
 Characterization and Segmenting the BusinesstoConsumer ECommerce Market Using Neural Networks
 A Neurofuzzy Model for Predicting Business Bankruptcy
 Neural Networks for Analysis of Financial Statements
 Developments in Accurate Consumer Risk Assessment Technology
 Strategies for Exploiting Neural Networks in Retail Finance
 Novel Techniques for Profiling and Fraud Detection in Mobile Telecommunications
 Detecting Payment Card Fraud with Neural Networks
 Money Laundering Detection with a NeuralNetwork
 Utilizing Fuzzy Logic and Neurofuzzy for Business Advantage
Complexity Control October 30, 2009
There are two issues that we have to avoid in Machine Learning:
 Overfitting
 Underfitting
Overfitting occurs when our model is heavily complex with so many degrees of freedom, that we can learn every detail of the training set. Such a model will have very high precision on the training set but will show very poor ability to predict outcomes of new instances, especially outside the domain of the training set.Dangerous for the overfitting:it will easily lead the predictions to the range that is far beyond the training data, and produce wild predictions in multilayer perceptrons even with noisefree data.The best way to avoid overfitting is to use lots of training data.
In a Neural Network if the depth is too much, the network will have many degrees of freedom and will learn every characteristic of the training data set. That means it will show a very precise outcome of the training set but will not be able to generalize the commonality of the training set to predict the outcome of new cases.
Underfitting occurs when the model we picked to describe the data is not complex enough, and has high error rate on the training set. There is always a tradeoff. If our model is too simple, underfitting could occur and if it is too complex, overfitting can occur.
Example
 Consider the example showed in the figure. We have a training set and we want to find a model which fits it the best. We can find a polynomial of high degree which almost passes through all the points in the training set. But, in fact the training set is coming from a line model. Now the problem is although the complex model has less error on the training set it diverges from the line in other ranges which we have no training point. Because of that the high degree polynomail has very poor predictive result on test cases. This is an example of overfitting model.
 Now consider a training set which comes from a polynomial of degree two model. If we model this training set with a polynomial of degree one, our model will have high error rate on the training set, and is not complex enough to describe the problem.
 Consider a simple classification example. If our classification rule takes as input only the colour of a fruit and concludes that it is a banana, then it is not a good classifier. The reason is that just because a fruit is a yellow, does not mean that it is a banana. We can add complexity to our model to make it a better classifier by considering more features typical of bananas, such as size and shape. If we continue to make our model more and more complex in order to improve our classifier, we will eventually reach a point where the quality of our classifier no longer improves, ie., we have overfit the data. This occurs when we have considered so many features that we have perfectly described the existing bananas, but if presented with a new banana of slightly different shape than the existing bananas, for example, it cannot be detected. This is the tradeoff; what is the right level of complexity?
Complexity Control  Nov 2, 2009
Overfitting occurs when the model becomes too complex and underfitting occurs when it is not complex enough, both of which are not desirable. To control complexity, it is necessary to make assumptions for the model before fitting the data. Assumptions that we can make for a model are with polynomials or a neural network. There are other ways as well.
We do not want a model to get too complex, so we control it by making an assumption on the model. With complexity control, we want a model or a classifier with a low error rate. The lecture will explain the tradeoff between Bias and variance for model complexity control.
How do we choose a good classifier?
Our goal is to find a classifier that minimizes the true error rate[math]\ L(h)[/math].
[math]\ L(h)=Pr\{h(x)\neq y\}[/math]
Recall the empirical error rate
[math]\ \hat L_{h}= \frac{1}{n} \sum_{i=1}^{n} I(h(x_{i}) \neq y_{i})[/math]
[math]\,h[/math] is a classifier and we want to minimize the error rate. So we apply [math]\displaystyle x_1[/math] to [math]\displaystyle x_n[/math] to [math]\displaystyle h[/math], and take the average to get the empirical true error rate estimation of probability that [math]h(x_{i}) \neq y_{i}[/math].
There is a downward bias to this estimate meaning that it is always less than the true error rate.
If there is a change in our complexity from low to high, our error rate is always decreasing. When we apply our model to the test data, our error rate will start to decrease to a point, but then it will increase since the model hasn't seen it before. This can be explained since training error will decrease when we fit the model better by increasing its complexity, but as we have seen, this complex model will not generalize well, resulting in a larger test error.
We use our test data (from the test sample line shown on Figure 2) to get our empirical error rate. Right complexity is defined as where error rate of the test data is minimum; and this is one idea behind complexity control.
We assume that we have samples [math]\,X_1, . . . ,X_n[/math] that follow some (possibly unknown) distribution. We want to estimate a parameter [math]\,f[/math] of the unknown distribution. This parameter may be the mean [math]\,E(X_i)[/math], the variance [math]\,var(X_i)[/math] or some other quantity.
The unknown parameter [math]\,f[/math] is a ﬁxed real number [math]f\in R[/math]. To estimate it, we use an estimator which is a function of our observations, [math]\hat{f}(X_1,...,X_n)[/math].
[math]Bias (\hat{f}) = E(\hat{f})  f[/math]
[math]MSE (\hat{f}) = E[(\hat{f}  f)^2]=Varince (\hat f)+Bias^2(\hat f )[/math]
[math]Variance (\hat{f}) = E[(\hat{f}  E(\hat{f}))^2][/math]
One property we desire of the estimator is that it is correct on average, that is, it is unbiased. [math]Bias (\hat{f}) = E(\hat{f})  f=0[/math]. However, there is a more important property for an estimator than just being unbiased: the mean squared error. In statistics, there are problems for which it may be good to use an estimator with a small bias. In some cases, an estimator with a small bias may have lesser mean squared error or be medianunbiased (rather than meanunbiased, the standard unbiasedness property). The property of medianunbiasedness is invariant under transformations while the property of meanunbiasedness may be lost under nonlinear transformations. For example, while using an unbiased estimator with large mean square error to estimate the parameter, we highly risk a big error. In contrast, a biased estimator with small mean square error will well improve the precision of our prediction.
Hence, our goal is to minimize [math]MSE (\hat{f})[/math].
From figure 3, we can see that the relationship of the three parameters is: [math]MSE (\hat{f})=Variance (\hat{f})+Bias ^2(\hat{f}) [/math]. Thus given the Mean Squared Error (MSE), if we have a low bias, then we will have a high variance and vice versa.
A Test error is a good estimation on MSE. We want to have a somewhat balanced bias and variance (not high on bias or variance), although it will have some bias.
Referring to Figure 2, overfitting happens after the point where training data (training sample line) starts to decrease and test data (test sample line) starts to increase. There are 2 main approaches to avoid overfitting:
1. Estimating error rate
[math]\hookrightarrow[/math] Empirical training error is not a good estimation
[math]\hookrightarrow[/math] Empirical test error is a better estimation
[math]\hookrightarrow[/math] CrossValidation is fast
[math]\hookrightarrow[/math] Computing error bound (analytically) using some probability inequality.
We will not discuss computing the error bound in class; however, a popular method for doing this computation is called VC Dimension (short for Vapnik–Chervonenkis Dimension). Information can be found from Andrew Moore and Steve Gunn.
2. Regularization
[math]\hookrightarrow[/math] Use of shrinkage method
[math]\hookrightarrow[/math] Decrease the chance of overfitting by controlling the weights
Example of under and overfitting in R
To give further intuition of over and underfitting, consider this example. A simple quadratic data set with some random noise is generated, and then polynomials of varying degrees are fitted. The errors for the training set and a test set are calculated.
>> x < rnorm(200,0,1) >> y < x^20.5*x+rnorm(200,0,0.3) >> xtest < rnorm(50,1,1) >> ytest < xtest^20.5*xtest+rnorm(50,0,0.3) >> p1 < lm(y~x) >> p2 < lm(y ~ poly(x,2)) >> pn < lm(y ~ poly(x,10)) >> psi < lm(y~I(sin(x))+I(cos(x)))

x
values for the training set are based on a [math]\,N(0,1)[/math] distribution, while the test set has a [math]\,N(1,1)[/math] distribution.y
values are determined by [math]\,y = x^2 + x + N(0,0.3)[/math], a quadratic function with some random variation. Polynomial least square fits of degree 1, 2, and 10 are calculated, as well as a fit of [math]\,sin(x)+cos(x)[/math].
>> > # calculate the mean squared error of degree 1 poly >> > sum((ypredict(p1,data.frame(x)))^2)/length(y) >> [1] 1.576042 >> > sum((ytestpredict(p1,data.frame(x=xtest)))^2)/length(ytest) >> [1] 7.727615
 Training and test mean squared errors for the linear fit. These are both quite high  and since the data is nonlinear, the different mean value of the test data increases the error quite a bit.
>> > # calculate the mean squared error of degree 2 poly >> > sum((ypredict(p2,data.frame(x)))^2)/length(y) >> [1] 0.08608467 >> > sum((ytestpredict(p2,data.frame(x=xtest)))^2)/length(ytest) >> [1] 0.08407432
 This fit is far better  and there is not much difference between the training and test error, either.
>> > # calculate the mean squared error of degree 10 poly >> > sum((ypredict(pn,data.frame(x)))^2)/length(y) >> [1] 0.07967558 >> > sum((ytestpredict(pn,data.frame(x=xtest)))^2)/length(ytest) >> [1] 156.7139
 With a highdegree polynomial, the training error continues to decrease, but not by much  and the test set error has risen again. The overfitting makes it a poor predictor. As the degree of the polynomial rises further, the accuracy of the computer becomes an issue  and a good fit is not even consistently produced for the training data.
>> > # calculate mse of sin/cos fit >> > sum((ypredict(psi,data.frame(x)))^2)/length(y) >> [1] 0.1105446 >> > sum((ytestpredict(psi,data.frame(x=xtest)))^2)/length(ytest) >> [1] 1.320404
 Fitting a function of the form sin(x)+cos(x) works pretty well on the training set, but because it is not the real underlying function, it fails on test data which doesn't lie on the same domain.
CrossValidation (CV)  Introduction
CrossValidation is used to estimate the error rate of a classifier with respect to test data rather than data used in the model. Here is a general introduction to CV:
[math]\hookrightarrow[/math] We have a set of collected data for which we know the proper labels
[math]\hookrightarrow[/math] We divide it into 2 parts, Training data (T) and Validation data (V)
[math]\hookrightarrow[/math] For our calculation, we pretend that we do not know the label of V and we use data in T to train the classifier
[math]\hookrightarrow[/math] We estimate an empirical error rate on V since the model hasn't seen V yet and we know the proper label of all elements in V to know how many were misclassified.
CV has different implementations which can reduce the variance of the calculated error rate, but sometimes with a tradeoff of a higher calculation time.
Complexity Control  Nov 4, 2009
Crossvalidation
Crossvalidation is the simplest and most widely used method to estimate the true error. It comes from the observation that although training error always decreases with the increasing complexity of the model, the test error starts to increase from a certain point, which is noted as overfitting (see figure 2 above). Since test error estimates MSE (mean square error) best, people came up with the idea to divide the data set into three parts: training set, validation set, and test set. training set is used to build the model, validation set is used to deside the parameters and the optimal model, and the test set is used to estimate the performance of the chosen model. A classical division is 50% for training set, and 25% each for validation set and test set. All of them are randomly selected from the original data set.
Training set: a set of examples used for learning: to fit the parameters of the classifier.
Validation set: a set of examples used to tune the parameters of a classifier.
Test set: a set of examples used only to assess the performance of a fully trained classifier.
Then, we only use the part of our data marked as the "training set" to train our algorithm, while keeping the remaining marked as the "validation set" untouched. As a result, the validation set will be totally unknown to the trained model. The error rate is then estimated by:
[math]\hat L(h) = \frac{1}{\nu}\sum_{X_i \in \nu}(h(x_i) \neq y_i)[/math], where [math]\,\nu[/math] is the cardinality of the validation set.
When we change the complexity, the error generated by the validation set will have the same behavior as the test set, so we are able to choose the best parameters to get the lowest error.
Kfold Crossvalidation
Above is the simplest form of complexity control. However, in reality, it may be hard to collect data ??and we usually suffer from the curse of dimensionality??, and a larger data set may be hard to come by. Consequently, we may not be able to afford to sacrifice part of the limited resources. In this case we use another method that addresses this problem, Kfold crossvalidation.The advantage of KFold Cross validation is that all the examples in the dataset are eventually used for both training and testing. We divide the data set into [math]\,K[/math] subsets roughly equal in size. The usual choice is [math]\,K = 10[/math].
Generally, how to choose [math]\,K[/math]:
if [math]\,K=n[/math], leave one out, low bias, high variance. Each subset contains a single element, so the model is trained with all except one point, and then validated using that point.
if [math]\,K=2[/math], say 2fold, 5fold, high bias, low variance. Each subset contains approximately [math]\,\frac{1}{2}[/math] or [math]\,\frac{1}{5}[/math] of the data.
For every [math]\,k[/math]th [math]( \,k \in [ 1, K ] )[/math] part, we use the other [math]\,K1[/math] parts to fit the model and test on the [math]\,k[/math]th part to estimate the prediction error [math]\hat L_k[/math], where
[math]\hat L(h) = \frac{1}{K}\sum_{k=1}^K\hat L_k[/math]
For example, suppose we want to fit a polynomial model to the data set and split the set into four equal subsets as shown in Figure 2. First we choose the degree to be 1, i.e. a linear model. Next we use the first three sets as training sets and the last as validation set, then the 1st, 2nd, 4th subsets as training set and the 3rd as validation set, so on and so forth until all the subsets have been the validation set once (all observations are used for both training and validation). After we get [math]\hat L_1, \hat L_2, \hat L_3, \hat L_4[/math], we can calculate the average [math]\hat L[/math] for degree 1 model. Similarly, we can estimate the error for n degree model and generate a simulating curve. Now we are able to choose the right degree which corresponds to the minimum error. Also, we can use this method to find the optimal unit number of hidden layers of neural networks. We can begin with 1 unit number, then 2, 3 and so on and so forth. Then find the unit number of hidden layers with lowest average error.
Generalized Crossvalidation
If the vector of observed values is denoted by [math]\mathbf{y}[/math], and the vector of fitted values by [math]\hat\mathbf{y}[/math].
[math]\mathbf{y} = \mathbf{H}\hat\mathbf{y}[/math],
where the hat matrix is given by
[math]\mathbf{H} = \mathbf{X}( \mathbf{X}^{T} \mathbf{X})^{1}\mathbf{X}^{T}[/math],
[math] \frac{1}{N}\sum_{i=1}^{N}[y_{i}  \hat f^{i}(\mathbf{x}_{i})]^{2}=\frac{1}{N}\sum_{i=1}^{N}[\frac{y_{i}\hat f(x_{i})}{1\mathbf{H}_{ii}}]^{2}[/math],
Then the GCV approximation is given by
[math] GCV(\hat f) = \frac{1}{N}\sum_{i=1}^{N}[\frac{y_{i}\hat f(x_{i})}{1trace(\mathbf{H})/N}]^{2}[/math],
Thus, one of the biggest advantages of the GCV is that the trace is more easily computed.
Leaveoneout Crossvalidation
Leaveoneout crossvalidation involves using all but one data point in the original training data set to train our model, then using the data point that we initially left out as a means to estimate true error. But repeating this process for every data point in our original data set, we can obtain a good estimation of true error.
In other words, leaveoneout crossvalidation is kfold crossvalidation in which we set the subset number [math]\,K[/math] to be the cardinality of the whole data set.
In the above example, we can see that kfold crossvalidation can be computationally expensive: for every possible value of the parameter, we must train the model [math]\,K[/math] times. This deficiency is even more obvious in leaveoneout crossvalidation, where we must train the model [math]\,n[/math] times, where [math]\,n[/math] is the number of data point in the data set.
Fortunately, when adding data points to the classifier is reversible, calculating the difference between two classifiers is computationally more efficient than calculating the two classifiers separately. So, if the classifier on all the data points is known, we simply undo the changes from a data point [math]\,K[/math] times to calculate the leaveoneout crossvalidation error rate.
How to decide the number of folds? For a large number of folds, the bias of the true error rate estimator will be small, the variance of it and the computing time will be large. For a small number of folds, everything will be opposite. When the datasets is large, 3fold cross validation will be enough, but if the datasets is very sparse we prefer to use leaveoneout.
Regularization for Neural Network — Weight Decay
Weight decay training is suggested as an implementation for achieving a robust neural network which is insensitive to noise. Since the number of hidden layers in NN is usually decided by certain domain knowledge, it may easily get into the problem of overfitting.
It can be seen from Figure 1 that when the weight is in the vicinity of zero, the operative part of the activation function shows linear behavior. The NN then collapses to an approximately linear model. Note that a linear model is the simplest model, we can avoid overfitting by constraining the weights to be small. This gives us a hint to initialize the random weights to be close to zero.
Formally, we penalize nonlinear weights by adding a penalty term in the error function. Now the regularized error function becomes:
[math]\,REG = err + \lambda(\sum_{i}w_i^2 + \sum_{jk}u_{jk}^2)[/math], where [math]\,err[/math] is the original error in backpropagation; [math]\,w_i[/math] is the weights of the output layer; [math]\,u_{jk}[/math] is the weights of the hidden layers.
Usually, too large [math]\,\lambda[/math] will make the weights [math]\,w_i[/math] and [math]\,u_{jk}[/math] too small. We can use cross validation to estimate [math]\,\lambda[/math].Another approach to choosing the [math]\,\lambda[/math] is to train several networks with different amounts of decay and estimates the generalization error for each; then choose the [math]\,\lambda[/math] that minimizes the estimated generalization error.
A similar penalty, weight elimination, is given by,
[math]\,REG = err + \lambda(\sum_{i}\frac{w_i^2}{1 + w_i^2} + \sum_{jk}\frac{u_{jk}^2}{1+u_{jk}^2})[/math].
As in backpropagation, we take partial derivative with respect to the weights:
[math]\frac{\partial REG}{\partial w_i} = \frac{\partial err}{\partial w_i} + 2\lambda w_i[/math]
[math]\frac{\partial REG}{\partial u_{jk}} = \frac{\partial err}{\partial u_{jk}} + 2\lambda u_{jk}[/math]
[math]w^{new} \leftarrow w^{old}  \rho\left(\frac{\partial err}{\partial w} + 2\lambda w\right)[/math]
[math]u^{new} \leftarrow u^{old}  \rho\left(\frac{\partial err}{\partial u} + 2\lambda u\right)[/math]
Note:
here [math]\,\lambda[/math] serves as a tradeoff parameter, tuning between the error rate and the linearity. Actually, we may also set [math]\,\lambda[/math] by crossvalidation. The tuning parameter is important since weights of zero will lead to zero derivatives and the algorithm will not change. On the other hand, starting with weights that are too large means starting with a nonlinear model which can often lead to poor solutions. <ref>Trevor Hastie, Robert Tibshirani, Jerome Friedman, Elements of Statistical Learning (Springer 2009) pp.398</ref>
We can standardize or normalize the inputs and targets, or adjust the penalty term for the standard deviations of all the inputs and targets in order to omit the biases and get good result from weight decay.
[math]\,\lambda[/math]is different for different types of weights in the NN. We can have different [math]\,\lambda[/math] for inputtohidden, hiddentohidden, and hiddentooutput weights.
Radial Basis Function (RBF) Networks  November 6, 2009
Introduction
A Radial Basis Function (RBF) network [19] is a type of artificial neural network with an output layer and a single hidden layer, with weights from the hidden layer to the output layer, and can be trained without back propagation since it has a closedform solution. The neurons in the hidden layer contain basis functions. One choice that has been widely used is that of radial basis functions, which have the property that each basis function depends only on the radial distance (typically Euclidean) from a center [math]\displaystyle\mu_{j}[/math], so that [math]\phi_{j}(x)= h({\Vert x  \mu_{j}\Vert})[/math].
RBFN were first used in solving multivariate interpolation problems and numerical analysis. Their prospect is similar in neural network applications, where the training and query targets are rather continuous. RBFN are artificial neural networks and it can be applied to Regression, Classification and Time series prediction.
[math]\ x_1 \cdot \cdot \cdot x_d[/math]: input layer of d dimension of training patterns
[math]\ \phi_1 \cdot \cdot \cdot\ phi_m[/math]: hidden layer of up to m locally tuned neurons centered over receptive fields
[math]\ y_1\cdot \cdot \cdot y_k[/math]: output layer that provides the response of the network
The output of an RBF network can be expressed as a weighted sum of its radial basis functions as follows:
[math]\hat y_{k} = \sum_{j=1}^M\phi_{j}(x) w_{jk}[/math]
The radial basis function is:
[math]\phi_{j}(x) = e^{\frac{\Vert x  \mu_{j}\Vert ^2}{2\sigma_{j}^2}}[/math]
(Gaussian without a normalization constant)
note:The hidden layer has a variable number of neurons (the optimal number is determined by the training process). As usual the more neurons in the hidden layer, the higher the model complexity. Each neuron consists of a radial basis function centered on a point with the same dimensions as the input data. The radii of the RBF functions may be different. The centers and radii can be determined through clustering or an EM algorithm. When the x vector is given from the input layer, the hidden neuron computes the radial distance from the neuron’s center point and then applies RBF function to this distance. The resulting value is passed to the the output layer and weighed together to form the output.
[math]\,y_{k}[/math] can be expressed in matrix form as:
[math]\hat Y = \Phi W [/math]
where
 [math]\hat{Y}_{n,k} = \left[ \begin{matrix} \hat{y}_{1,1} & \hat{y}_{1,2} & \cdots & \hat{y}_{1,k} \\ \hat{y}_{2,1} & \hat{y}_{2,2} & \cdots & \hat{y}_{2,k} \\ \vdots &\vdots & \ddots & \vdots \\ \hat{y}_{n,1} & \hat{y}_{n,2} & \cdots & \hat{y}_{n,k} \end{matrix}\right] [/math] is the matrix of output variables.
 [math]\Phi_{n,m} = \left[ \begin{matrix} \phi_{1,1} & \phi_{1,2} & \cdots & \phi_{1,m} \\ \phi_{2,1} & \phi_{2,2} & \cdots & \phi_{2,m} \\ \vdots & \vdots & \ddots & \vdots \\ \phi_{n,1} & \phi_{n,2} & \cdots & \phi_{n,m} \end{matrix}\right] [/math] is the matrix of Radial Basis Functions.
 [math]W_{m,k} = \left[ \begin{matrix} w_{1,1} & w_{1,2} & \cdots & w_{1,k} \\ w_{2,1} & w_{2,2} & \cdots & w_{2,k} \\ \vdots & \vdots & \ddots & \vdots \\ w_{m,1} & w_{m,2} & \cdots & w_{m,k} \end{matrix}\right] [/math] is the matrix of weights.
Here, [math]k[/math] is the number of outputs, [math]n[/math] is the number of data points, and [math]m[/math] is the number of hidden units. If [math]k = 1[/math], [math]\hat Y[/math] and [math]W[/math] are column vectors.
related reading:
Introduction of the Radial Basis Function (RBF) Networks [20]
Paper about the BBFN for multitask learning [21]
Radial Basis Function (RBF) Networks [22] [23] [24]
Advantage of RBFN: 1. First, it can model any nonlinear function using a single hidden layer, which removes some designdecisions about numbers of layers. 2.Second, the simple linear transformation in the output layer can be optimized fully using traditional linear modeling techniques
Estimation of weight matrix W
We minimize the training error, [math]\Vert Y  \hat{Y}\Vert^2[/math] in order to find [math]\,W[/math].
From a previous result in linear algebra we know that
[math]\Vert A \Vert^2 = Tr(A^{T}A)[/math]
Thus we have a problem similar to linear regression:
[math]\ err = \Vert Y  \Phi W\Vert^{2} = Tr[(Y  \Phi W)^{T}(Y  \Phi W)][/math]
[math]\ err = Tr[Y^{T}Y  Y^{T}\Phi W  W^{T} \Phi^{T} Y + W^{T}\Phi^{T} \Phi W][/math]
Useful properties of matrix differentiation
[math]\frac{\partial Tr(AX)}{\partial X} = A^{T}[/math]
[math]\frac{\partial Tr(X^{T}A)}{\partial X} = A[/math]
[math]\frac{\partial Tr(X^{T}AX)}{\partial X} = (A^{T} + A)X[/math]
Solving for W
We find the minimum over [math]\,W[/math] by setting [math]\frac{\partial err}{\partial W}[/math] equal to zero and using the aforementioned properties of matrix differentiation.
[math]\frac{\partial err}{\partial W} = 0[/math]
[math]\ 0  \Phi^{T}Y  \Phi^{T}Y + 2\Phi^{T}\Phi W = 0[/math]
[math]\ 2 \Phi^{T}Y + 2\Phi^{T}\Phi W = 0[/math]
[math]\ W = (\Phi^{T}\Phi)^{1}\Phi^{T}Y[/math]
[math]\hat{Y} = \Phi W = \Phi(\Phi^{T}\Phi)^{1}\Phi^{T}Y = HY[/math]
where [math]\ H = \Phi(\Phi^{T}\Phi)^{1}\Phi^{T}[/math]
[math]\,H[/math] is the hat matrix for this model. This gives us a nice results since the solution has a closed form and we do not have to worry about convexity problems in this case.
Including an additional bias
[math]\,y_{k}[/math] can be expressed in matrix form as:
[math]\hat Y = \Phi W [/math]
where
 [math]\hat Y = \left[ \begin{matrix} y_{11} & y_{12} & \cdots & y_{1k} \\ y_{21} & y_{22} & \cdots & y_{2k} \\ \vdots & & \ddots & \vdots \\ y_{n1} & y_{n2} & \cdots & y_{nk} \end{matrix}\right] [/math] is the matrix(n by k) of output variables.
 [math]\Phi = \left[ \begin{matrix} \phi_{10} &\phi_{11} & \phi_{12} & \cdots & \phi_{1M} \\ \phi_{20} & \phi_{21} & \phi_{22} & \cdots & \phi_{2M} \\ \vdots & & \ddots & \vdots \\ \phi_{n0} &\phi_{n1} & \phi_{n2} & \cdots & \phi_{nM} \end{matrix}\right] [/math] is the matrix(n by M+1) of Radial Basis Functions.
 [math]W = \left[ \begin{matrix} w_{01} & w_{02} & \cdots & w_{0k} \\ w_{11} & w_{12} & \cdots & w_{1k} \\ w_{21} & w_{22} & \cdots & w_{2k} \\ \vdots & & \ddots & \vdots \\ w_{M1} & w_{M2} & \cdots & w_{Mk} \end{matrix}\right] [/math] is the matrix(M+1 by k) of weights.
where the extra basis function [math]\Phi_{0}[/math] is set to 1.
Normalized RBF
In addition to the above unnormalized architecture, the normalized RBF can be represented as:
[math]\hat{y}_{k}(X) = \frac{\sum_{j=1}^{M} w_{jk}\Phi_{j}(X)}{\sum_{r=1}^{M}\Phi_{r}(X)}[/math]
Actually, [math]\Phi^{\ast}_{j}(X) = \frac{\Phi_{j}(X)}{\sum_{r=1}^{M}\Phi_{r}(X)}[/math] is known as a normalized radial basis function. Giving the familiar form,
[math]\hat{y}_{k}(X) = \sum_{j=1}^{M} w_{jk}\Phi^{\ast}_{j}(X)[/math]
Conceptualizing RBF networks
In the past, we have classified data using models that were explicitly linear, quadratic, or otherwise definite. In RBF networks, like in Neural Networks, we can fit an arbitrary model. How can we do this without changing the equations being used?
Recall a trick that was discussed in the October 7 lecture: if we add new features to our original data set, we can project into higher dimensions, use a linear algorithm, and get a quadratic result by collapsing to a lower dimension afterward. In RBF networks, something similar can happen.
Think of [math]\,\Phi[/math], our matrix of radial basis functions, as a feature space of the input. Each hidden unit, then, can be thought to represent a feature; we can see that, if there are more hidden units than input units, we can essentially project to a higherdimensional space, as we did in our earlier trick. However, this does not mean that an RBF network will actually do this, it is merely a way to convince yourself that RBF networks (and neural networks) can fit arbitrary models. Nevertheless, it is also noticed that just because of such great power, the problem of overfitting appears more important. We have to control its complexity so that it does not fit to an arbitrary training model but to a general one.
RBF networks for classification  a probabilistic paradigm
An RBF network is akin to fitting a Gaussian mixture model to data. We assume that each class can be modelled by a single function [math]\,\phi[/math] and data is generated by a mixture model. According to Bayes Rule,
[math]Pr(Y = y_{k}  X = x) = \frac {Pr(xy_{k})*Pr(y_{k})}{Pr(x)}[/math]
While all classifiers that we have seen thus far in the course have been in discriminative form, the RBF network is a generative model that can be represented using a directed graph.
We can replace the class conditional density in the above conditional probability expression by marginalizing [math]\,x[/math] over [math]\,j[/math]: [math]\Pr(xy_{k}) = \sum_{j} Pr(xj)*Pr(jy_{k})[/math]
 Note We made the assumption that each class can be modelled by a single function [math]\displaystyle\Phi[/math] and that the data was generated by a mixture model. The Gaussian mixture model has the form:
[math]f(x)=\sum_{m=1}^M \alpha_m \phi(x;\mu_m,\Sigma_m)[/math] where [math]\displaystyle\alpha_m[/math] are mixing proportions, [math]\displaystyle\sum_m \alpha_m=1[/math], and [math]\displaystyle\mu_m[/math] and [math]\displaystyle\Sigma_m[/math] are the mean and covariance of each Gaussian density respectively. <ref>H. Trevor, R. Tibshirani, J. Friedman, The Elements of Statistical Learning (Springer 2009), pp. 214. </ref> The generative model in Figure 1 shows graphically how each Gaussian in the mixture model is chosen to sample from.
Radial Basis Function (RBF) Networks  November 9th, 2009
RBF Network for classification (A probabilistic point of view)
Using RBF Network[25] to do classification, we usually treat it as regression problem. We want to set a threshold to decide what the data’s class membership is. However, to find some insight to the classification problem of what we are doing in terms of RBF Network, we often think of mixture models and make certain assumptions. Before we mainly using deterministic model to describe data, which means a given input always generate the same output, now we are going to consider generative model of data. In this case, some hidden variables are incorporated and joint probability is assigned between the nodes so that we can derive results through the bayes rule.
We assume, as we can see in the graph on the right hand side, that we have three random variables, [math]\displaystyle y_k[/math], [math]\displaystyle j[/math], and [math]\displaystyle x[/math] where [math]\displaystyle y_k[/math] denotes class [math]\,k[/math], [math]\displaystyle x[/math] is what we observed here, and [math]\displaystyle j[/math] is some hidden random variable. There is a process that there are different classes, and each class can trigger a different hidden random variable [math]\displaystyle j[/math]. To understand this, we can assume that, for instance, this random variable [math]\displaystyle j[/math] has a Gaussian distribution (it could have any other distribution as well) and that all the [math]\displaystyle j[/math]’s have the same distribution (Gaussian), but with different parameters. From each Gaussian distribution triggered by each class, we are going to sample some data points. Therefore, in the end, we are going to get a set of data, which are not strictly Gaussian, but are actually a mixture of Gaussians.
Again, we look at the posterior distribution from Bayes' Rule.
[math]Pr(Y = y_{k}  X = x) = \frac {Pr(X = x  Y = y_{k})*Pr(Y = y_{k})}{Pr(X = x)}[/math]
Since we made the assumption that the data has been generated from a mixture model, we can estimate this conditional probability by
[math]\Pr(X = x  Y = y_{k}) = \sum_{j} Pr(X = x  j)*Pr(j  Y = y_{k})[/math],
which is the class conditional distribution (or probability) of the mixture model. Note, here, if we only have a simple model from [math]\displaystyle y_k[/math] to [math]\displaystyle x[/math], then we won’t have this summation.
We can substitute this class conditional distribution into Bayes' formula. We can see that the posterior of class [math]\displaystyle k[/math] is the summation over [math]\displaystyle j[/math] of the probability of [math]\displaystyle x[/math] given [math]\displaystyle j[/math] times the probability of [math]\displaystyle j[/math] given [math]\displaystyle y_k[/math], times the prior distribution of class [math]\displaystyle k[/math], and lastly divided by the marginal probability of [math]\displaystyle x[/math]. That is,
[math]\Pr(y_k  x) = \frac {\sum_{j} Pr(x  j)*Pr(j  y_{k})*Pr(y_{k})}{Pr(x)}[/math].
Since, the prior probability of class [math]\displaystyle k[/math], [math]\displaystyle Pr(y_{k})[/math], does not have an index of [math]\displaystyle j[/math], it can be taken out of the summation. This yields,
[math]\Pr(y_k  x) = \frac {Pr(y_{k})\sum_{j} Pr(x  j)*Pr(j  y_{k})}{Pr(x)}[/math].
We multiply this by [math]\displaystyle 1 = \frac {Pr(j)}{Pr(j)}[/math]. Then, it becomes,
[math]\Pr(y_k  x) = \frac {Pr(y_{k})\sum_{j} Pr(x  j)*Pr(j  y_{k})}{Pr(x)} * \frac {Pr(j)}{Pr(j)}[/math].
Next, note that [math]\displaystyle Pr(j  x) = \frac {Pr(x  j)*Pr(j)}{Pr(x)}[/math], and [math]\displaystyle Pr(y_k  j) = \frac {Pr(j  y_k)*Pr(y_k)}{Pr(j)}[/math]. Then rearranging the terms, we finally have the posterior:
[math]\displaystyle Pr(y_k  x) = \sum_{j} Pr(j  x)Pr(y_k  j)[/math].
where [math]\displaystyle Pr(j  x) [/math] is the probability of future given data, [math]\displaystyle Pr(y_k  j) [/math] is the probability of class membership given a future.
Interestingly, this is just the product of the posterior of the two functions that are summed.
Interpretation of RBF Network classification
We want to relate the results that we derived above to our RBF Network. In a RBF Network, as we can see on the right hand side, we have a set of data, [math]\displaystyle x_1[/math] to [math]\displaystyle x_d[/math], and the hidden basis function, [math]\displaystyle \phi_{1}[/math] to [math]\displaystyle \phi_{M}[/math], and then we have some output, [math]\displaystyle y_1[/math] to [math]\displaystyle y_k[/math]. Also, we have weights from the hidden layer to output layer. The output is just the linear sum of [math]\displaystyle \phi[/math]’s.
Now consider probability of [math]\displaystyle j[/math] given [math]\displaystyle x[/math] to be [math]\displaystyle \phi[/math], and the probability of [math]\displaystyle y_k[/math] given [math]\displaystyle j[/math] to be the weights [math]\displaystyle w_{jk}[/math], then the posterior can be written as,
[math]\displaystyle Pr(y_k  x) = \sum_{j} \phi_{j}(x)*w_{jk}[/math].
Now, let us look at an example in one dimensional case. Suppose,
[math]\phi_{j}(x) = e^{\frac{\Vert x  \mu_{j}\Vert ^2}{2\sigma_{j}^2}}[/math], and [math]\displaystyle j[/math] is from 1 to 2.
We know that [math]\displaystyle \phi[/math] is a radial basis function. It's as if we put some Gaussian over data. And for each Gaussian, we consider the center [math]\displaystyle \mu[/math]. Then, what [math]\displaystyle \phi[/math] computes is the similarity of any data point to the center.
We can see the graph on the left which plots the density of [math]\displaystyle \phi_{1}[/math] and [math]\displaystyle \phi_{2}[/math]. Take [math]\displaystyle \phi_{1}[/math] for instance, if the point gets far from the center [math]\displaystyle \mu_{1}[/math], then it will reduce [math]\displaystyle \phi_{1}[/math] to become nearly zero. Remember that, we can usually find a nonlinear regression or classification of input space by doing a linear one in some extended space or some feature space (more details in Aside). Here, the [math]\displaystyle \phi[/math]’s actually produce that feature space.
So, one way to look at this is that this [math]\displaystyle \phi[/math] is telling us that given an input, how likely the probability of presence of a particular feature is. Say, for example, we define the features as the centers of these Gaussian distributions. Then, this [math]\displaystyle \phi[/math] function somehow computes the possibility given certain data points, of this kind of feature appearing. If the data point is right at the center, then the value of that [math]\displaystyle \phi[/math] would be one, i.e. the probability is 1. If the point is far from the center, then the probability ([math]\displaystyle \phi[/math] function value) will be close to zero, that is, it’s less likely. Therefore, we can treat [math]\displaystyle Pr(j  x)[/math] as the probability of a particular feature given data.
When we have those features, then [math]\displaystyle y[/math] is the linear combination of the features. Hence, any of the weights [math]\displaystyle w[/math], which is equal to [math]\displaystyle Pr(y_k  j)[/math], tells us how likely this particular [math]\displaystyle y[/math] will appear given those features. Therefore, the weight [math]\displaystyle w_{jk}[/math] shows the probability of class membership given feature.
Hence, we have found a probabilistic point of view to look at RBF Network!
 Note There are some inconsistencies with this probabilistic point of view. There are no restrictions that force [math]\displaystyle Pr(y_k  x) = \sum_{j} \phi_{j}(x)*w_{jk}[/math] to be between 0 and 1. So if least squares is used to solve this, [math]\displaystyle w_{jk}[/math] cannot be interpreted as a probability.
Aside
 Feature Space:
 One way to produce a feature space is LDA
 Suppose, we have n data points [math]\mathbf{x}_1[/math] to [math]\mathbf{x}_n [/math]. Each data point has d features. And these n data points consist of the [math]X[/math] matrix,
 [math]X = \left[ \begin{matrix} x_{11} & x_{21} & \cdots & x_{n1} \\ x_{12} & x_{22} & \cdots & x_{n2} \\ \vdots & & \ddots & \vdots \\ x_{1d} & x_{2d} & \cdots & x_{nd} \end{matrix}\right] [/math]
 Also, we have feature space,
 [math]\Phi^{T} = \left[ \begin{matrix} \phi_{1}(\mathbf{x_1}) & \phi_{1}(\mathbf{x_2})& \cdots & \phi_{1}(\mathbf{x_n})\\ \phi_{2}(\mathbf{x_1})& \phi_{2}(\mathbf{x_2})& \cdots & \phi_{2}(\mathbf{x_n}) \\ \vdots & & \ddots & \vdots \\ \phi_{M}(\mathbf{x_1}) & \phi_{M}(\mathbf{x_2}) & \cdots & \phi_{M}(\mathbf{x_n}) \end{matrix}\right] [/math]
 If we want to solve a regression problem for the input data, we don’t perform Least Square on this [math]\displaystyle X[/math] matrix, we do Least Square on the feature space, i.e. on the [math]\displaystyle \Phi^{T}[/math] matrix. The dimensionality of [math]\displaystyle \Phi^{T}[/math] is M by n. We can add [math]\ \Phi_0=1 [/math] which is not any function of [math]\ x_1 \cdot \cdot \cdot x_j [/math]
 Now, we still have n data points, but we define these n data points in terms of a new set of features. So, originally, we define our data points by d features, but now, we define them by M features. And what are those M features telling us?
 Let us look at the first column of [math]\displaystyle \Phi^{T}[/math] matrix. The first entry is [math]\displaystyle \phi_1[/math] applied to [math]\mathbf{x_1}[/math], and so on, until the last entry is [math]\displaystyle \phi_M[/math] applied to [math]\mathbf{x_1}[/math]. Suppose each of these [math]\displaystyle \phi_j[/math] is defined by
 [math]\phi_{j}(x) = e^{\frac{\Vert x  \mu_{j}\Vert ^2}{2\sigma_{j}^2}}[/math].
 Then, each [math]\displaystyle \phi_j[/math] checks the similarity of the data point with its center. Hence, the new set of features are actually representing M centers in our data set, and for each data point, its new features check how this point is similar to the first center; how it is similar to the second center; and how it is similar to the [math]\displaystyle M^{th}[/math] center. And this checking process will apply to all data points. Therefore, feature space gives another representation of our data set.
Methords for selecting center [math]\ \mu [/math]:
 Subsampling: Randomlychosen training points are copied to the radial units. Since they are randomly selected, they will represent the distribution of the training data in a statistical sense.
 KMeans algorithm: Given K radial units, it adjusts the positions of the centers so that: Each training point belongs to a cluster center, and is nearer to this center than to any other center; Each cluster center is the centroid of the training points that belong to it.
The size of the deviation (smoothing factor) determines how spiky the Gaussian functions are. Deviations should typically be chosen so that Gaussians overlap with a few nearby centers.
Methods for choosing deviation are:
 Choose the deviation by oursleves.
 Select the deviation to reflect the number of centers and the volume of space they occupy
 KNearest Neighbor algorithm: Each unit's deviation is individually set to the mean distance to its K nearest neighbors.
If the Gaussians are too spiky, the network will not interpolate between known points, and the network loses the ability to generalize. If the Gaussians are very broad, the network loses fine detail.
useful resouces:
1.some examples, advantages & disadvantages:Basis Function (RBF) Networks
2.comparision between BP & RBF:[26]
Model selection or complexity control for RBF Network  a brief introduction
In order to obtain a better fit for the training data, we often want to increase the complexity of our RBF Network. By its construction, we know that to change the complexity of a RBF Network, the only way is to add or decrease the number of basis functions. A large number of basis function yields a more complex network. In theory, if we add enough basis functions, the RFB Network would work for any training; however, it doesn't mean this RBF Network model can generalize well. Therefore, for the purpose of avoiding overfitting problem (see Notes below), we only want to increase the number of basis function to certain point, i.e. its optimal level.
For the model selection, what we usually do is estimate the training error. After working through the training error, we’ll see that the training error in fact can be decomposed, and one component of training error is called Mean Squared Error (MSE). In the later notes, we will find that our final goal is to get a good estimate of MSE. Moreover, in order to find an optimal model for our data, we select the model with the smallest MSE.
Now, let us introduce some notations that we will use in the analysis:
 [math]\hat f[/math]  the prediction model estimated by a RBF network from the training data
 [math]\displaystyle f[/math]  the real model (not null), and ideally, we want [math]\hat f[/math] to be close to [math]\displaystyle f[/math]
 [math]\displaystyle err[/math]  the training error
 [math]\displaystyle Err[/math]  the testing error
 [math]\displaystyle MSE[/math]  the Mean Squared Error
Notes
 Being more complex isn’t always a good thing. Sometime, overfitting causes the model to lose its generality. For example in the graph on left hand side, the data points are sampled from the model [math]\displaystyle y_i= f(x_i)+\epsilon_i[/math], where [math]\displaystyle f(x_i)[/math] is a linear function, which is shown by the blue line, and [math]\displaystyle \epsilon_i[/math]is additive Gaussian noise from [math]~N(0,\sigma^2)[/math]. The red curve displayed in the graph shows the overfitted model. Clearly, this overfitted model only works for any training data, and is useless for any further prediction when new data points are introduced.
> n<20; > x<seq(1,10,length=n); > alpha<2.5; > beta<1.75; > y<alpha+beta*x+rnorm(n); > plot(y~x, pch=16, lwd=3, cex=0.5, main='Overfitting'); > abline(alpha, beta, col='blue'); > lines(spline(x, y), col = 2);
 More details on this topic later on.
Model Selection(Stein's Unbiased Risk Estimate) November 11th, 2009
Model Selection
Model selection is a task of selecting a model of optimal complexity for a given data. Learning a radial basis function network from data is a parameter estimation problem. One difficulty with this problem is selecting parameters that show good performance on both training and testing data. In principle, a model is selected to have parameters associated with the best observed performance on training data, although our goal really is to achieve good performance on unseen testing data. Not surprisingly, a model selected on the basis of training data does not necessarily exhibit comparable performance on the testing data. When squared error is used as the performance index, a zeroerror model on the training data can always be achieved by using a sufficient number of basis functions.
But, training error and testing error do not demonstrate a linear relationship. In particular, a smaller training error does do not necessarily result in a smaller testing error. In practice, one often observes that, up to a certain point, the model error on testing data tends to decrease as the training error decreases. However, if one attempts to decrease the training error too far by increasing model complexity, the testing error often can take a dramatic increase.
The basic reason behind this phenomenon is that in the process of minimizing training error, after a certain point, the model begins to overfit the training set. Overfitting in this context means fitting the model to training data at the expense of losing generality. In the extreme form, a set of [math]\displaystyle N[/math] training data points can be modeled exactly with [math]\displaystyle N[/math] radial basis functions. Such a model follows the training data perfectly. However, the model is not representative features of the true underlying data source, and this is why it fails to correctly model new data points.
In general, the training error rate will be less than the testing error on the new data. A model typically adapts to the training data, and hence the training error will be an overly optimistic estimate of the testing error. An obvious way to well estimate testing err is to add a penalty term to the training error to compensate. Actually, SURE is developed based on this.
Stein's unbiased risk estimate (SURE)
Important Notation[27]
Let:
 [math]\hat f(X)[/math] denote the prediction model, which is estimated from a training sample by the RBF neural network model.
 [math]\displaystyle f(X)[/math] denote the true model.
 [math]\displaystyle err=\sum_{i=1}^N (\hat y_iy_i)^2 [/math] denote the training error,which is the average loss over the training sample.
 [math]\displaystyle Err=\sum_{i=1}^M (\hat y_iy_i)^2 [/math] denote the test error, which is the expected prediction error on an independent test sample.
 [math]\displaystyle MSE=E(\hat ff)^2[/math] denote the mean squared error, where [math]\hat f(X)[/math] is the estimated model and [math]\displaystyle f(X)[/math] is the true model.
The BiasVariance Decomposition:
 [math] \begin{align} \displaystyle MSE = E(\hat ff)^2 &= E[(\hat fE(\ hat f))+(E(\hat f)f)]^2\\ &= E[(\hat fE(\hat f))^2+2*(\hat fE(\hat f))*(E(\hat f)f)+(E(\hat f)f)^2]\\ &= E[(\hat fE(\hat f))^2]+E[2*(\hat fE(\hat f))*(E(\hat f)f)]+E[(E(\hat f)f)^2]\\ &= Var(\hat f)+Bias^2(\hat f) \end{align} [/math]
Since, [math]\displaystyle E[2*(\hat fE(\hat f))*(E(\hat f)f)]=2*Cov[E(\hat f)f, \hat fE(\hat f)][/math], which is equal to zero.
Suppose the observations [math]\displaystyle y_i= f(x_i)+\epsilon_i[/math], where [math]\displaystyle \epsilon_i[/math]is additive Gaussian noise [math]~N(0,\sigma^2)[/math].We need to estimate [math]\hat f[/math] from training data set [math]T=\{(x_i,y_i)\}^2_{i=1}[/math]. Let [math]\hat f_i=\hat f_(x_i)[/math] and [math]\displaystyle f_i= f(x_i)[/math]， then
[math]\displaystyle E[(\hat y_iy_i)^2 ]=E[(\hat f_if_i\epsilon_i)^2][/math][math]=E[(\hat f_if_i)^2]+E[\epsilon_i^2]2E[\epsilon_i(\hat f_if_i)][/math]
[math]\displaystyle E[(\hat y_iy_i)^2 ]=E[(\hat f_if_i)^2]+\sigma^22E[\epsilon_i(\hat f_if_i)][/math] [math]\displaystyle (1)[/math]
The last term can be written as:
[math]\displaystyle E[\epsilon_i(\hat f_if_i)]=E[(y_if_i)(\hat f_if_i)]=cov(y_i,\hat f)[/math], where[math]\displaystyle y_i[/math] and [math]\hat f_i[/math] both have same mean [math]\displaystyle f_i[/math].
Stein's Lemma
If [math]\,Z[/math] is [math]\,N(\mu,\sigma^2)[/math] and if [math]\displaystyle g(Z)[/math] is weakly differentiable,such that[math]\displaystyle E[\vert g'(Z)\vert]\lt \infty[/math], then [math]\displaystyle E[g(Z)(Z\mu)]=\sigma^2E(g'(Z))[/math].
According to Stein's Lemma, the last cross term of [math]\displaystyle (1)[/math], [math]\displaystyle E[\epsilon_i(\hat f_if_i)][/math] can be written as [math]\sigma^2 E[\frac {\partial \hat f}{\partial y_i}][/math]. The derivation is as follows.
[math]\displaystyle Proof[/math]: Let [math]\,Z = \epsilon[/math]. Then [math]g(Z) = \hat ff[/math], since [math]\hat y = f + \epsilon[/math], and [math]\,f[/math] is a constant. So [math]\,\mu = 0[/math] and [math]\,\sigma^2[/math] is the variance in [math]\,\epsilon[/math]. [math]\displaystyle E[g(Z)(Z\mu)]=E[(\hat ff)\epsilon]=\sigma^2E(g'(Z))=\sigma^2 E[\frac {\partial (\hat ff)}{\partial y_i}]=\sigma^2 E[\frac {\partial \hat f}{\partial y_i}\frac {\partial f}{\partial y_i}][/math]
Since [math]\displaystyle f[/math] is the true model, not the function of the observations [math]\displaystyle y_i[/math], then [math]\frac {\partial f}{\partial y_i}=0[/math].
So,[math]\displaystyle E[\epsilon_i(\hat f_if_i)]=\sigma^2 E[\frac {\partial \hat f}{\partial y_i}][/math] [math]\displaystyle (2)[/math]
Two Different Cases
SURE in RBF, Automatic basis selection for RBF networks using Stein’s unbiased risk estimator,Ali Ghodsi Dale Schuurmans
Case 1
Consider the case in which a new data point has been introduced to the estimated model, i.e. [math](x_i,y_i)\not\in\tau[/math]; this new point belong to the validation set [math]\displaystyle \nu[/math],i.e.[math](x_i,y_i)\in\nu[/math]. Since [math]\displaystyle y_i[/math] is a new point, [math]\hat f[/math] and [math]\displaystyle y_i[/math] are independent. Therefore [math]\displaystyle cov(y_i,\hat f)=0[/math] (or think about [math]\frac{\partial \hat f}{\partial y_i}[/math], when [math]\,y_i[/math]is a new point, then it has nothing with [math]\hat f[/math] because the estimation of [math]\hat f[/math] is from the training data, so [math]\frac{\partial \hat f}{\partial y_i}=0[/math] ) and [math]\displaystyle (1)[/math] in this case can be written as:
[math]\displaystyle E[(\hat y_iy_i)^2 ]=E[(\hat f_if_i)^2]+\sigma^2[/math].
This expectation means [math]\frac {1}{m}\sum_{i=1}^m (\hat y_iy_i)^2 = \frac {1}{m}\sum_{i=1}^m (\hat f_if_i)^2+ \sigma^2[/math].
[math]\sum_{i=1}^m (\hat y_iy_i)^2 = \sum_{i=1}^m (\hat f_if_i)^2+ m\sigma^2[/math]
Based on the notation we denote above, then we obtain: [math]\displaystyle MSE=Errm\sigma^2[/math]
This is the justification behind the technique of cross validation. since [math]\displaystyle \sigma^2[/math] is constant, to minimize [math]\displaystyle MSE[/math] is equal to minimize the test err [math]\displaystyle Err[/math]. In cross vaildation to avoid overfitting or underfitting, a validation data set is independent from the estimated model.
Case 2
A more interesting case is the case in which we do not use new data points to assess the performance of the estimated model. and the training data is used for both estimating and assessing a model [math]\hat f_i[/math]. In this case the cross term in [math]\displaystyle (1)[/math] cannot be ignored because [math]\hat f_i[/math] and [math]\displaystyle y_i[/math] are not independent. Therefore the cross term can be estimated by Stein's lemma, which was originally proposed to estimated the mean of a Guassian distribution.
Suppose [math](x_i,y_i)\in\tau[/math], then by applying Stein's lemma, we obtain [math]\displaystyle (2)[/math] proved above.
[math]\displaystyle E[(\hat y_iy_i)^2 ]=E[(\hat f_if_i)^2]+\sigma^22\sigma^2E[\frac {\partial \hat f}{\partial y_i}][/math].
This expectation means [math]\frac {1}{N}\sum_{i=1}^N (\hat y_iy_i)^2 = \frac {1}{N}\sum_{i=1}^N (\hat f_if_i)^2+ \sigma^2\frac {2}{N}\sum_{i=1}^N \frac {\partial \hat f}{\partial y_i} [/math].
[math]\sum_{i=1}^N (\hat y_iy_i)^2 = \sum_{i=1}^N (\hat f_if_i)^2+ N\sigma^22\sigma^2\sum_{i=1}^N \frac {\partial \hat f}{\partial y_i} [/math].
[math]\displaystyle err=MSE+N\sigma^22\sigma^2\sum_{i=1}^N \frac {\partial \hat f}{\partial y_i}[/math]
[math]\displaystyle MSE=errN\sigma^2+2\sigma^2\sum_{i=1}^N \frac {\partial \hat f}{\partial y_i}[/math] [math]\displaystyle (3)[/math]
In statistics, this is known as Stein's unbiased risk estimate (SURE) is an unbiased estimator of the meansquared error of a given estimator, in a deterministic estimation scenario. In other words, it provides an indication of the accuracy of a given estimator. This is important since, in deterministic estimation, the true meansquared error of an estimator generally depends on the value of the unknown parameter, and thus cannot be determined completely.
SURE for RBF Network
Based on SURE, the optimum number of basis functions should be assigned to have the minimum generalization err [math]\displaystyle err[/math]. Based on the Radial Basis Function Network, by setting [math]\frac{\partial err}{\partial W}[/math] equal to zero , we get the least squared solution of[math]\ W = (\Phi^{T}\Phi)^{1}\Phi^{T}Y[/math].Then we have [math]\hat{Y} = \Phi W = \Phi(\Phi^{T}\Phi)^{1})\Phi^{T}Y = HY[/math],where [math]\ H = \Phi(\Phi^{T}\Phi)^{1})\Phi^{T}[/math], [math]\,H[/math] is the hat matrix for this model.
[math]\hat f=\,H_{i1}y_2+\,H_{i2}y_2+\cdots+\,H_{in}y_n[/math] [math]\displaystyle (3)[/math]
where [math]\,H[/math] depends on the input vector [math]\displaystyle x_i[/math] but not on [math]\displaystyle y_i[/math].
By taking the derivative of [math]\hat f_i[/math] with respect to [math]\displaystyle y_i[/math], we can easily obtain:
[math]\sum_{i=1}^N \frac {\partial \hat f}{\partial y_i}=\sum_{i=1}^N \,H_{ii}[/math]
Now, substituing this into[math]\displaystyle (3)[/math] ,we get
[math]\displaystyle MSE=errN\sigma^2+2\sigma^2\sum_{i=1}^N \,H_{ii}[/math]
Here, we can tell that [math]\sum_{i=1}^N \,H_{ii}= \,Trace(H)[/math], the sum of the diagonal elements of [math]\,H[/math]. Thus, we can obtain the further simplification that [math]\,Trace(H)= Trace(\Phi(\Phi^{T}\Phi)^{1})\Phi^{T})= Trace(\Phi^{T}\Phi(\Phi^{T}\Phi)^{1})=d[/math], where[math]\displaystyle d[/math] is the dimension of [math]\displaystyle \Phi[/math]. Since [math]\displaystyle \Phi[/math] is a projection of input matrix [math]\,X[/math] onto a basis set spanned by [math]\,M[/math], the number of basis functions. If considering intercept, then [math]\,Trace(H)= M+1[/math].
Then,[math]\displaystyle MSE=errN\sigma^2+2\sigma^2(M+1)[/math].
SURE Algorithm
We use this method to find the optimum number of basis function by choosing the model with smallest MSE over the set of models considered. Given a set of models [math]\hat f_M(x)[/math] indexed by the number of basis functions, [math]\displaystyle err(M)[/math].
Then, [math]\displaystyle MSE(M)=err(M)N\sigma^2+2\sigma^2(M+1)[/math]
where [math]\displaystyle N[/math] is the number of training samples and the noise,[math]\sigma^2[/math], can be estimated from the training data as
[math]\hat \sigma^2=\frac {1}{N1}\sum_{i=1}^N (\hat yy)^2[/math].
By applying SURE algorithm to SPECT Heart data, we get the optimal number of basis functions is [math]\displaystyle M=4[/math].
Pls take a look at figure 27.1 on the right, which shows that[math]\displaystyle MSE[/math] is smallest when [math]\displaystyle M=4[/math].
Calculating the SURE value is easy if you have access to [math]\,\sigma[/math].
sure_Err = error  num_data_point * sigma .^ 2 + 2 * sigma .^2 * (num_basis_functions + 1);
If [math]\,\sigma[/math] is not known, it can be estimated using the error.
error = (output  expected_output) .^ 2; sigma = (dot(err, ones(1, num_data_point)) / (num_data_point)) / (num_data_point  1); sure_Err = error  num_data_point * sigma .^ 2 + 2 * sigma .^2 * (num_basis_functions + 1);
SURE for RBF network & Support Vector Machine  November 13th, 2009
SURE for RBF network
Minimizing MSE
By Stein's unbiased risk estimate (SURE) for Radial Basis Function (RBF) Network we get:
[math]\displaystyle MSE=errN\sigma^2+2\sigma^2(M+1) [/math] (28.1)
 [math]\displaystyle MSE[/math](mean square error)= [math]\sum_{i=1}^N (\hat y_iy_i)^2 [/math]
 [math]\displaystyle err[/math](training error)= [math]\sum_{i=1}^N (\hat f_if_i)^2 [/math]
 [math]\displaystyle (M+1) [/math]( number of hidden units)= [math]\sum_{i=1}^N \frac {\partial \hat f}{\partial y_i} [/math]
Goal: To minimize MSE
1. If [math]\displaystyle \sigma [/math] is known, then it is no impact (i.e. a constant), and we can ignore it. Only need to minimize [math]\displaystyle MSE=err +2\sigma^2(M+1)[/math].
2. In reality, we do not know [math]\displaystyle \sigma[/math], and the estimate [math]\,\hat \sigma[/math] changes when [math]\displaystyle (M+1) [/math] changes. However, we can estimate [math]\displaystyle \sigma [/math].
[math]\displaystyle y_i= f(x_i)+\epsilon_i[/math], where [math]\displaystyle \epsilon_i[/math]is additive Gaussian noise [math]~N(0,\sigma^2)[/math]. Suppose we do not know the variance of [math]\displaystyle \epsilon[/math]. Then,
[math]\displaystyle \sigma^2=\frac{1}{N1}\sum_{i=1}^N (\hat y_iy_i)^2 =\frac{1}{N1}err[/math] (28.2)
Substitute (28.2) into (28.1), get
[math]\displaystyle MSE=errN\frac{1}{N1}err+2\frac{1}{N1}err(M+1)[/math]
[math]\displaystyle MSE=err(1\frac{N}{N1}+\frac{2(M+1)}{N1})[/math]
[math]\displaystyle MSE=err(\frac{N1N+2M+2}{N1})[/math]
[math]\displaystyle MSE=err(\frac{2M+1}{N1}) [/math] (28.3)
Figure 28.1: the training error will decrease and the MSE will increase when increasing the number of hidden units (i.e. the model is more complex).
When the number of hidden units gets larger and larger, the training error will decrease until it approaches to [math]\displaystyle 0 [/math]. If training error equals [math]\displaystyle 0 [/math], then no matter how large [math]\displaystyle (M+1) [/math] is, from (28.3) we can see the estimate of MSE will approach [math]\displaystyle 0 [/math] as well. However, in fact it does not happen since when training error is close to [math]\displaystyle 0 [/math] overfitting happens, and MSE should increase instead of being close to [math]\displaystyle 0 [/math]. We can see it from the Figure 28.1.
We can see [math]\displaystyle \sigma^2 [/math] is the average of [math]\displaystyle err [/math]. In order to deal with this problem, we can take the average for [math]\displaystyle err[/math] of each hidden unit. For example: we can first take 1 hidden unit, and take 10 hidden units in the next. Since in reality the value of [math]\, \sigma^2[/math] is a constant adjustment to the data points, and doesn't depend on [math]\,M+1[/math], using the average [math]\,\sigma^2[/math] value for 1 to 10 hidden units has a firm theoretical basis.
We can also see that unlike the classical Cross Validation (CV) or Leave one out (LOO) techniques, the SURE technique does not need to do the validation to find the optimal model. Hence, SURE technique uses less data than CV or LOO. It is suitable for the case that there is not enough data for validation. However, to implement SURE we need to find [math]\frac {\partial \hat f}{\partial y_i}[/math], which may not be trivial for models that do not have a closedform solution.
Kmeans Clustering
Description:
Kmeans clustering is a method of cluster analysis which aims to partition n observations into k clusters in which each observation belongs to the cluster with the nearest mean.
 The number of hidden units is equal to the number of clusters, which is [math]\displaystyle \phi [/math]
 [math]\displaystyle \phi_{j}(x) = e^{\frac{\Vert x  \mu_{j}\Vert ^2}{2\sigma_{j}^2}}[/math], we set it equal for all clusters.
The basic details for [math]K[/math]means clustering are given:
The [math]K[/math] initial centers are randomly chosen from the training data.
Then the following two steps are iterated alternately until convergence.
1. For each existing center, we reidentify its cluster (every point in this cluster should be closer to this center than to others).
2. Compute the mean for each cluster and make it as the new center for each cluster.
Kmeans Clustering algorithm
 For a given cluster assignment [math]\displaystyle C[/math], the total cluster variance is minimized with respect to [math]\displaystyle \lbrace m_1,m_2,...,m_k \rbrace[/math] yielding the means of the currently assigned clusters.
 Given a current set of means[math]\displaystyle \lbrace m_1,m_2,...,m_k \rbrace[/math], [math]\displaystyle min \sum_{k=1}^K \N_k \sum_{C(i)=k} \x_im_k \^2[/math] is minimized by assigning each observation to the closet (current) cluster mean. That is, [math]\C(i)=argmin \Vert x_im_k \Vert ^2[/math].
 Steps 1 and 2 are iterated until the assignments do not change.
Example:
Partition data into 2 clusters (2 hidden values)
>> X=rand(30,80); >> [IDX,C,sumD,D]=kmeans(X,2); >> size(IDX) >> 30 1 >> size(C) >> 2 80 >> size(sumD) >> 2 1 >> c1=sum(IDX==1) >> 14 >> c2=sum(IDX==2) >> 16 >> sumD >> 85.6643 >> 101.0419 >> v1=sumD(1,1)/c1 >> 6.1189 >> v2=sumD(2,1)/c2 >> 6.3151
Comments:
We create [math]X[/math] randomly as a training set with 80 data points and 30 dimensions, and then apply “kmeans” method to separate X into 2 clusters. IDX is a vector contains 1 or 2 which indicates 2 clusters, and its size is 30*1. [math]\displaystyle C [/math] is the center (mean) of each cluster with size 2*80; sumD is sum of the square distance between the data points and center of its cluster. The [math]\displaystyle c1 [/math] and [math]\displaystyle c2 [/math] indicate the number of data points in cluster 1 and 2. [math]\displaystyle V1 [/math] is the variance of the first cluster [math]\displaystyle (v1=\sigma_1)[/math]; [math]\displaystyle V2 [/math] is the variance of the second cluster [math]\displaystyle (v2=\sigma_2)[/math]. Now we can get [math]\displaystyle \phi [/math], [math]\displaystyle w [/math], hat matrix [math]\displaystyle H [/math] and [math]\displaystyle \hat Y [/math] by following equations. Finally, we will get the [math]\displaystyle MSE [/math] and predict the test set.
[math]\displaystyle \phi_{j}(x) = e^{\frac{\Vert x  \mu_{j}\Vert ^2}{2\sigma_{j}^2}}[/math]
[math]\displaystyle W = (\phi^{T}\Phi)^{1}\phi^{T}Y[/math]
[math]\displaystyle H = \phi(\Phi^{T}\Phi)^{1}\phi^{T}[/math]
[math]\displaystyle \hat{Y} = \phi W = \phi(\phi^{T}\phi)^{1}\phi^{T}Y = HY[/math]
Aside:
Similar in spirit to [math]K[/math]means, there is EM algorithm with respect to Gaussian mixture model. Generally speaking, the Gaussian mixture model is referred to as a soft clustering while [math]K[/math]means is hard clustering.
Similar to [math]K[/math]means, the following two steps are iterated alternately until convergernce.
Estep, each point is assigned a weight for each cluster based on the likelihood of each of the corresponding Gaussians. Observations is assigned 1 for one cluster if they are closer to the center of that cluster, and is assigned 0 for other clusters.
Mstep, compute the weighted means and covariances and make them as the new means and covariances for every cluster.
>>[P,mu,phi,1Pxtr]=mdgEM(X,2,200,0);
Support Vector Machine
Introduction
We have seen that linear discriminant analysis and logistic regression both estimate linear decision boundaries in similar but slightly different ways. Separating hyperplane classifiers provide the basis for the support vector machine (SVM). An SVM constructs linear decision boundaries that explicitly try to separate the data into different classes while maximizing the margin of separation. The techniques that make the extensions to the nonseparable case, where the classes overlap, are generalized to what is known as the support vector machine. It produces nonlinear boundaries by constructing a linear boundary in a highdimensional, transformed version of the feature space. It is also calculated based on only a fraction of the data rather than a calculation on every point in the data, much like the difference between the median and the mean.
The original basis for SVM was published in the 1960s by Vapnik, Chervonenkis et al., however the ideas did not gain any attention until strong results were shown in the early 1990s.
Definition:
Support Vector Machines (SVM) are a set of related supervised learning methods used for classification and regression. A support vector machine constructs a maximum margin hyperplane or set of hyperplanes in a higher or infinite dimensional space. The set of points near the class boundaries, support vectors, define the model which can be used for classification, regression or other tasks.
Optimal Seperating Hyperplane
Figure 28.2 An example with two classes separated by a hyperplane. The blue line is the least squares solution, which misclassifies one of the training points. Also shown are the black separating hyperplanes found by the perceptron learning algorithm with different random starts.
We can see the data points can be separated by a linear boundary are in two classes in [math]\displaystyle \mathbb{R}^{2} [/math]. Suppose a dataset is indeed linearly separable, then there exits infinitely many possible separating hyperplanes including the black lines in the figure as two of them for training data. However, which solution is the best when we introduce the new data.
Aside:
The blue line is the least squares solution to the problem,obtained by regressing the [math]\displaystyle 1/+1 [/math] response [math]\displaystyle Y [/math] on [math]\displaystyle X [/math] (with intercept); the line is given by
[math]\displaystyle {X:\hat\beta_0+\hat\beta_1X_1+\hat\beta_2X_2=0}[/math].
This least squares solution does not do a perfect job in separating the points, and makes one error. This is the same boundary found by linear discriminant analysis, in light of its equivalence with linear regression in the twoclass case.
Classifiers such as (28.4) that compute a linear combination of the input features and return the sign were called perceptrons in the engineering literature in the late 1950s.
Identifications:
 Hyperplane: separate two classes
[math]\displaystyle x^{T}\beta+\beta_0=0[/math]
 Margin: the distance between the hyperplane and the closest point.
[math]\displaystyle d_i=x_i^{T}\beta+\beta_0 [/math] where [math]\displaystyle i=1,....,N[/math]
Note: since distance is positive, if the data is on [math]\displaystyle +1 [/math] side the distance is [math]\displaystyle d_i(+1)[/math]. If the data is on the [math]\displaystyle 1 [/math] side the distance is [math]\displaystyle d_i(1)[/math].
 Data points: [math]\displaystyle y_i\in\{1,+1\}[/math]we can classify points as[math]\displaystyle sign\{d_i\}[/math] if [math]\displaystyle \beta,\beta_0 [/math] is known.
Maximum Margin Classifiers in the Linearly separable case
Choose the line farthest from both classes or choose the line which has the maximum distance from the closest point. (i.e maximize the margin)
[math]\displaystyle Margin=min\{y_id_i\}[/math] [math]\displaystyle i=1,2,....,N [/math]
where [math]\displaystyle y_i [/math] is label and [math]\displaystyle d_i [/math] is distance
Fiqure 28.3 depicts a hyperplane defined by the equation [math]\displaystyle x^{T}\beta+\beta_0=0[/math]. Since they are in [math]\displaystyle \mathbb{R}^{2} [/math], the hyperplane is a line.
Let us rewrite [math]\displaystyle Margin=min\{y_id_i\}[/math] by using the following properties:
1. [math]\displaystyle \beta [/math] is orthogonal to the hyperplane
Two points[math]\displaystyle x_1,x_2[/math] lying on the hyperplane.
[math]\displaystyle \beta^{T}x_1+\beta_0=0[/math]
[math]\displaystyle \beta^{T}x_2+\beta_0=0[/math]
[math]\displaystyle \beta^{T}x_1+\beta_0  (\beta^{T}x_2+\beta_0)=0[/math]
[math]\displaystyle \beta^{T}(x_1x_2)=0[/math]
Hence,[math]\displaystyle \beta [/math] is orthogonal to [math]\displaystyle (x_1x_2)[/math], and[math]\displaystyle \beta^*=\frac{\beta}{\\beta\} [/math] is the vector normal to the hyperplane.
2. For any point [math]\displaystyle x_1 [/math] on the hyperplane,
[math]\displaystyle \beta^{T}x_1+\beta_0=0[/math]
[math]\displaystyle \beta^{T}x_1=\beta_0[/math]
For any point on the hyperplane, multiplying by [math]\displaystyle \beta^{T}[/math] gives negative value of the intercept of the hyperplane.
3. The signed distance for any point [math]\displaystyle x [/math] to the hyperplane is [math]\displaystyle d_i=\beta^{T}(x_ix_0)[/math].
Since the length of [math]\displaystyle \beta [/math] is not known, let it be unit vector.
[math]\displaystyle d_i=\frac{\beta^{T}(x_ix_0)}{\\beta\} [/math] [math]\displaystyle i=1,2,....,N [/math]
[math]\displaystyle d_i=\frac{\beta^{T}x_i\beta^{T}x_0}{\\beta\} [/math]
by property 2
[math]\displaystyle d_i=\frac{\beta^{T}x_i+\beta_0}{\\beta\} [/math]
We had [math]\displaystyle Margin=min(y_id_i)[/math] [math]\displaystyle i=1,2,....,N [/math], and since we now know how to compute [math]\displaystyle d_i \Rightarrow[/math]
[math]\displaystyle Margin=min\{y_i\frac{\beta^{T}x_i+\beta_0}{\\beta\}\} [/math]
Suppose [math]\displaystyle x_i [/math] is not on the hyperplane
[math]\displaystyle y_i(\beta^{T}x_i+\beta_0)\gt 0 [/math]
[math]\displaystyle y_i(\beta^{T}x_i+\beta_0)\geq c [/math]for [math]\displaystyle c\gt 0 [/math]
[math]\displaystyle y_i(\frac{\beta^{T}x_i}{c}+\frac{\beta_0}{c})\geq1[/math]
This is known as the canonical representation of the decision hyperplane.
For [math]\displaystyle \beta^{T} [/math] only the direction is important, so [math]\displaystyle \frac{\beta^{T}}{c} [/math] does not change its direction, the hyperplane will be the same.
[math]\displaystyle y_i(\beta^{T}x_i+\beta_0)\geq1 [/math]
[math]\displaystyle y_i\frac{\beta^{T}x_i+\beta_0}{\\beta\}\geq\frac{1}{\\beta\} [/math]
[math]\displaystyle Margin=min(\frac{1}{\\beta\} )[/math]
which is equivalent as [math]\displaystyle Margin=max(\\beta\ )[/math]
Reference:
Hastie,T.,Tibshirani,R., Friedman,J.,(2008).The Elements of Statistical Learning:129130
ExtensionMulticlass SVM[28]
SVM is only directly applicable for twoclass case. We want to generalize this algorithm to multiclass tasks.
Multiclass SVM aims to assign labels to instances by using support vector machines, where the labels are drawn from a finite set of several elements. The dominating approach for doing so is to reduce the single multiclass problem into multiple binary problems. Each of the problems yields a binary classifier, which is assumed to produce an output function that gives relatively large values for examples from the positive class and relatively small values for examples belonging to the negative class. Two common methods to build such binary classifiers are where each classifier distinguishes between (i) one of the labels to the rest (oneversusall) or (ii) between every pair of classes (oneversusone). Classification of new instances for oneversusall case is done by a winnertakesall strategy, in which the classifier with the highest output function assigns the class (it is important that the output functions be calibrated to produce comparable scores). For the oneversusone approach, classification is done by a maxwins voting strategy, in which every classifier assigns the instance to one of the two classes, then the vote for the assigned class is increased by one vote, and finally the class with most votes determines the instance classification.
Optimizing The Support Vector Machine  November 16th, 2009
We currently derive Support Vector Machine for the case where two classes are separable in the given feature space. This margin can be written as [math]\,min\{y_id_i\}[/math], or the distance of each point from the hyperplane, where [math]\,d_i[/math] is the distance and [math]\,y_i[/math] is used as the sign.
Margin Maximizing Problem for the Support Vector Machine
[math]\,Margin=min\{y_id_i\}[/math] can be rewritten as [math]\,min\left\{\frac{y_i\left(\beta^Tx_i+\beta_0\right)}{\beta}\right\}[/math]. <br\>Note that the term [math]\,y_i\left(\beta^Tx_i+\beta_0\right) = 0[/math] if [math]\,x_i[/math] is on the hyperplane, but [math]\,y_i\left(\beta^Tx_i+\beta_0\right) \gt 0[/math] if [math]\,x_i[/math] is not on the hyperplane.
This implies [math]\,\exists C\gt 0[/math] such that [math]\,y_i\left(\beta^Tx_i+\beta_0\right) \geq C[/math].
Divide through by C to produce [math]\,y_i\left(\frac{\beta^T}{C}x_i + \frac{\beta_0}{C}\right) \geq 1[/math].
[math]\,\beta, \beta_0[/math] compose a hyperplane that can have different values but we care about the direction, dividng through by a constant does not change the direction of the hyperplane. Thus, by assuming scaled values for [math]\,\beta, \beta_0[/math] we eliminate C, so that [math]\,y_i\left(\beta^Tx_i+\beta_0\right) \geq 1[/math]. Implying that the lower bound on [math]\,y_i\left(\beta^Tx_i+\beta_0\right)[/math] is [math]\displaystyle 1[/math]
Now in order to maximize the margin, we simply need to find [math]\,min\left\{\frac{1}{\beta}\right\}[/math].
In other words, our optimization problem is now to find maximum [math]\,\beta[/math], under the constraint that [math]\,min_i\{y_i(\beta^Tx_i+\beta_0)\} = 1[/math].
Note that we're dealing with the norm of [math]\,\beta[/math]. There are many different choices of possible norms, in general pnorm. The 1norm of a vector is simply the sum of the absolute value of each element (also known as the taxicab or Manhattan distance), and is apparently more accurate, but also has a discontinuity in the derivative. 2norm or the Euclidean norm (the intuitive measure of the length of a vector), is easier to work with  that is [math]\,\\beta\_2 = (\beta^T\beta)^{1/2}[/math]. For convenience, we will maximize [math]\,\frac{1}{2}\\beta\^2 = \frac{1}{2}\beta^T\beta[/math] where the constant 1/2 has been added for simplification and that the maximizing the function is the same as maximizing the square root of that function.
This is an example of a quadratic programming problem and we will minimize a quadratic function subject to linear inequality constraints.
Writing Lagrangian Form of Support Vector Machine
The Lagrangian form is introduced to ensure that the optimization conditions are satisfied, as well as finding an optimal solution (the optimal saddle point of the Lagrangian for the classic quadratic optimization). The problem will be solved in dual space by introducing [math]\,\alpha_i[/math] as dual constraints, this is in contrast to solving the problem in primal space as function of the betas. A simple algorithm for iteratively solving the Lagrangian has been found to run well on very large data sets, making SVM more usable. Note that this algorithm is intended to solve Support Vector Machines with some tolerance for errors  not all points are necessarily classified correctly. Several papers by Mangasarian explore different algorithms for solving SVM.
[math]\,L(\beta,\beta_0,\alpha) = \frac{1}{2}\\beta\^2  \sum_{i=1}^n{\alpha_i\left(y_i(\beta^Tx_i+\beta_0)1\right)}[/math]. To find the optimal value, set the derivative equal to zero.
[math]\,\frac{\partial L}{\partial \beta} = 0[/math], [math]\,\frac{\partial L}{\partial \beta_0} = 0[/math]. Note that [math]\,\frac{\partial L}{\partial \alpha_i}[/math] is equivalent to the constraints [math]\left(y_i(\beta^Tx_i+\beta_0)1\right) \geq 0, \,\forall\, i[/math]
First, [math]\,\frac{\partial L}{\partial \beta} = \frac{\partial}{\partial \beta}\frac{1}{2}\\beta\^2  \sum_{i=1}^n{\left\{\frac{\partial}{\partial \beta}(\alpha_iy_i\beta^Tx_i)+\frac{\partial}{\partial \beta}\alpha_iy_i\beta_0\frac{\partial}{\partial \beta}\alpha_iy_i\right\}}[/math]
 [math]\frac{\partial}{\partial \beta}\frac{1}{2}\\beta\^2 = \beta[/math].
 [math]\,\frac{\partial}{\partial \beta}(\alpha_iy_i\beta^Tx_i) = \alpha_iy_ix_i[/math]
 [math]\,\frac{\partial}{\partial \beta}\alpha_iy_i\beta_0 = 0[/math].
 [math]\,\frac{\partial}{\partial \beta}\alpha_iy_i = 0[/math].
So this simplifies to [math]\,\frac{\partial L}{\partial \beta} = \beta  \sum_{i=1}^n{\alpha_iy_ix_i} = 0[/math]. In other words,
[math]\,\beta = \sum_{i=1}^n{\alpha_iy_ix_i}[/math], [math]\,\beta^T = \sum_{i=1}^n{\alpha_iy_ix_i^T}[/math]
Similarly, [math]\,\frac{\partial L}{\partial \beta_0} = \sum_{i=1}^n{\alpha_iy_i} = 0[/math].
This allows us to rewrite the Lagrangian without [math]\,\beta[/math].
[math]\,\frac{1}{2}\sum_{i=1}^n{\sum_{j=1}^n{\alpha_i\alpha_jy_iy_jx_i^Tx_j}}  \sum_{i=1}^n{\alpha_iy_i(\sum_{j=1}^n{\alpha_jy_jx_j^Tx_i} + \beta_0)  1}[/math].
Because [math]\,\sum_{i=1}^n{\alpha_iy_i} = 0[/math], and [math]\,\beta_0[/math] is constant, [math]\,\sum_{i=1}^n{\alpha_iy_i\beta_0} = 0[/math]. So this simplifies further, to
[math]L(\alpha) = \,\frac{1}{2}\sum_{i=1}^n{\sum_{j=1}^n{\alpha_i\alpha_jy_iy_jx_i^Tx_j}} + \sum_{i=1}^n{\alpha_i}[/math] A dual representation of the maximum margin.
Because [math]\,\alpha_i[/math] is the Lagrange multiplier, [math]\,\alpha_i \geq 0 \forall i[/math].
This is a much simpler optimization problem
Extension:Global Optimization of Support Vector Machines(Using Genetic Algorithms for Bankruptcy Prediction)
One of the most important research issues in finance is building accurate corporate bankruptcy prediction models since they are essential for the risk management of financial institutions. Thus, researchers have applied various datadriven approaches to enhance prediction performance including statistical and artificial intelligence techniques. Recently, support vector machines(SVMs) are becoming popular because they use a risk function consisting of the empirical error and a regularized term which is derived from the structural risk minimization principle. In addition, they don’t require huge training samples and have little possibility of overfitting. However, in order to use SVM, a user should determine several factors such as the parameters of a kernel function, appropriate feature subset, and proper instance subset by heuristics, which hinders accurate prediction results when using SVM.
There is a paper [[29]]proposing a novel approach to enhance the prediction performance of SVM for the prediction of financial distress. Their suggestion is the simultaneous optimization of the feature selection and the instance selection as well as the parameters of a kernel function for SVM by using genetic algorithms (GAs). They apply their model to a realworld case. Experimental results show that the prediction accuracy of conventional SVM may be improved significantly by using their model.
Extension: Finding Optimal Parameter Values
The accuracy of an SVM model dependents on the selection of the model parameters. DTREG provides two methods for finding optimal parameter values. Grid search: tries values of each parameter across the specified search range using geometric steps. Pattern search/compass search/line search:starts at the center of the search range and makes trial steps in each direction for each parameter.
If the fit of the model improves, the search center moves to the new point and the process is repeated. If no improvement is found, the step size is reduced and the search is tried again. The pattern search stops when the search step size is reduced to a specified tolerance.
Positives and Negatives When Optimizing SVM[30]
 (Pos) Appears to avoid overfitting in high dimensional spaces and generalize well using a small training set (the complexity of SVM is characterized by the number of support vectors rather than the dimensionality of the transformed space  no formal theory to justify this).
 (Pos) Global optimization method, no local optima (SVM are based on exact optimization, not approximate methods).
 (Neg) Applying trained classifiers can be expensive.
The Support Vector Machine algorithm  November 18, 2009
Lagrange Duality
In convex optimization, consider the primal optimization problem:
[math]\,\min_\beta[/math] [math]\,f(\omega)[/math]
[math]\,s.t.[/math] [math]\,g_i(\omega) \leq 0[/math]
[math]\,h_i(\omega) = 0[/math]
Define the generalized Lagrangian to be
[math]L(\omega,\alpha,\beta) = f(\omega)+\sum_{i=1}^k\alpha_ig_i(\omega)+\sum_{i=1}^k\beta_ih_i(\omega)[/math]
Then the dual optimization problem is
[math]\ \max{\alpha,\beta} [/math] [math]\ \min_{\omega}[/math] [math]\ L(\omega,\alpha,\beta) [/math]
Now instead of solving the primal problem, we can solve the dual problem without changing the solution as long as it subjects to the KarushKuhnTucker (KKT) conditions:
[math]\frac{\partial}{\partial\omega_i}L(\omega,\alpha,\beta)=0[/math]
[math]\frac{\partial}{\partial\beta_i}L(\omega,\alpha,\beta)=0[/math]
[math]\,\alpha_ig_i(\omega)=0[/math]
[math]g_i(\omega) \leq 0[/math]
[math]\,\alpha_i \geq 0[/math]
We are interested in the dual form because it gives bound on the primal problem and in some cases is easier to solve. For more information about convex optimization, see the book by Boyd: [31]
Solving the Lagrangian
Continuing from the above derivation, we now have the equation that we need to minimize, as well as two constraints.
The Support Vector Machine problem boils down to:
[math]\min_{\alpha} L(\alpha) = \sum_{i=1}^n{\alpha_i}  \frac{1}{2}\sum_{i=1}^n{\sum_{j=1}^n{\alpha_i\alpha_jy_iy_jx_i^Tx_j}}[/math]
 such that [math]\alpha_i \geq 0[/math]
 and [math]\sum_{i=1}^n{\alpha_i y_i} = 0[/math]
We are looking to minimize [math]\,\alpha[/math], which is our only unknown. Once we know [math]\,\alpha[/math], we can easily find [math]\,\beta[/math] and [math]\,\beta_0[/math] (see the Support Vector algorithm below for complete details).
If we examine the Lagrangian equation, we can see that [math]\,\alpha[/math] is multiplied by itself; that is, the Lagrangian is quadratic with respect to [math]\,\alpha[/math]. Our constraints are linear. This is therefore a problem that can be solved through quadratic programming techniques. We will examine how to do this in Matlab shortly.
We can write the Lagrangian equation in matrix form:
[math]L(\alpha) = \underline{\alpha}^T\underline{1}  \frac{1}{2}\underline{\alpha}^TS\underline{\alpha}[/math]
 such that [math]\underline{\alpha} \geq \underline{0}[/math]
 and [math]\underline{\alpha}^T\underline{y} = 0[/math]
Where:
 [math]\underline{\alpha}[/math] denotes an [math]\,n \times 1[/math] vector; [math]\underline{\alpha}^T = [\alpha_1, ..., \alpha_n][/math]
 Matrix [math]S = y_iy_jx_i^Tx_j = (y_ix_i)^T(y_ix_i)[/math]
 [math]\,\underline{0}[/math] and [math]\,\underline{1}[/math] are vectors containing all 0s or all 1s respectively
Using this matrix notation, we can use Matlab's built in quadratic programming routine, quadprog.
Quadprog example
Let's use quadprog
to find the solution to [math]\,L(\alpha)[/math].
Matlab's quadprog
function minimizes an equation of the following form:
 [math]\min_x\frac{1}{2}x^THx+f^Tx[/math]
 such that: [math]\,A \cdot x \leq b[/math], [math]\,A_{eq} \cdot x = b_{eq}[/math] and [math]\,lb \leq x \leq ub[/math]
We can now see why we kept the [math]\frac{1}{2}[/math] constant in the original derivation of the equation.
The function is called as such: x = quadprog(H,f,A,b,Aeq,beq,lb,ub)
. The variables correspond to values in the equation above.
We can now try to find the solution to [math]\,L(\alpha)[/math] (though, it should be noted, that in [math]\,L(\alpha)[/math] we subtract the first term rather than add it; I'm not sure if this difference would change how we call quadprog
).
We'll use a simple onedimensional data set, which is essentially y = 1 or 1 + Gaussian noise. (Note: you could easily put the values straight into the quadprog equation; they are separated for clarity.)
x = [mvnrnd([1],[0.01],100); mvnrnd([1],[0.01],100)]'; y = [ones(100,1); ones(100,1)]; S = (y*x)' * (y*x); f = ones(200,1); A = ones(1,200); b = 0; Aeq = y'; beq = 0; lb = 0; ub = []; % There is no upper bound alpha = quadprog(S,f,A,b,Aeq,beq,lb,ub);
This gives us the optimal [math]\,\alpha[/math]... or at least I think it should, but it does not appear to work for me (that is, despite setting the lower boundary of 0, a number of values are still negative. Whether this is just the nature of quadprog
or an error on my part is an exercise for the reader).
Examining K.K.T. conditions
KarushKuhnTucker conditions (more info) give us a closer look into the Lagrangian equation and the associated conditions.
Suppose we are looking to minimize [math]\,f(x)[/math] such that [math]\,g_i(x) \geq 0, \forall{x}[/math]. If [math]\,f[/math] and [math]\,g[/math] are differentiable, then the necessary conditions for [math]\hat{x}[/math] to be a local minimum are:
 At the optimal point, [math]\frac{\partial L}{\partial \hat{x}} = 0[/math]; i.e. [math]f'(\hat{x})  \sum{\alpha_ig'(\hat{x})}=0[/math]
 [math]\alpha_i \geq 0[/math]. (Dual Feasibility)
 [math]\alpha_ig_i(\hat{x}) = 0, \forall{i}[/math] (Complementary Slackness)
 [math]g_i(\hat{x}) \geq 0[/math] (Primal Feasibility)
If any of these conditions are violated, then the problem is deemed not feasible.
These are all trivial except for condition 3. Let's examine it further in our support vector machine problem.
Support Vectors
Support vectors are the training points that determine the optimal separating hyperplane that we seek. Also, they are the most difficult points to classify or the most informative for the classification.
In our case, the [math]g_i(\hat{x})[/math] function is:
 [math]\,g_i(x) = y_i(\beta^Tx_i+\beta_0)1[/math]
Substituting [math]\,g_i[/math] into KKT condition 3, we get [math]\,\alpha_i[y_i(\beta^Tx_i+\beta_0)1] = 0[/math]. <br\>In order for this condition to be satisfied either
[math]\,\alpha_i= 0[/math] or
[math]\,y_i(\beta^Tx_i+\beta_0)=1[/math]
All points [math]\,x_i[/math] will be either 1 or greater than 1 distance away from the hyperplane.
Case 1: a point [math]\displaystyle x_i \gt 1[/math] away from the margin
If [math]\,y_i(\beta^Tx_i+\beta_0) \gt 1 \Rightarrow \alpha_i = 0[/math].
If point [math]\, x_i[/math] is not on the margin, then the corresponding [math]\,\alpha_i=0[/math].
Case 2: a point [math]\displaystyle x_i = 1[/math] away from the margin
If [math]\,\alpha_i \gt 0 \Rightarrow y_i(\beta^Tx_i+\beta_0) = 1[/math] <br\>If point [math]\, x_i[/math] is on the margin, then the corresponding [math]\,\alpha_i\gt 0[/math].
Points on the margin, with corresponding [math]\,\alpha_i \gt 0[/math], are called support vectors.
Using support vectors
Support vectors are important because they allow the support vector machine algorithm to be insensitive to outliers. If [math]\,\alpha = 0[/math], then the cost function is also 0, and won't contribute to the solution of the SVM problem; only points on the margin — support vectors — contribute. Hence the model given by SVM in entirely defined by the set of support vectors, a subset of the entire training set. This is interesting because in the NN methods(and can be generalize to classical statistical learning) previous to this the configuration of the network needed to be specified. In this case we have a datadriven or 'nonparametric' model in which is the training set and algorithm will determine the support vectors, instead of fitting a set of parameters using CV or other error minimization functions.
References: Wang, L, 2005. Support Vector Machines: Theory and Applications, Springer, 3
The support vector machine algorithm
 Solve the quadratic programming problem:[math]\min_{\alpha} L(\alpha) = \sum_{i=1}^n{\alpha_i}  \frac{1}{2}\sum_{i=1}^n{\sum_{j=1}^n{\alpha_i\alpha_jy_iy_jx_i^Tx_j}}[/math] such that [math]\alpha_i \geq 0[/math] and [math]\sum_{i=1}^n{\alpha_i y_i} = 0[/math]
 Use Matlab's quadprog to find the optimal [math]\,\underline{\alpha}[/math]
 Find [math]\beta = \sum_{i=1}^n{\alpha_iy_i\underline{x_i}}[/math]
 Find [math]\,\beta_0[/math] by choosing a support vector (a point with [math]\,\alpha_i \gt 0[/math]) and solving [math]\,y_i(\beta^Tx_i+\beta_0) = 1[/math]
Example in Matlab
The following code, taken verbatim from the lecture, shows how to use Matlab builtin SVM routines (found in the Bioinformatics toolkit) to do classification through support vector machines.
load 2_3; [U,Y] = princomp(X'); data = Y(:,1:2); l = [ones(1,200) ones(1,200)]; [train,test] = crossvalind('holdOut',400); % Gives indices of train and test; so, train is a matrix of 0 or 1, 1 where the point should be used as part of the training set svmStruct = svmtrain(data(train,:), l(train), 'showPlot', true);
yh = svmclassify(svmStruct, data(test,:), 'showPlot', true);
SVM in Gene Selection
DNA microarrays now permit scientists to screen thousands of genes simultaneously and determine whether those genes are active, hyperactive or silent in normal or cancerous tissue. Because these new microarray devices generate bewildering amounts of raw data, new analytical methods must be developed to sort out whether cancer tissues have distinctive signatures of gene expression over normal tissues or other types of cancer tissues.
Extention:Support Vector Machines for Pattern Recognition
[32] This paper talks about linear Support Vector Machines for separable and nonseparable data by working through a nontrivial example in detail, and also it describes a mechanical analog and when SVM solutions are unique and when they are global. From this paper we can know support vector training can be practically implemented, and the kernel mapping technique which is used to construct SVM solutions which are nonlinear in the data.
Results of some experiments which were inspired by these arguments are also presented. The writer gives numerous examples and proofs of most of the key theorems, he hopes the people can find old material is cast in a fresh light since the paper includes some new material.
Limitation of SVM algorithm [33]
 The biggest limitation of SVM lies in the choice of the kernel (the best choice of kernel for a given problem is still a research problem).
 A second limitation is speed and size (mostly in training  for large training sets, it typically selects a small number of support vectors, thereby minimizing the computational requirements during testing).
Nonlinear hypersurfaces and NonSeparable classes  November 20, 2009
Kernel Trick
We talked about the curse of dimensionality at the beginning of this course, however, we now turn to the power of high dimensions in order to find a linearly separable hyperplane between two classes of data points. To understand this, imagine a two dimensional prison where a two dimensional person is constrained. Suppose magically we give the person a third dimension, then he can escape from the prison. In other words, the prison and the person are linearly separable now with respect to the third dimension. The intuition behind the "kernel trick" is basically to map data to a higher dimension so that they are linearly separable by a hyperplane.
The original optimal hyperplane algorithm proposed by Vladimir Vapnik in 1963 was a linear classifier. However, in 1992, Bernhard Boser, Isabelle Guyon and Vapnik suggested a way to create nonlinear classifiers by applying the kernel trick to maximummargin hyperplanes. The algorithm is very similar, except that every dot product is replaced by a nonlinear kernel function as below. This allows the algorithm to fit the maximummargin hyperplane in a transformed feature space. We have seen SVM as a linear classification problem that finds the maximum margin hyperplane in the given input space. However, for many real world problems a more complex decision boundary is required. The following simple method was devised in order to solve the same linear classification problem but in a higher dimensional space, a 'feature space', under which the maximum margin hyperplane is better suited.
Let [math]\,\phi[/math] be a mapping,
[math]\phi:\mathbb{R}^d \rightarrow \mathbb{R}^D [/math]
We wish to find a [math]\,\phi[/math] such that our data will be suited for separation by a hyperplane. Given this function, we are lead to solve the previous constrained quadratic optimization on the transformed dataset,
[math]\min_{\alpha} L(\alpha) = \sum_{i=1}^n{\alpha_i}  \frac{1}{2}\sum_{i=1}^n{\sum_{j=1}^n{\alpha_i\alpha_jy_iy_j\phi(x_i)^T\phi(x_j)}}[/math] such that [math]\alpha_i \geq 0[/math] and [math]\sum_{i=1}^n{\alpha_i y_i} = 0[/math]
The solution to this optimization problem is now well known; however a workable [math]\,\phi[/math] must be determined. Possibly the largest drawback in this method is that we must compute the inner product of two vectors in the high dimensional space. As the number of dimensions in the initial data set increases, the inner product becomes computationally intensive or impossible.
However, we have a very useful result that says that there exists a class of functions, [math]\,\Phi[/math], which satisfy the above requirements and that for any function [math]\,\phi \in \Phi[/math],
[math]\,\phi(x_i)^T\phi(x_j) = K(x_i,x_j) [/math]
Where K is the kernel function in the input space satisfying Mercer's condition (to guarantee that it indeed corresponds to certain mapping function [math]\,\phi[/math]). As a result, if the objective function depends on inner products but not on coordinates, we can always use the kernel function to implicitly calculate in the feature space without storing the huge data. Not only does this solve the computation problems but it no longer requires us to explicitly determine a specific mapping function in order to use this method. In fact, it is now possible to use an infinite dimensional feature space in SVM without even knowing the function [math]\,\phi[/math].
Three popular kernel choices in the SVM
The SVM only relies on the innerproduct between vectors [math]\ x_ix_j [/math] If every data point is mapped into highdimensional space via some transformation [math]\ \phi [/math], the innerproduct becomes: [math]\ K(x_i,x_j)= \phi(x_i)\phi (x_j)[/math] [math]\ K(x_i,x_j) [/math] is called the kernel function. For SVM, we only need specify the kernel [math]\ K(x_i,x_j) [/math], without need to know the corresponding nonlinear mapping, [math]\ \phi (x)[/math].
There are many types of kernels that can be used in Support Vector Machines models. These include linear, polynomial, radial basis function (RBF) and sigmoid functions.
Linear: [math]\ K(\underline{x}_{i},\underline{x}_{j})= \underline{x}_{i}^T\underline{x}_{j}[/math],
Polynomial: [math]\ K(\underline{x}_{i},\underline{x}_{j})= (\gamma\underline{x}_{i}^T\underline{x}_{j}+r)^{d}, \gamma \gt 0[/math],
Radial Basis: [math]\ K(\underline{x}_{i},\underline{x}_{j})= exp(\gamma \\underline{x}_i  \underline{x}_j\^{2}), \gamma \gt 0[/math],
Gaussian kernel: [math]\ K(x_i,x_j)=exp(\frac{x_ix_j^2}{2\sigma^2 })[/math],
Twolayer perceptron: [math]\ K(x_i,x_j)=tanh(\alpha x_ix_j+\beta)[/math],
Sigmoid: [math]\ K(\underline{x}_{i},\underline{x}_{j})= tanh(\gamma\underline{x}_{i}^T\underline{x}_{j}+r)[/math].
Here, [math]\ \gamma [/math], [math]\ r[/math], and [math]\ d[/math] are all kernel parameters.
The RBF is by far the most popular choice of kernel types used in Support Vector Machines. This is mainly because of their localized and finite responses across the entire range of the real xaxis.The art of flexible modeling using basis expansions consists of picking an appropriate family of basis functions, and then controlling the complexity of the representation by selection, regularization, or both. Some of the families of basis functions have elements that are defined locally; for example, [math]\displaystyle B[/math]splines are defined locally in [math]\displaystyle R[/math]. If more flexibility is desired in a particular region, then that region needs to be represented by more basis functions(which in the case of [math]\displaystyle B[/math]splines translates to more knots).Kernel methods achieve flexibility by fitting simple models in a region local to the target point [math]\displaystyle x_0[/math]. Localization is achieved via a weighting kernl [math]\displaystyle K[/math] and individual observations receive weights [math]\displaystyle K(x_0,x_i)[/math]. RBF combine these ideas, by treating the kernel functions as basis functions.
Once we have chosen the Kernel function, we don't need to figure out what [math]\,\phi[/math] is, just use [math]\,\phi(\underline{x}_i)^T\phi(\underline{x}_j) = K(\underline{x}_i,\underline{x}_j) [/math] to replace [math]\,\underline{x}_i^T\underline{x}_j[/math]
Since the transformation chosen is dependent on the shape of the data, the only automated way to choose an appropriate kernel is by trial and error. Otherwise it is chosen manually.
Mercer's Theorem in detail
Let [math]\,\phi[/math] be a mapping to a high dimensional Hibert Space [math]\,H[/math]
[math]\phi:x \in \mathbb{R}^d \rightarrow H [/math]
The transformed coordinates can be defined as,
[math]\phi_1(x)\dots\phi_d(x)\dots [/math]
By Hilbert  Schmidt theory we can represent an inner product in Hilbert space as,
[math]\,\phi(x_i)^T\phi(x_j) = \sum_{r=1}^{\infty}a_k\phi_r(x_i)\phi(x_j) \Leftrightarrow K(x_i,x_j), \ a_r \ge 0 [/math]
where K is symmetric, then Mercer's theorem gives necessary and sufficient conditions on K for it to satisfy the above relation.
Let C be a compact subset of [math]\mathbb{R}^d[/math] and K a function [math] \in L^2(C) [/math], if
[math]\, \int_C\int_C K(u,v)g(u)g(v)dudv \ge 0, \ \forall g \in L^2(C)[/math]
then,
[math]\sum_{r=1}^{\infty}a_k\phi_r(u)\phi(v)[/math] converges absolutely and uniformly to a symmetric function [math]\,K(u,v)[/math]
References:
Vapnik, V., 1998. Statistical Learning Theory. John Wiley & Sons, {423}
Mercer, J., 1909. Functions of positive and negative type and their connection with the theory of integral equations. Philos. Trans. Roy. Soc. London, A 209:415{446}
Kernel Functions
There are various kernel functions, for example:
 Linear kernel: [math]\,k(x,y)=x \cdot y[/math]
 Polynomial kernel: [math]\,k(x,y)=(x \cdot y)^d[/math]
 Gaussian kernel: [math]e^{\frac{xy^2}{2\sigma^2}}[/math]
If [math]\,X[/math] is a [math]\,d \times n[/math] matrix in the original space, and [math]\,\phi(X)[/math] is a [math]\,D \times n[/math] matrix in the Hilbert space (good explanation video: part 1 part 2), then [math]\,\phi^T(X) \cdot \phi(X)[/math] is an [math]\,n \times n[/math] matrix. The inner product is also illustrated as correlation, which measures the similarity between data points. This gives us some insight in how to choose the kernel. The choice depends on certain prior knowledge of the problem and on how we believe the similarity of our data should be measured. In practice, the Gaussian (RBF) kernel usually works best. Besides the most common kernel functions mentioned above, many novel kernels are also suggested for different problem domains like text classification, gene classification and so on.
These kernel functions can be applied to many algorithms to derive the "kernel version". For example, kernel PCA, kernel LDA, etc.
SVM: nonseparable case
We have seen how SVMs are able to find an optimally separating hyperplane of two separable classes of data, in which case the margin contains no data points. However, in the real world, data of different classes are usually mixed together at the boundary and it's hard to find a perfect boundary to totally separate them. To address this problem, we slacken the classification rule to allow data cross the margin. Mathematically the problem becomes,
 [math]\min_{\beta, \beta_0} \frac{1}{2}\beta^2[/math]
 [math]\,y_i(\beta^Tx_i+\beta_0) \geq 1\xi_i[/math]
 [math]\xi_i \geq 0[/math]
Now each data point can have some error [math]\,\xi_i[/math]. However, we only want data to cross the boundary when they have to and make the minimum sacrifice; thus, a penalty term is added correspondingly in the objective function to constrain the number of points that cross the margin. The optimization problem now becomes:
 [math]\min_{\alpha} \frac{1}{2}\beta^2+\gamma\sum_{i=1}^n{\xi_i}[/math]
 [math]\,s.t.[/math] [math]y_i(\beta^Tx+\beta_0) \geq 1\xi_i[/math]
 [math]\xi_i \geq 0[/math]
<br\>Note that [math]\,\xi_i[/math] is not necessarily smaller than one, which means data can not only enter the margin but can also cross the separating hyperplane.
<br\>Note that [math]\,\gamma \Rightarrow \infty [/math] is feasible in the separable case, as all [math]\,\xi_i = 0[/math]. In general, for higher [math]\,\gamma[/math], the sets are more separable.
Aside: More information about SVM and Kernal
SVMs are currently among the best performers for many benchmark datasets and have been extended to a number of tasks such as regression. It seems the kernel trick is the most attracting site of SVMs. This idea has now been applied to many other learning models where the innerproduct is concerned, and they are called ‘kernel’ methods. Tuning SVMs remains to be the main research focus: how to an optimal kernel? Kernel should match the smooth structure of the data.
Support Vector Machine algorithm for nonseparable cases  November 23, 2009
With the program formulation above, we can form the lagrangian, apply KKT conditions, and come up with a new function to optimize. As we will see, the equation that we will attempt to optimize in the SVM algorithm for nonseparable data sets is the same as the optimization for the separable case, with slightly different conditions.
Forming the Lagrangian
In this case we have have two constraints in the Lagrangian and therefore we optimize with respect to two dual variables [math]\,\alpha[/math] and [math]\,\lambda[/math],
 [math]L: \frac{1}{2} \beta^2 + \gamma \sum_{i=1}^n \xi_i  \sum_{i=1}^n \alpha_i[y_i(\beta^T x_i+\beta_0)1+\xi_i]\sum_{i=1}^n \lambda_i \xi_i[/math]
 [math]\alpha_i \geq 0, \lambda_i \geq 0[/math]
Applying KKT conditions[34]
 [math]\frac{\partial L}{\partial p} = 0[/math] at an optimal solution [math]\, \hat p[/math], for each primal variable [math]\,p = \{\beta, \beta_0, \xi\}[/math]
[math]\frac{\partial L}{\partial \beta}=\beta  \sum_{i=1}^n \alpha_i y_i x_i = 0 \Rightarrow \beta=\sum_{i=1}^n\alpha_i y_i x_i[/math] <br\>[math]\frac{\partial L}{\partial \beta_0}=\sum_{i=1}^n \alpha_i y_i =0 \Rightarrow \sum_{i=1}^n \alpha_i y_i =0[/math] since the sign does not make a difference
[math]\frac{\partial L}{\partial \xi_i}=\gamma  \alpha_i  \lambda_i \Rightarrow \gamma = \alpha_i+\lambda_i[/math]. This is the only new condition added here  [math]\,\alpha_i \geq 0, \lambda_i \geq 0[/math], dual feasibility
 [math]\alpha_i g_i(\hat p) = 0[/math], complementary slackness: [math]\,\alpha_i[y_i(\beta^T x_i+\beta_0)1+\xi_i]=0[/math] and [math]\,\alpha_i \xi_i=0[/math]
 [math]\, g_i(\hat p) \geq 0[/math], primal feasibility. [math]\,y_i( \beta^T x_i+ \beta_0)1+ \xi_i \geq 0[/math]
Putting it all together
With our KKT conditions and the Lagrangian equation, we can now use quadratic programming to find [math]\,\alpha[/math]. <br\> Similar to what we did for the separable case after apply KKT conditions, replace the primal variables in terms of dual variables into the Lagrangian equations and simplify.
In matrix form, we want to solve the following optimization:
 [math]L(\alpha) = \underline{\alpha}^T\underline{1}  \frac{1}{2}\underline{\alpha}^TS\underline{\alpha}[/math]
 [math]\,s.t.[/math] [math]\underline{0} \leq \underline{\alpha} \leq \gamma[/math], [math]\underline{\alpha}^T\underline{y} = 0[/math]
Solving this gives us [math]\,\underline{\alpha}[/math], which we can use to find [math]\,\underline{\beta}[/math] as before:
 [math]\,\underline{\beta} = \sum{\alpha_i y_i \underline{x_i}}[/math]
However, we cannot find [math]\,\beta_0[/math] in the same way as before, even if we choose a point with [math]\,\alpha_i \gt 0[/math], because we do not know the value of [math]\,\xi_i[/math] in the equation
 [math]\,y_i(\underline{\beta}^Tx_i + \beta_0)  1 + \xi_i = 0[/math]
From our discussion on the KKT conditions, we know that [math]\,\lambda_i \xi_i = 0[/math] and [math]\,\gamma = \alpha_i + \lambda_i[/math].
So, if [math]\,\alpha_i \lt \gamma[/math] then [math]\,\lambda_i \gt 0[/math] and consequently [math]\,\xi_i = 0[/math].
Therefore, we can solve for [math]\,\beta_0[/math] if we choose a point where:
 [math]\,0 \lt \alpha_i \lt \gamma[/math]
Note
 When [math]\,0 \lt \alpha_i \lt \gamma[/math], we are considering a point that is on the margin.
 If [math]\,0 \lt \alpha_i = \gamma[/math] then [math]\,\lambda_i = 0 \Rightarrow \xi_i \gt 0[/math] and we're dealing with a point that has crossed the margin.
 In this case, the local minimization is also the global minimization. Since [math]\,S[/math] is positive semidefinite, then [math]\,L(\alpha)[/math] is convex.
The SVM algorithm for nonseparable data sets
The algorithm, then, for nonseparable data sets is:
 Use
quadprog
(or another quadratic programming technique) to solve the above optimization and find [math]\,\alpha[/math]  Find [math]\,\underline{\beta}[/math] by solving [math]\,\underline{\beta} = \sum{\alpha_i y_i x_i}[/math]
 Find [math]\,\beta_0[/math] by choosing a point where [math]\,0 \lt \alpha_i \lt \gamma[/math] and then solving [math]\,y_i(\underline{\beta}^Tx_i + \beta_0)  1 = 0[/math]
Potential drawbacks
Potential drawbacks of the SVM are the following two aspects:
1.Uncalibrated Class membership probabilities[35]
2.The SVM is only directly applicable for twoclass tasks. Therefore, algorithms that reduce the multiclass task to several binary problems have to be applied, see the Multiclass SVM section.
3.How to select the kernel function parameters  for Gaussian kernels the width parameter [math]\,\sigma [/math], and the value of [math]\,\epsilon [/math] in the [math]\,\epsilon [/math]insensitive loss function have not entirely solved yet.
some resouces:
1. introduction of SVM[36]
2. SVM in computational biology[37]
Finishing up SVM  November 25, 2009
Does SVM find a global minimum?
When we discussed KKT conditions, we listed the necessary conditions for [math]\hat{x}[/math] to be a local minimum. However, it would be ideal if we could show that SVM finds a global minimum (unlike, say, neural networks that find a local minimum).
Recall that our conditions, for the nonseparable case, are [math]\,0 \leq \underline{\alpha} \leq \gamma[/math] and [math]\,\underline{\alpha}^T\underline{y} = 0[/math]. These are both convex.
Our lagrangian equation is [math]L(\alpha) = \underline{\alpha}^T\underline{1}  \frac{1}{2}\underline{\alpha}^TS\underline{\alpha}[/math]. Since this is quadratic, it might be convex, but it also may not be; it depends on the matrix [math]\,S[/math]. If [math]\,S[/math] is a positive semidefinite matrix, then the lagrangian equation is convex.
Recall that [math]\,S[/math] is the product of [math]\,L^TL[/math], where [math]L_{d\times n} = \begin{bmatrix} y_1x_{11}& & y_nx_{1n} \\ \vdots&\cdots& \vdots\\ y_1x_{d1}& & y_nx_{dn} \\ \end{bmatrix}[/math]. Similar to the notion that squaring any number will give us a positive number in the end, a matrix that is the product of a matrix transposed times itself will result in a positive semidefinite matrix.
So, we know that [math]\,S[/math] is positive semidefinite. The lagrangian equation is convex, and therefore, the SVM algorithm finds a global minimum.
Naive Bayes, Decision Trees, K Nearest Neighbours, Boosting, and Bagging  November 25, 2009
Now that we've covered a number of more advanced classification algorithms, we can look at some of the simpler classification algorithms that are usually discussed at the beginning of a discussion on classification.
Naive Bayes Classifiers
Recall that one of the major drawbacks of the Bayes classifier was the difficulty in estimating a joint density in a multidimensional space. Naive Bayes classifiers are one possible solution to the problem. They are especially popular for problems with highdimension feature problems.
A naive Bayes classifier applies a strong independence assumption to the class density [math]\,f_{k}(x)[/math]. It assumes that inputs within each class are conditionally independent. In other words, it assumes that the value of one feature in a class is unrelated to that of any other feature.
[math]\ f_{k}(x) = \prod_{j=1}^d f_{jk}(x_{j})[/math]
Each of the d marginal densities can be estimated separately using onedimensional density estimates. If one of the components [math]\,x_{j}[/math] is discrete then its density can be estimated using a histogram. We can thus mix discrete and continuous variables in a naive Bayes classifier.
Naive Bayes classifiers often perform extremely well in practice despite these 'naive' and seemingly optimistic assumptions. This is because while individual class density estimates could be biased, the bias does not carry through to the posterior probabilities.
It is also possible to train naive Bayes classifiers using maximum likelihood estimation.
Decision Trees
Decision trees[38] are highly intuitive learning methods that can be thought of as partitioning the feature space into a number of rectangles. Trees can be used for classification, regression, or both. Trees map features of a decision problem onto a conclusion, or label.
We fit a tree model by minimizing some measure of impurity. For a single covariate [math]\,X_{1}[/math] we choose a point t on the real line that splits the real line into two sets R1 = [math](\infty,t][/math], R2 = [math][t,\infty)[/math] in a way that minimizes impurity.
We denote by [math] \hat p_{s}(j) [/math] the proportion of observations in [math]\ R_{s}[/math] that [math]\ Y_{i} = j[/math].
[math] \hat p_{s}(j) = \frac{\sum_{i = 1}^{n} I(Y_{i} = j,X_{i} \in R_{s})}{\sum_{i = 1}^{n} I(X_{i} \in R_{s})}[/math]
Extension: Decision Tree Analysis Decision Trees from Mind Tools
useful link:
Algorithm, Overfitting, Examples:[39],[40],[41]
Common Node Impurity Measures
Some common node impurity measures are:
 Misclassification error:
[math] 1  \hat p_{s}(j) [/math]
 Gini Index:
[math] \sum_{j \neq i} \hat p_{s}(j)\hat p_{s}(i)[/math]
 Crossentropy:
[math]  \sum_{j = 1}^{K} \hat p_{s}(j) log(\hat p_{s}(j))[/math]
KNearest Neighbours Classification
[math]K[/math]nearest neighbours is a very simple algorithm that classifies points based on a majority vote of the [math]\ k[/math] nearest points in the feature space, with the object being assigned to the class most common among its [math]\ k[/math] nearest neighbors. [math]\ k[/math] is a positive integer, typically small which is chosen by cross validation. If [math]\ k=1[/math], then the object is simply assigned to the class of its nearest neighbor.
1. Ties are broken at random.
2. If we assume the features are real, we can use the Euclidean distance in feature space.
3. Since the features are measured in different units, we can standardize the features to have mean zero and variance 1.
Property[42]
Kmearest neighbor algorithm has some strong results. As the number of data points goes infinity, the algorithm is guaranteed to yield an error rate no worse than twice the Bayes error rate (the minimum achievable error rate given the distribution of the data). Knearest neighbor is guaranteed to approach the Bayes error rate, for some value of k (where k increases as a function of the number of data points).
Boosting
Boosting algorithms are a class of machine learning metaalgorithms that can improve weak classifiers. If we have a weak classifier which slightly does better than random classification, then by assigning larger weights to points which are misclassified and trying to minimize the new cost function, we probably can get a new classifier which classifies with less error. This procedure can be repeated for a finite number of times and then a new classifier which is a weighed aggregation of the generated classifiers will be used as the boosted classifier. The better each generated classifier is the more its weight is in the final classifier.
Paper about Booting: Boosting is a general method for improving the accuracy of any given learning algorithm. This paper introduces the boosting algorithm AdaBoost, and explains the underlying theory of boosting, including an explanation of why boosting often does not suffer from overfitting as well as boosting’s relationship to supportvector machines. Finally, this paper gives some examples of applications of boosting recently.
AdaBoost Algorithm
Let's first look at the original boosting algorithm:
 Set all the weights of all points equal [math]w_i\leftarrow \frac{1}{n}[/math] where we have [math]\,n[/math] points.
 For [math]j=1,\dots, J[/math]
 Find [math]h_j:X\rightarrow \{1,+1\}[/math] that minimizes the weighted error [math]\,L_j[/math]
[math]h_j=\mbox{argmin}_{h_j} L_j [/math] where [math]L_j=\frac{\sum_{i=1}^n w_i I[y_i\neq h_j(x_i)]}{\sum_{i=1}^n w_i}[/math]  Let [math]\alpha_j\leftarrow\log(\frac{1L_j}{L_j})[/math]
 Update the weights: [math]w_i\leftarrow w_i e^{a_j I[y_j\neq h_j(x_i)]}[/math]
 Find [math]h_j:X\rightarrow \{1,+1\}[/math] that minimizes the weighted error [math]\,L_j[/math]
 The final classifier is [math]h(x)=\mbox{sign}\left(\sum_{j=1}^J \alpha_j h_j(x)\right)[/math]
When applying boosting to different classifiers, the first step in 2 may be different since we can define the most proper misclassification error according to the problem. However, the major idea is to give higher weight to misclassified examples, which does not change across classifiers.
Boosting works very well in practice, and there are a lot of research and published works on why it works this well. One possible explanation is that it actually maximizes the margin of classifiers.
We can see that in AdaBoost if training points are accurately classified, then their weights of being used in the next classifier is kept unchanged, while if points are not accurately classified, their weights of being used again is raised. At a result easier examples get classified in the very first few classifiers and hard examples are learned later with increasing emphasis. Finally, all the classifiers are combined through a majority vote, which is also weighted by their accuracy, taking consideration of both the easy and hard points. Thus the AdaBoost focuses on the more informative or difficult points.
AnyBoost
Many boosting algorithms belong to a class called AnyBoost which are gradient descent algorithms for choosing linear combinations of elements of an inner product space in order to minimize some cost functional.
We are primarily interested in voted combinations of classifiers [math]H(x) = sgn(\sum_{j=1}^J \alpha_j h_j(x))[/math]
We want to find H such that the cost functional [math]C(F) = \frac{1}{m}\sum_{i=1}^m c(y_i F(x_i))[/math] is minimized for a suitable cost function [math]c[/math]
[math]h_j:X\rightarrow \{1,+1\}[/math] are weak base classifiers from some class [math]\ H[/math] and [math] \alpha_j[/math] are classifier weights. The margin of an example [math](x_i,y_i)[/math] is defined by [math]y_i H(x_i)[/math].
The base hypotheses h and their linear combinations H can be considered to be elements of an inner product function space [math](S,\langle,\rangle)[/math].
We define the inner product as [math]\langle F,G \rangle = \frac{1}{m}\sum_{i=1}^m F(x_i) G(x_i)[/math] but the AnyBoost algorithm is valid for any cost function and inner product. We have a function [math]H[/math] as a linear combination of base classifiers and wish to add a base classifier h to H so that cost [math]\ C(H + \epsilon h)[/math] decreases for arbitrarily small [math] \epsilon[/math]. The direction we seek is found by maximizing [math]\langle\nabla C(H),h\rangle[/math]
AnyBoost algorithm:
 [math]\ H_0(x) = 0[/math]
 For [math]j=0,\dots, J[/math]
 Find [math]h_{j+1}:X\rightarrow \{1,+1\}[/math] that maximizes the inner product [math]\langle\nabla C(H),h_{j+1}\rangle[/math]
 If [math]\langle\nabla C(H),h_{j+1}\rangle \leq 0 [/math] then
 Return [math]\ H_j[/math]
 Choose step size [math]\ \alpha_{j+1}[/math]
 [math]\ H_{j+1} = H_j + \alpha_{j+1} h_{j+1}[/math]
 The final classifier is [math]\ H_{J+1}[/math]
Other voting methods, including AdaBoost, can be viewed as special cases of this algorithm.
Bagging
Bagging, or bootstrap aggregating, is another metatechnique used to reduce the variance of classifiers with high variability. It exploits the fact that a bootstrap mean is approximately equal to the posterior average. It is most effective for highly nonlinear classifiers such as decision trees. In particular because of the highly unstable nature of these classifiers, they stand most likely to benefit from bagging.
The idea is to train classifiers [math]\ h_{1}(x)[/math] to [math]\ h_{B}(x)[/math] using B bootstrap samples from the data set. The final classification is obtained using an average or 'plurality vote' of the B classifiers as follows:
 [math]\, h(x)= \left\{\begin{matrix} 1 & \frac{1}{B} \sum_{i=1}^{B} h_{b}(x) \geq \frac{1}{2} \\ 0 & \mathrm{otherwise} \end{matrix}\right.[/math]
Many classifiers, such as trees, already have underlying functions that estimate the class probabilities at [math]\,x[/math]. An alternative strategy is to average these class probabilities instead of the final classifiers. This approach can produce bagged estimates with lower variance and usually better performance.
References: Breiman L, Bagging Predictors, Machine Learning, 24, 123–140 (1996)
Example
Random Forests
A classifier consisting of a collection of tree–structured classifiers [math]\displaystyle {h(x;\theta_k)k=1,2,…},[/math] where [math]{\theta_k }[/math] are independently and identically distributed random vectors. The nature and dimensionality of [math]\theta[/math] depends on its use in the tree construction. Bagging is such an example. Currently, random forest use randomly selected inputs or combinations of inputs at each node to grow the tree.
How to compare without a test set? Randomly partition the data into 10% and 90% sets. Use the 90% as training data, to grow tree model using cross validation (1se rule) and also grow a random forest. Then predict the 10% test data. Repeat the procedure 100 times and average the result.R code is as follows.
>misv.tree<rep(0,100);
>sizev.tree<rep(0,100);
>misv.forest<rep(0,100);
>for (j in 1:100)
{ list<sample(seq(1:683),70,replace=F) train<data.frame(bc[ list,]); test<data.frame(bc[list,]) tr0<rpart(factor(Y)~.,data=train, control=rpart.control( minsplit=10, minbucket=5, cp=0.0,xval=10)) x<printcp(tr0) bs<x[1,2]+1; min<x[1,4]; cpv<x[1,1]; stde<x[1,5] for (i in 1:length(x[,1])) { if(x[i,4]<min ) { min<x[i,4] bs<x[i,2]+1 cpv<x[i,1] stde<x[i,5] } } limit<min+stde ; index<0 ; for(i in 1:length(x[,1])) { if(index<1){ if(x[i,4]>limit) { bs<x[i+1,2]+1 cpv<x[i+1,1] } else index<2 } } tr<prune(tr0,cp=cpv) # prune tree fity<predict(tr,newdata=test,type='class') table<table(test[,1],fity) mis<table[1,2]+table[2,1] misv.tree[j]<mis sizev.tree[j]<bs forest<randomForest(factor(Y)~.,data=train,mtry=4,ntree=100) # random forest fity<predict(forest,newdata=test,type='class') table<table(test[,1],fity) mis<table[1,2]+table[2,1] misv.forest[j]<mis
}