stat946w18/AmbientGAN: Generative Models from Lossy Measurements: Difference between revisions

From statwiki
Jump to navigation Jump to search
m (Basic information)
(Introduction into the basic idea of AmbientGAN)
Line 1: Line 1:
= Introduction =
= Introduction =
Generative Adversarial Networks operate by simulating complex distributions but training them requires access to large amounts of high quality data. Ashish Bora, Eric Price and Alexandros G. Dimakis propose AmbientGAN as a way to recover the true underlying distribution from lossy data under certain conditions. Even when these conditions are not met the results from AmbientGAN still produce high quality results.  
Generative Adversarial Networks operate by simulating complex distributions but training them requires access to large amounts of high quality data. Often times we only have access to noisy or partial observations, from here on referred to as measurements of the true data. If know the measurement function and we would like to train a generative model for the true data there are several ways to continue with varying degrees of success. We will use noisy MNIST data as an illustrative example. Suppose we only see MNIST data that have been run through a Gaussian kernel (blurred) with some noise from a <math>N(0, 0.5^2)</math> distribution added to each pixel:


= Problem Specification =
<gallery mode="packed">
File:mnist.png| True Data (Unobserved)
File:mnistmeasured.png| Measured Data (Observed)
</gallery>
 
 
=== Ignore the problem ===
[[File:GANignore.png|500px]] [[File:mnistignore.png|300px]]
 
Train a generative model directly on the measured data. This will obviously be unable to generate the true distribution before measurement has occurred.
 
 
=== Try to recover the information lost ===
[[File:GANrecovery.png|420px]] [[File:mnistrecover.png|300px]]
 
Works better than ignoring the problem but depends on how easily the measurement function can be inversed.
 
=== AmbientGAN ===
[[File:GANambient.png|500px]] [[File:mnistambient.png|300px]]
 
Ashish Bora, Eric Price and Alexandros G. Dimakis propose AmbientGAN as a way to recover the true underlying distribution from measurements of the true data.
 
AmbientGAN works by training a generator which attempts to have the measruements of the output it generates fool the discriminator. The discriminator must distinguish between real and generated measurements.


= Model =
= Model =
AmbientGAN compared to GAN (AmbientGAN passes the generator output through a measurement function)  
For the following variables superscript <math>r</math> represents the true distributions while superscript <math>g</math> represents the generated distributions. Let <math>x</math>, represent the underlying space and <math>y</math> for the measurement.
 
Thus <math>p_x^r</math> is the real underlying distribution over <math>\mathbb{R}^n</math> that we are interested in. However if we assume our (known) measurement functions, <math>f_\theta: \mathbb{R}^n \to \mathbb{R}^m</math> are parameterized by <math>\theta \sim p_\theta</math>, we can only observe <math>y = f_\theta(x)</math>.
 
 
 
AmbientGAN has the objective function:
 
<math>\min_G \max_D \mathbb{E}_{Y^r \sim p_y^r}[q(D(Y^r))] + \mathbb{E}_{Z \sim p_z, \Theta \sim p_\theta}[q(1 - D(f_\Theta(G(Z))))] </math>


= Results =
= Empirical Results =


= Open questions =
= Theoretical Results =


= Criticisms =
= Criticisms =

Revision as of 04:29, 27 February 2018

Introduction

Generative Adversarial Networks operate by simulating complex distributions but training them requires access to large amounts of high quality data. Often times we only have access to noisy or partial observations, from here on referred to as measurements of the true data. If know the measurement function and we would like to train a generative model for the true data there are several ways to continue with varying degrees of success. We will use noisy MNIST data as an illustrative example. Suppose we only see MNIST data that have been run through a Gaussian kernel (blurred) with some noise from a [math]\displaystyle{ N(0, 0.5^2) }[/math] distribution added to each pixel:


Ignore the problem

Train a generative model directly on the measured data. This will obviously be unable to generate the true distribution before measurement has occurred.


Try to recover the information lost

Works better than ignoring the problem but depends on how easily the measurement function can be inversed.

AmbientGAN

Ashish Bora, Eric Price and Alexandros G. Dimakis propose AmbientGAN as a way to recover the true underlying distribution from measurements of the true data.

AmbientGAN works by training a generator which attempts to have the measruements of the output it generates fool the discriminator. The discriminator must distinguish between real and generated measurements.

Model

For the following variables superscript [math]\displaystyle{ r }[/math] represents the true distributions while superscript [math]\displaystyle{ g }[/math] represents the generated distributions. Let [math]\displaystyle{ x }[/math], represent the underlying space and [math]\displaystyle{ y }[/math] for the measurement.

Thus [math]\displaystyle{ p_x^r }[/math] is the real underlying distribution over [math]\displaystyle{ \mathbb{R}^n }[/math] that we are interested in. However if we assume our (known) measurement functions, [math]\displaystyle{ f_\theta: \mathbb{R}^n \to \mathbb{R}^m }[/math] are parameterized by [math]\displaystyle{ \theta \sim p_\theta }[/math], we can only observe [math]\displaystyle{ y = f_\theta(x) }[/math].


AmbientGAN has the objective function:

[math]\displaystyle{ \min_G \max_D \mathbb{E}_{Y^r \sim p_y^r}[q(D(Y^r))] + \mathbb{E}_{Z \sim p_z, \Theta \sim p_\theta}[q(1 - D(f_\Theta(G(Z))))] }[/math]

Empirical Results

Theoretical Results

Criticisms

References

https://openreview.net/forum?id=Hy7fDog0b