SlideShare une entreprise Scribd logo
1  sur  71
Télécharger pour lire hors ligne
Plug & Play Generative Networks:
Conditional Iterative Generation of Images in Latent Space
Anh Nguyen, Jason Yosinski, Yoshua
Bengio, Alexey Dosovitskiy, Jeff Clune
[GitHub] [Arxiv]
Slides by Víctor Garcia
UPC Computer Vision Reading Group (27/01/2017)
Index
● Introduction
● Probabilistic Interpretation of the method
● Methods and Experiments
○ PPGN-x: DAE model of p(x)
○ DGN-AM: sampling without a learned prior
○ PPGN-h: Generator and DAE model of p(h)
○ Joint PPGN-h: joint Generator and DAE
● Further Experiments
○ Image Generation: Captioning
○ Image Generation: Multifaceted Feature Visualization
○ Image inpainting
● Conclusions
Introduction
Interpretation of different frameworks to generate images maximizing:
p(x, y) = p(x)*p(y|x)
Prior Condition
Encourages to
look realistic
Encourages to
look from a
particular class
Introduction
Image Generation:
● High Resolution Images
(227x227)
GANs struggle to Generate >64x64 Images
Introduction
Image Generation:
● High Resolution Images
● Intra-Class Variance
Introduction
Image Generation:
● High Resolution Images
● Intra-Class Variance
● Inter-Class Variance
(1000-ImageNet classes)
Index
● Introduction
● Probabilistic Interpretation of the method
● Methods and Experiments
○ PPGN-x: DAE model of p(x)
○ DGN-AM: sampling without a learned prior
○ PPGN-h: Generator and DAE model of p(h)
○ Joint PPGN-h: joint Generator and DAE
● Further Experiments
○ Image Generation: Captioning
○ Image Generation: Multifaceted Feature Visualization
○ Image inpainting
● Conclusions
Probabilistic Interpretation of the method
Metropolis-adjusted Langevin algorithm (MALA) which is a MCMC algorithm for
iteratively producing random samples from a distribution p(x):
Probabilistic Interpretation of the method
Metropolis-adjusted Langevin algorithm (MALA) which is a MCMC algorithm for
iteratively producing random samples:
Current state
Probabilistic Interpretation of the method
Metropolis-adjusted Langevin algorithm (MALA) which is a MCMC algorithm for
iteratively producing random samples:
Future State Current state
Probabilistic Interpretation of the method
Metropolis-adjusted Langevin algorithm (MALA) which is a MCMC algorithm for
iteratively producing random samples:
Future State Current state Gradient to the
natural manifold of
p(x)
Probabilistic Interpretation of the method
Metropolis-adjusted Langevin algorithm (MALA) which is a MCMC algorithm for
iteratively producing random samples:
Gradient to the
natural manifold of
p(x)
NoiseFuture State Current state
Probabilistic Interpretation of the method
Future State Current state Gradient to the
natural manifold
of p(x)
Noise
Probabilistic Interpretation of the method
p(x)
Probabilistic Interpretation of the method
p(x)
Step towards an image that
causes the classifier to produce
a higher score for class C
Step towards a more
generic image
Noise
Probabilistic Interpretation of the method
xt
Rough
example
Probabilistic Interpretation of the method
y_co = Content activations y_st = Style activations
Rough
example
Probabilistic Interpretation of the method
xt+i
Rough
example
Index
● Introduction
● Probabilistic Interpretation of the method
● Methods and Experiments
○ PPGN-x: DAE model of p(x)
○ DGN-AM: sampling without a learned prior
○ PPGN-h: Generator and DAE model of p(h)
○ Joint PPGN-h: joint Generator and DAE
● Further Experiments
○ Image Generation: Captioning
○ Image Generation: Multifaceted Feature Visualization
○ Image inpainting
● Conclusions
Method
Why Plug & Play ?
Index
● Introduction
● Probabilistic Interpretation of the method
● Methods and Experiments
○ PPGN-x: DAE model of p(x)
○ DGN-AM: sampling without a learned prior
○ PPGN-h: Generator and DAE model of p(h)
○ Joint PPGN-h: joint Generator and DAE
● Further Experiments
○ Image Generation: Captioning
○ Image Generation: Multifaceted Feature Visualization
○ Image inpainting
● Conclusions
Method | PPGN-x: DAE model of p(x)
What a Denoising Autoencoder is?
x
h(x)
R(x)
Method | PPGN-x: DAE model of p(x)
What a Denoising Autoencoder is?
x_noise
h(x)
x
N(0,σ^2)
R(x)
Method | PPGN-x: DAE model of p(x)
What a Denoising Autoencoder is?
x_noise
h(x)
x
N(0,σ^2)
R(x)
Method | PPGN-x: DAE model of p(x)
Method | PPGN-x: DAE model of p(x)
Method | PPGN-x: DAE model of p(x)
1) Poorly modeled data, blurry 2) Slow changes
Index
● Introduction
● Probabilistic Interpretation of the method
● Methods and Experiments
○ PPGN-x: DAE model of p(x)
○ DGN-AM: sampling without a learned prior
○ PPGN-h: Generator and DAE model of p(h)
○ Joint PPGN-h: joint Generator and DAE
● Further Experiments
○ Image Generation: Captioning
○ Image Generation: Multifaceted Feature Visualization
○ Image inpainting
● Conclusions
Method | DGN-AM: sampling without a learned prior
Deep Generator Network-based Activation Maximization
It is faster if we move over h subspace instead of the x
fc6
AlexNet
Method | DGN-AM: sampling without a learned prior
Deep Generator Network-based Activation Maximization
Discriminator 1/0
AlexNet
fc6
Method | DGN-AM: sampling without a learned prior
Once we trained the network G we find the equation for the MALA algorithm
Method | DGN-AM: sampling without a learned prior
Once we trained the network G we find the equation for the MALA algorithm
Method | DGN-AM: sampling without a learned prior
Once we trained the network G we find the equation for the MALA algorithm
Method | DGN-AM: sampling without a learned prior
Once we trained the network G we find the equation for the MALA algorithm
No learned prior No noise
Method | DGN-AM: sampling without a learned prior
+ Different modes from different starts
- Same image after many steps
- Low mixing speed
Index
● Introduction
● Probabilistic Interpretation of the method
● Methods and Experiments
○ PPGN-x: DAE model of p(x)
○ DGN-AM: sampling without a learned prior
○ PPGN-h: Generator and DAE model of p(h)
○ Joint PPGN-h: joint Generator and DAE
● Further Experiments
○ Image Generation: Captioning
○ Image Generation: Multifaceted Feature Visualization
○ Image inpainting
● Conclusions
Method | PPGN-h: Generator and DAE model of p(h)
A 7 layers DAE is added to model the prior p(h) in order to increase the mixing speed
Method | PPGN-h: Generator and DAE model of p(h)
The equation is the following:
Prior p(h) Conditioned
Gradient
Noise
Method | PPGN-h: Generator and DAE model of p(h)
- Similar to the last case. Low diversity
- p(h) model learned by DAE is too simple
Index
● Introduction
● Probabilistic Interpretation of the method
● Methods and Experiments
○ PPGN-x: DAE model of p(x)
○ DGN-AM: sampling without a learned prior
○ PPGN-h: Generator and DAE model of p(h)
○ Joint PPGN-h: joint Generator and DAE
● Further Experiments
○ Image Generation: Captioning
○ Image Generation: Multifaceted Feature Visualization
○ Image inpainting
● Conclusions
Method | Joint PPGN-h: joint Generator and DAE
In order to model p(h) in a more complex way
DAE: h/fc6 → ? → h/fc6
Method | Joint PPGN-h: joint Generator and DAE
In order to model p(h) in a more complex way
DAE: h/fc6 → ? → h/fc6
Joint Generator and DAE: h/fc6 x h/fc6
G E
Method | Joint PPGN-h: joint Generator and DAE
In order to model p(h) in a more complex way
DAE: h/fc6 → ? → h/fc6
Joint Generator and DAE: h/fc6 x h/fc6
G E
With the same existing network we train the Generator G to act as a DAE in conjunction with the E
network
Method | Joint PPGN-h: joint Generator and DAE
AlexNet
Equation is the
same than before
Method | Joint PPGN-h: joint Generator and DAE
- Faster mixing
- Better quality
Method | Joint PPGN-h: joint Generator and AE
AlexNet
Equation is the
same than before
Method | Joint PPGN-h: joint Generator and AE
- Faster mixing
- Better quality
Method | Joint PPGN-h: joint Generator and DAE
Noise sweeps
For the last model we test the reconstruction of different h/fc6 vectors when adding different noise levels:
fc6
N(0, ) +
Method | Joint PPGN-h: joint Generator and AE
Noise sweeps
For the last model we test the reconstruction of different h/fc6 vectors when adding different noise levels:
Method | Joint PPGN-h: joint Generator and AE
Noise sweeps
Method | Joint PPGN-h: joint Generator and AE
Noise sweeps
We can still recover large information from the image when mapping with a lot of noise.
Many → one.
Method | Joint PPGN-h: joint Generator and DAE
Combination of Losses
Comparison of Losses:
● Real Images
●
●
●
●
Method | Joint PPGN-h: joint Generator and DAE
Combination of Losses
Method | Joint PPGN-h: joint Generator and DAE
Combination of Losses
Method | Joint PPGN-h: joint Generator and DAE
Evaluating: Qualitatively
Method | Joint PPGN-h: joint Generator and DAE
Evaluating: Qualitatively
Method | Joint PPGN-h: joint Generator and DAE
Evaluating: Qualitatively
Index
● Introduction
● Probabilistic Interpretation of the method
● Methods and Experiments
○ PPGN-x: DAE model of p(x)
○ DGN-AM: sampling without a learned prior
○ PPGN-h: Generator and DAE model of p(h)
○ Joint PPGN-h: joint Generator and DAE
● Further Experiments
○ Image Generation: Captioning
○ Image Generation: Multifaceted Feature Visualization
○ Image inpainting
● Conclusions
Further Experiments | Captioning
MS-COCO Dataset
Further Experiments | Captioning
Index
● Introduction
● Probabilistic Interpretation of the method
● Methods and Experiments
○ PPGN-x: DAE model of p(x)
○ DGN-AM: sampling without a learned prior
○ PPGN-h: Generator and DAE model of p(h)
○ Joint PPGN-h: joint Generator and DAE
● Further Experiments
○ Image Generation: Captioning
○ Image Generation: Multifaceted Feature Visualization
○ Image inpainting
● Conclusions
Further Experiments | MFV
Multifaceted Feature Visualization
Multifaceted Feature Visualization
Further Experiments | MFV
Index
● Introduction
● Probabilistic Interpretation of the method
● Methods and Experiments
○ PPGN-x: DAE model of p(x)
○ DGN-AM: sampling without a learned prior
○ PPGN-h: Generator and DAE model of p(h)
○ Joint PPGN-h: joint Generator and DAE
● Further Experiments
○ Image Generation: Captioning
○ Image Generation: Multifaceted Feature Visualization
○ Image inpainting
● Conclusions
Further Experiments | Inpainting
Multifaceted Feature Visualization
Further Experiments | Inpainting
Multifaceted Feature Visualization
Further Experiments | Inpainting
Multifaceted Feature Visualization
Further Experiments | Inpainting
Multifaceted Feature Visualization
Further Experiments | Inpainting
Multifaceted Feature Visualization
Conclusions
● Only using GANs for the reconstruction, GANs collapse into fewer modes, far
from the original p(x).
● Using extra Losses it is possible to better reconstruct the images even for 1000
classes and for higher resolution. Mapping one-to-one helps to prevent typical
latent → missing modes.
● It would be great to generate also the embedding space for this
super-resolution multi-class images instead of using a supervised learned
space.
Plug & Play Generative Networks: Conditional Iterative Generation of Images in Latent Space (UPC Reading Group)

Contenu connexe

En vedette

End-to-end Speech Recognition with Recurrent Neural Networks (D3L6 Deep Learn...
End-to-end Speech Recognition with Recurrent Neural Networks (D3L6 Deep Learn...End-to-end Speech Recognition with Recurrent Neural Networks (D3L6 Deep Learn...
End-to-end Speech Recognition with Recurrent Neural Networks (D3L6 Deep Learn...Universitat Politècnica de Catalunya
 
Generative Adversarial Networks and Their Applications
Generative Adversarial Networks and Their ApplicationsGenerative Adversarial Networks and Their Applications
Generative Adversarial Networks and Their ApplicationsArtifacia
 
Advanced Neural Machine Translation (D4L2 Deep Learning for Speech and Langua...
Advanced Neural Machine Translation (D4L2 Deep Learning for Speech and Langua...Advanced Neural Machine Translation (D4L2 Deep Learning for Speech and Langua...
Advanced Neural Machine Translation (D4L2 Deep Learning for Speech and Langua...Universitat Politècnica de Catalunya
 
Image-to-Image Translation with Conditional Adversarial Nets (UPC Reading Group)
Image-to-Image Translation with Conditional Adversarial Nets (UPC Reading Group)Image-to-Image Translation with Conditional Adversarial Nets (UPC Reading Group)
Image-to-Image Translation with Conditional Adversarial Nets (UPC Reading Group)Universitat Politècnica de Catalunya
 
Neural Machine Translation (D3L4 Deep Learning for Speech and Language UPC 2017)
Neural Machine Translation (D3L4 Deep Learning for Speech and Language UPC 2017)Neural Machine Translation (D3L4 Deep Learning for Speech and Language UPC 2017)
Neural Machine Translation (D3L4 Deep Learning for Speech and Language UPC 2017)Universitat Politècnica de Catalunya
 
Generative Adversarial Networks (D2L5 Deep Learning for Speech and Language U...
Generative Adversarial Networks (D2L5 Deep Learning for Speech and Language U...Generative Adversarial Networks (D2L5 Deep Learning for Speech and Language U...
Generative Adversarial Networks (D2L5 Deep Learning for Speech and Language U...Universitat Politècnica de Catalunya
 
A pixel to-pixel segmentation method of DILD without masks using CNN and perl...
A pixel to-pixel segmentation method of DILD without masks using CNN and perl...A pixel to-pixel segmentation method of DILD without masks using CNN and perl...
A pixel to-pixel segmentation method of DILD without masks using CNN and perl...남주 김
 
SSD: Single Shot MultiBox Detector (ECCV2016)
SSD: Single Shot MultiBox Detector (ECCV2016)SSD: Single Shot MultiBox Detector (ECCV2016)
SSD: Single Shot MultiBox Detector (ECCV2016)Takanori Ogata
 
Generative adversarial networks
Generative adversarial networksGenerative adversarial networks
Generative adversarial networks남주 김
 
Deep convnets for global recognition (Master in Computer Vision Barcelona 2016)
Deep convnets for global recognition (Master in Computer Vision Barcelona 2016)Deep convnets for global recognition (Master in Computer Vision Barcelona 2016)
Deep convnets for global recognition (Master in Computer Vision Barcelona 2016)Universitat Politècnica de Catalunya
 

En vedette (16)

End-to-end Speech Recognition with Recurrent Neural Networks (D3L6 Deep Learn...
End-to-end Speech Recognition with Recurrent Neural Networks (D3L6 Deep Learn...End-to-end Speech Recognition with Recurrent Neural Networks (D3L6 Deep Learn...
End-to-end Speech Recognition with Recurrent Neural Networks (D3L6 Deep Learn...
 
Deep Learning for Computer Vision: Object Detection (UPC 2016)
Deep Learning for Computer Vision: Object Detection (UPC 2016)Deep Learning for Computer Vision: Object Detection (UPC 2016)
Deep Learning for Computer Vision: Object Detection (UPC 2016)
 
Generative Adversarial Networks and Their Applications
Generative Adversarial Networks and Their ApplicationsGenerative Adversarial Networks and Their Applications
Generative Adversarial Networks and Their Applications
 
Advanced Neural Machine Translation (D4L2 Deep Learning for Speech and Langua...
Advanced Neural Machine Translation (D4L2 Deep Learning for Speech and Langua...Advanced Neural Machine Translation (D4L2 Deep Learning for Speech and Langua...
Advanced Neural Machine Translation (D4L2 Deep Learning for Speech and Langua...
 
Generative adversarial text to image synthesis
Generative adversarial text to image synthesisGenerative adversarial text to image synthesis
Generative adversarial text to image synthesis
 
Image-to-Image Translation with Conditional Adversarial Nets (UPC Reading Group)
Image-to-Image Translation with Conditional Adversarial Nets (UPC Reading Group)Image-to-Image Translation with Conditional Adversarial Nets (UPC Reading Group)
Image-to-Image Translation with Conditional Adversarial Nets (UPC Reading Group)
 
Speaker ID II (D4L1 Deep Learning for Speech and Language UPC 2017)
Speaker ID II (D4L1 Deep Learning for Speech and Language UPC 2017)Speaker ID II (D4L1 Deep Learning for Speech and Language UPC 2017)
Speaker ID II (D4L1 Deep Learning for Speech and Language UPC 2017)
 
Neural Machine Translation (D3L4 Deep Learning for Speech and Language UPC 2017)
Neural Machine Translation (D3L4 Deep Learning for Speech and Language UPC 2017)Neural Machine Translation (D3L4 Deep Learning for Speech and Language UPC 2017)
Neural Machine Translation (D3L4 Deep Learning for Speech and Language UPC 2017)
 
Generative Adversarial Networks (D2L5 Deep Learning for Speech and Language U...
Generative Adversarial Networks (D2L5 Deep Learning for Speech and Language U...Generative Adversarial Networks (D2L5 Deep Learning for Speech and Language U...
Generative Adversarial Networks (D2L5 Deep Learning for Speech and Language U...
 
A pixel to-pixel segmentation method of DILD without masks using CNN and perl...
A pixel to-pixel segmentation method of DILD without masks using CNN and perl...A pixel to-pixel segmentation method of DILD without masks using CNN and perl...
A pixel to-pixel segmentation method of DILD without masks using CNN and perl...
 
Open-ended Visual Question-Answering
Open-ended  Visual Question-AnsweringOpen-ended  Visual Question-Answering
Open-ended Visual Question-Answering
 
SSD: Single Shot MultiBox Detector (ECCV2016)
SSD: Single Shot MultiBox Detector (ECCV2016)SSD: Single Shot MultiBox Detector (ECCV2016)
SSD: Single Shot MultiBox Detector (ECCV2016)
 
ECCV 2016 速報
ECCV 2016 速報ECCV 2016 速報
ECCV 2016 速報
 
Generative adversarial networks
Generative adversarial networksGenerative adversarial networks
Generative adversarial networks
 
Deep convnets for global recognition (Master in Computer Vision Barcelona 2016)
Deep convnets for global recognition (Master in Computer Vision Barcelona 2016)Deep convnets for global recognition (Master in Computer Vision Barcelona 2016)
Deep convnets for global recognition (Master in Computer Vision Barcelona 2016)
 
SSD: Single Shot MultiBox Detector (UPC Reading Group)
SSD: Single Shot MultiBox Detector (UPC Reading Group)SSD: Single Shot MultiBox Detector (UPC Reading Group)
SSD: Single Shot MultiBox Detector (UPC Reading Group)
 

Plus de Universitat Politècnica de Catalunya

The Transformer in Vision | Xavier Giro | Master in Computer Vision Barcelona...
The Transformer in Vision | Xavier Giro | Master in Computer Vision Barcelona...The Transformer in Vision | Xavier Giro | Master in Computer Vision Barcelona...
The Transformer in Vision | Xavier Giro | Master in Computer Vision Barcelona...Universitat Politècnica de Catalunya
 
Towards Sign Language Translation & Production | Xavier Giro-i-Nieto
Towards Sign Language Translation & Production | Xavier Giro-i-NietoTowards Sign Language Translation & Production | Xavier Giro-i-Nieto
Towards Sign Language Translation & Production | Xavier Giro-i-NietoUniversitat Politècnica de Catalunya
 
Learning Representations for Sign Language Videos - Xavier Giro - NIST TRECVI...
Learning Representations for Sign Language Videos - Xavier Giro - NIST TRECVI...Learning Representations for Sign Language Videos - Xavier Giro - NIST TRECVI...
Learning Representations for Sign Language Videos - Xavier Giro - NIST TRECVI...Universitat Politècnica de Catalunya
 
Generation of Synthetic Referring Expressions for Object Segmentation in Videos
Generation of Synthetic Referring Expressions for Object Segmentation in VideosGeneration of Synthetic Referring Expressions for Object Segmentation in Videos
Generation of Synthetic Referring Expressions for Object Segmentation in VideosUniversitat Politècnica de Catalunya
 
Learn2Sign : Sign language recognition and translation using human keypoint e...
Learn2Sign : Sign language recognition and translation using human keypoint e...Learn2Sign : Sign language recognition and translation using human keypoint e...
Learn2Sign : Sign language recognition and translation using human keypoint e...Universitat Politècnica de Catalunya
 
Convolutional Neural Networks - Xavier Giro - UPC TelecomBCN Barcelona 2020
Convolutional Neural Networks - Xavier Giro - UPC TelecomBCN Barcelona 2020Convolutional Neural Networks - Xavier Giro - UPC TelecomBCN Barcelona 2020
Convolutional Neural Networks - Xavier Giro - UPC TelecomBCN Barcelona 2020Universitat Politècnica de Catalunya
 
Self-Supervised Audio-Visual Learning - Xavier Giro - UPC TelecomBCN Barcelon...
Self-Supervised Audio-Visual Learning - Xavier Giro - UPC TelecomBCN Barcelon...Self-Supervised Audio-Visual Learning - Xavier Giro - UPC TelecomBCN Barcelon...
Self-Supervised Audio-Visual Learning - Xavier Giro - UPC TelecomBCN Barcelon...Universitat Politècnica de Catalunya
 
Attention for Deep Learning - Xavier Giro - UPC TelecomBCN Barcelona 2020
Attention for Deep Learning - Xavier Giro - UPC TelecomBCN Barcelona 2020Attention for Deep Learning - Xavier Giro - UPC TelecomBCN Barcelona 2020
Attention for Deep Learning - Xavier Giro - UPC TelecomBCN Barcelona 2020Universitat Politècnica de Catalunya
 
Generative Adversarial Networks GAN - Xavier Giro - UPC TelecomBCN Barcelona ...
Generative Adversarial Networks GAN - Xavier Giro - UPC TelecomBCN Barcelona ...Generative Adversarial Networks GAN - Xavier Giro - UPC TelecomBCN Barcelona ...
Generative Adversarial Networks GAN - Xavier Giro - UPC TelecomBCN Barcelona ...Universitat Politècnica de Catalunya
 
Q-Learning with a Neural Network - Xavier Giró - UPC Barcelona 2020
Q-Learning with a Neural Network - Xavier Giró - UPC Barcelona 2020Q-Learning with a Neural Network - Xavier Giró - UPC Barcelona 2020
Q-Learning with a Neural Network - Xavier Giró - UPC Barcelona 2020Universitat Politècnica de Catalunya
 
Language and Vision with Deep Learning - Xavier Giró - ACM ICMR 2020 (Tutorial)
Language and Vision with Deep Learning - Xavier Giró - ACM ICMR 2020 (Tutorial)Language and Vision with Deep Learning - Xavier Giró - ACM ICMR 2020 (Tutorial)
Language and Vision with Deep Learning - Xavier Giró - ACM ICMR 2020 (Tutorial)Universitat Politècnica de Catalunya
 
Image Segmentation with Deep Learning - Xavier Giro & Carles Ventura - ISSonD...
Image Segmentation with Deep Learning - Xavier Giro & Carles Ventura - ISSonD...Image Segmentation with Deep Learning - Xavier Giro & Carles Ventura - ISSonD...
Image Segmentation with Deep Learning - Xavier Giro & Carles Ventura - ISSonD...Universitat Politècnica de Catalunya
 

Plus de Universitat Politècnica de Catalunya (20)

Deep Generative Learning for All - The Gen AI Hype (Spring 2024)
Deep Generative Learning for All - The Gen AI Hype (Spring 2024)Deep Generative Learning for All - The Gen AI Hype (Spring 2024)
Deep Generative Learning for All - The Gen AI Hype (Spring 2024)
 
Deep Generative Learning for All
Deep Generative Learning for AllDeep Generative Learning for All
Deep Generative Learning for All
 
The Transformer in Vision | Xavier Giro | Master in Computer Vision Barcelona...
The Transformer in Vision | Xavier Giro | Master in Computer Vision Barcelona...The Transformer in Vision | Xavier Giro | Master in Computer Vision Barcelona...
The Transformer in Vision | Xavier Giro | Master in Computer Vision Barcelona...
 
Towards Sign Language Translation & Production | Xavier Giro-i-Nieto
Towards Sign Language Translation & Production | Xavier Giro-i-NietoTowards Sign Language Translation & Production | Xavier Giro-i-Nieto
Towards Sign Language Translation & Production | Xavier Giro-i-Nieto
 
The Transformer - Xavier Giró - UPC Barcelona 2021
The Transformer - Xavier Giró - UPC Barcelona 2021The Transformer - Xavier Giró - UPC Barcelona 2021
The Transformer - Xavier Giró - UPC Barcelona 2021
 
Learning Representations for Sign Language Videos - Xavier Giro - NIST TRECVI...
Learning Representations for Sign Language Videos - Xavier Giro - NIST TRECVI...Learning Representations for Sign Language Videos - Xavier Giro - NIST TRECVI...
Learning Representations for Sign Language Videos - Xavier Giro - NIST TRECVI...
 
Open challenges in sign language translation and production
Open challenges in sign language translation and productionOpen challenges in sign language translation and production
Open challenges in sign language translation and production
 
Generation of Synthetic Referring Expressions for Object Segmentation in Videos
Generation of Synthetic Referring Expressions for Object Segmentation in VideosGeneration of Synthetic Referring Expressions for Object Segmentation in Videos
Generation of Synthetic Referring Expressions for Object Segmentation in Videos
 
Discovery and Learning of Navigation Goals from Pixels in Minecraft
Discovery and Learning of Navigation Goals from Pixels in MinecraftDiscovery and Learning of Navigation Goals from Pixels in Minecraft
Discovery and Learning of Navigation Goals from Pixels in Minecraft
 
Learn2Sign : Sign language recognition and translation using human keypoint e...
Learn2Sign : Sign language recognition and translation using human keypoint e...Learn2Sign : Sign language recognition and translation using human keypoint e...
Learn2Sign : Sign language recognition and translation using human keypoint e...
 
Intepretability / Explainable AI for Deep Neural Networks
Intepretability / Explainable AI for Deep Neural NetworksIntepretability / Explainable AI for Deep Neural Networks
Intepretability / Explainable AI for Deep Neural Networks
 
Convolutional Neural Networks - Xavier Giro - UPC TelecomBCN Barcelona 2020
Convolutional Neural Networks - Xavier Giro - UPC TelecomBCN Barcelona 2020Convolutional Neural Networks - Xavier Giro - UPC TelecomBCN Barcelona 2020
Convolutional Neural Networks - Xavier Giro - UPC TelecomBCN Barcelona 2020
 
Self-Supervised Audio-Visual Learning - Xavier Giro - UPC TelecomBCN Barcelon...
Self-Supervised Audio-Visual Learning - Xavier Giro - UPC TelecomBCN Barcelon...Self-Supervised Audio-Visual Learning - Xavier Giro - UPC TelecomBCN Barcelon...
Self-Supervised Audio-Visual Learning - Xavier Giro - UPC TelecomBCN Barcelon...
 
Attention for Deep Learning - Xavier Giro - UPC TelecomBCN Barcelona 2020
Attention for Deep Learning - Xavier Giro - UPC TelecomBCN Barcelona 2020Attention for Deep Learning - Xavier Giro - UPC TelecomBCN Barcelona 2020
Attention for Deep Learning - Xavier Giro - UPC TelecomBCN Barcelona 2020
 
Generative Adversarial Networks GAN - Xavier Giro - UPC TelecomBCN Barcelona ...
Generative Adversarial Networks GAN - Xavier Giro - UPC TelecomBCN Barcelona ...Generative Adversarial Networks GAN - Xavier Giro - UPC TelecomBCN Barcelona ...
Generative Adversarial Networks GAN - Xavier Giro - UPC TelecomBCN Barcelona ...
 
Q-Learning with a Neural Network - Xavier Giró - UPC Barcelona 2020
Q-Learning with a Neural Network - Xavier Giró - UPC Barcelona 2020Q-Learning with a Neural Network - Xavier Giró - UPC Barcelona 2020
Q-Learning with a Neural Network - Xavier Giró - UPC Barcelona 2020
 
Language and Vision with Deep Learning - Xavier Giró - ACM ICMR 2020 (Tutorial)
Language and Vision with Deep Learning - Xavier Giró - ACM ICMR 2020 (Tutorial)Language and Vision with Deep Learning - Xavier Giró - ACM ICMR 2020 (Tutorial)
Language and Vision with Deep Learning - Xavier Giró - ACM ICMR 2020 (Tutorial)
 
Image Segmentation with Deep Learning - Xavier Giro & Carles Ventura - ISSonD...
Image Segmentation with Deep Learning - Xavier Giro & Carles Ventura - ISSonD...Image Segmentation with Deep Learning - Xavier Giro & Carles Ventura - ISSonD...
Image Segmentation with Deep Learning - Xavier Giro & Carles Ventura - ISSonD...
 
Curriculum Learning for Recurrent Video Object Segmentation
Curriculum Learning for Recurrent Video Object SegmentationCurriculum Learning for Recurrent Video Object Segmentation
Curriculum Learning for Recurrent Video Object Segmentation
 
Deep Self-supervised Learning for All - Xavier Giro - X-Europe 2020
Deep Self-supervised Learning for All - Xavier Giro - X-Europe 2020Deep Self-supervised Learning for All - Xavier Giro - X-Europe 2020
Deep Self-supervised Learning for All - Xavier Giro - X-Europe 2020
 

Dernier

How we prevented account sharing with MFA
How we prevented account sharing with MFAHow we prevented account sharing with MFA
How we prevented account sharing with MFAAndrei Kaleshka
 
Student Profile Sample report on improving academic performance by uniting gr...
Student Profile Sample report on improving academic performance by uniting gr...Student Profile Sample report on improving academic performance by uniting gr...
Student Profile Sample report on improving academic performance by uniting gr...Seán Kennedy
 
Defining Constituents, Data Vizzes and Telling a Data Story
Defining Constituents, Data Vizzes and Telling a Data StoryDefining Constituents, Data Vizzes and Telling a Data Story
Defining Constituents, Data Vizzes and Telling a Data StoryJeremy Anderson
 
NO1 Certified Black Magic Specialist Expert Amil baba in Lahore Islamabad Raw...
NO1 Certified Black Magic Specialist Expert Amil baba in Lahore Islamabad Raw...NO1 Certified Black Magic Specialist Expert Amil baba in Lahore Islamabad Raw...
NO1 Certified Black Magic Specialist Expert Amil baba in Lahore Islamabad Raw...Amil Baba Dawood bangali
 
专业一比一美国俄亥俄大学毕业证成绩单pdf电子版制作修改
专业一比一美国俄亥俄大学毕业证成绩单pdf电子版制作修改专业一比一美国俄亥俄大学毕业证成绩单pdf电子版制作修改
专业一比一美国俄亥俄大学毕业证成绩单pdf电子版制作修改yuu sss
 
RABBIT: A CLI tool for identifying bots based on their GitHub events.
RABBIT: A CLI tool for identifying bots based on their GitHub events.RABBIT: A CLI tool for identifying bots based on their GitHub events.
RABBIT: A CLI tool for identifying bots based on their GitHub events.natarajan8993
 
GA4 Without Cookies [Measure Camp AMS]
GA4 Without Cookies [Measure Camp AMS]GA4 Without Cookies [Measure Camp AMS]
GA4 Without Cookies [Measure Camp AMS]📊 Markus Baersch
 
Decoding the Heart: Student Presentation on Heart Attack Prediction with Data...
Decoding the Heart: Student Presentation on Heart Attack Prediction with Data...Decoding the Heart: Student Presentation on Heart Attack Prediction with Data...
Decoding the Heart: Student Presentation on Heart Attack Prediction with Data...Boston Institute of Analytics
 
Learn How Data Science Changes Our World
Learn How Data Science Changes Our WorldLearn How Data Science Changes Our World
Learn How Data Science Changes Our WorldEduminds Learning
 
modul pembelajaran robotic Workshop _ by Slidesgo.pptx
modul pembelajaran robotic Workshop _ by Slidesgo.pptxmodul pembelajaran robotic Workshop _ by Slidesgo.pptx
modul pembelajaran robotic Workshop _ by Slidesgo.pptxaleedritatuxx
 
DBA Basics: Getting Started with Performance Tuning.pdf
DBA Basics: Getting Started with Performance Tuning.pdfDBA Basics: Getting Started with Performance Tuning.pdf
DBA Basics: Getting Started with Performance Tuning.pdfJohn Sterrett
 
Data Analysis Project : Targeting the Right Customers, Presentation on Bank M...
Data Analysis Project : Targeting the Right Customers, Presentation on Bank M...Data Analysis Project : Targeting the Right Customers, Presentation on Bank M...
Data Analysis Project : Targeting the Right Customers, Presentation on Bank M...Boston Institute of Analytics
 
9711147426✨Call In girls Gurgaon Sector 31. SCO 25 escort service
9711147426✨Call In girls Gurgaon Sector 31. SCO 25 escort service9711147426✨Call In girls Gurgaon Sector 31. SCO 25 escort service
9711147426✨Call In girls Gurgaon Sector 31. SCO 25 escort servicejennyeacort
 
Advanced Machine Learning for Business Professionals
Advanced Machine Learning for Business ProfessionalsAdvanced Machine Learning for Business Professionals
Advanced Machine Learning for Business ProfessionalsVICTOR MAESTRE RAMIREZ
 
FAIR, FAIRsharing, FAIR Cookbook and ELIXIR - Sansone SA - Boston 2024
FAIR, FAIRsharing, FAIR Cookbook and ELIXIR - Sansone SA - Boston 2024FAIR, FAIRsharing, FAIR Cookbook and ELIXIR - Sansone SA - Boston 2024
FAIR, FAIRsharing, FAIR Cookbook and ELIXIR - Sansone SA - Boston 2024Susanna-Assunta Sansone
 
原版1:1定制南十字星大学毕业证(SCU毕业证)#文凭成绩单#真实留信学历认证永久存档
原版1:1定制南十字星大学毕业证(SCU毕业证)#文凭成绩单#真实留信学历认证永久存档原版1:1定制南十字星大学毕业证(SCU毕业证)#文凭成绩单#真实留信学历认证永久存档
原版1:1定制南十字星大学毕业证(SCU毕业证)#文凭成绩单#真实留信学历认证永久存档208367051
 
Thiophen Mechanism khhjjjjjjjhhhhhhhhhhh
Thiophen Mechanism khhjjjjjjjhhhhhhhhhhhThiophen Mechanism khhjjjjjjjhhhhhhhhhhh
Thiophen Mechanism khhjjjjjjjhhhhhhhhhhhYasamin16
 
办理(UWIC毕业证书)英国卡迪夫城市大学毕业证成绩单原版一比一
办理(UWIC毕业证书)英国卡迪夫城市大学毕业证成绩单原版一比一办理(UWIC毕业证书)英国卡迪夫城市大学毕业证成绩单原版一比一
办理(UWIC毕业证书)英国卡迪夫城市大学毕业证成绩单原版一比一F La
 
INTERNSHIP ON PURBASHA COMPOSITE TEX LTD
INTERNSHIP ON PURBASHA COMPOSITE TEX LTDINTERNSHIP ON PURBASHA COMPOSITE TEX LTD
INTERNSHIP ON PURBASHA COMPOSITE TEX LTDRafezzaman
 
MK KOMUNIKASI DATA (TI)komdat komdat.docx
MK KOMUNIKASI DATA (TI)komdat komdat.docxMK KOMUNIKASI DATA (TI)komdat komdat.docx
MK KOMUNIKASI DATA (TI)komdat komdat.docxUnduhUnggah1
 

Dernier (20)

How we prevented account sharing with MFA
How we prevented account sharing with MFAHow we prevented account sharing with MFA
How we prevented account sharing with MFA
 
Student Profile Sample report on improving academic performance by uniting gr...
Student Profile Sample report on improving academic performance by uniting gr...Student Profile Sample report on improving academic performance by uniting gr...
Student Profile Sample report on improving academic performance by uniting gr...
 
Defining Constituents, Data Vizzes and Telling a Data Story
Defining Constituents, Data Vizzes and Telling a Data StoryDefining Constituents, Data Vizzes and Telling a Data Story
Defining Constituents, Data Vizzes and Telling a Data Story
 
NO1 Certified Black Magic Specialist Expert Amil baba in Lahore Islamabad Raw...
NO1 Certified Black Magic Specialist Expert Amil baba in Lahore Islamabad Raw...NO1 Certified Black Magic Specialist Expert Amil baba in Lahore Islamabad Raw...
NO1 Certified Black Magic Specialist Expert Amil baba in Lahore Islamabad Raw...
 
专业一比一美国俄亥俄大学毕业证成绩单pdf电子版制作修改
专业一比一美国俄亥俄大学毕业证成绩单pdf电子版制作修改专业一比一美国俄亥俄大学毕业证成绩单pdf电子版制作修改
专业一比一美国俄亥俄大学毕业证成绩单pdf电子版制作修改
 
RABBIT: A CLI tool for identifying bots based on their GitHub events.
RABBIT: A CLI tool for identifying bots based on their GitHub events.RABBIT: A CLI tool for identifying bots based on their GitHub events.
RABBIT: A CLI tool for identifying bots based on their GitHub events.
 
GA4 Without Cookies [Measure Camp AMS]
GA4 Without Cookies [Measure Camp AMS]GA4 Without Cookies [Measure Camp AMS]
GA4 Without Cookies [Measure Camp AMS]
 
Decoding the Heart: Student Presentation on Heart Attack Prediction with Data...
Decoding the Heart: Student Presentation on Heart Attack Prediction with Data...Decoding the Heart: Student Presentation on Heart Attack Prediction with Data...
Decoding the Heart: Student Presentation on Heart Attack Prediction with Data...
 
Learn How Data Science Changes Our World
Learn How Data Science Changes Our WorldLearn How Data Science Changes Our World
Learn How Data Science Changes Our World
 
modul pembelajaran robotic Workshop _ by Slidesgo.pptx
modul pembelajaran robotic Workshop _ by Slidesgo.pptxmodul pembelajaran robotic Workshop _ by Slidesgo.pptx
modul pembelajaran robotic Workshop _ by Slidesgo.pptx
 
DBA Basics: Getting Started with Performance Tuning.pdf
DBA Basics: Getting Started with Performance Tuning.pdfDBA Basics: Getting Started with Performance Tuning.pdf
DBA Basics: Getting Started with Performance Tuning.pdf
 
Data Analysis Project : Targeting the Right Customers, Presentation on Bank M...
Data Analysis Project : Targeting the Right Customers, Presentation on Bank M...Data Analysis Project : Targeting the Right Customers, Presentation on Bank M...
Data Analysis Project : Targeting the Right Customers, Presentation on Bank M...
 
9711147426✨Call In girls Gurgaon Sector 31. SCO 25 escort service
9711147426✨Call In girls Gurgaon Sector 31. SCO 25 escort service9711147426✨Call In girls Gurgaon Sector 31. SCO 25 escort service
9711147426✨Call In girls Gurgaon Sector 31. SCO 25 escort service
 
Advanced Machine Learning for Business Professionals
Advanced Machine Learning for Business ProfessionalsAdvanced Machine Learning for Business Professionals
Advanced Machine Learning for Business Professionals
 
FAIR, FAIRsharing, FAIR Cookbook and ELIXIR - Sansone SA - Boston 2024
FAIR, FAIRsharing, FAIR Cookbook and ELIXIR - Sansone SA - Boston 2024FAIR, FAIRsharing, FAIR Cookbook and ELIXIR - Sansone SA - Boston 2024
FAIR, FAIRsharing, FAIR Cookbook and ELIXIR - Sansone SA - Boston 2024
 
原版1:1定制南十字星大学毕业证(SCU毕业证)#文凭成绩单#真实留信学历认证永久存档
原版1:1定制南十字星大学毕业证(SCU毕业证)#文凭成绩单#真实留信学历认证永久存档原版1:1定制南十字星大学毕业证(SCU毕业证)#文凭成绩单#真实留信学历认证永久存档
原版1:1定制南十字星大学毕业证(SCU毕业证)#文凭成绩单#真实留信学历认证永久存档
 
Thiophen Mechanism khhjjjjjjjhhhhhhhhhhh
Thiophen Mechanism khhjjjjjjjhhhhhhhhhhhThiophen Mechanism khhjjjjjjjhhhhhhhhhhh
Thiophen Mechanism khhjjjjjjjhhhhhhhhhhh
 
办理(UWIC毕业证书)英国卡迪夫城市大学毕业证成绩单原版一比一
办理(UWIC毕业证书)英国卡迪夫城市大学毕业证成绩单原版一比一办理(UWIC毕业证书)英国卡迪夫城市大学毕业证成绩单原版一比一
办理(UWIC毕业证书)英国卡迪夫城市大学毕业证成绩单原版一比一
 
INTERNSHIP ON PURBASHA COMPOSITE TEX LTD
INTERNSHIP ON PURBASHA COMPOSITE TEX LTDINTERNSHIP ON PURBASHA COMPOSITE TEX LTD
INTERNSHIP ON PURBASHA COMPOSITE TEX LTD
 
MK KOMUNIKASI DATA (TI)komdat komdat.docx
MK KOMUNIKASI DATA (TI)komdat komdat.docxMK KOMUNIKASI DATA (TI)komdat komdat.docx
MK KOMUNIKASI DATA (TI)komdat komdat.docx
 

Plug & Play Generative Networks: Conditional Iterative Generation of Images in Latent Space (UPC Reading Group)

  • 1. Plug & Play Generative Networks: Conditional Iterative Generation of Images in Latent Space Anh Nguyen, Jason Yosinski, Yoshua Bengio, Alexey Dosovitskiy, Jeff Clune [GitHub] [Arxiv] Slides by Víctor Garcia UPC Computer Vision Reading Group (27/01/2017)
  • 2. Index ● Introduction ● Probabilistic Interpretation of the method ● Methods and Experiments ○ PPGN-x: DAE model of p(x) ○ DGN-AM: sampling without a learned prior ○ PPGN-h: Generator and DAE model of p(h) ○ Joint PPGN-h: joint Generator and DAE ● Further Experiments ○ Image Generation: Captioning ○ Image Generation: Multifaceted Feature Visualization ○ Image inpainting ● Conclusions
  • 3. Introduction Interpretation of different frameworks to generate images maximizing: p(x, y) = p(x)*p(y|x) Prior Condition Encourages to look realistic Encourages to look from a particular class
  • 4. Introduction Image Generation: ● High Resolution Images (227x227) GANs struggle to Generate >64x64 Images
  • 5. Introduction Image Generation: ● High Resolution Images ● Intra-Class Variance
  • 6. Introduction Image Generation: ● High Resolution Images ● Intra-Class Variance ● Inter-Class Variance (1000-ImageNet classes)
  • 7. Index ● Introduction ● Probabilistic Interpretation of the method ● Methods and Experiments ○ PPGN-x: DAE model of p(x) ○ DGN-AM: sampling without a learned prior ○ PPGN-h: Generator and DAE model of p(h) ○ Joint PPGN-h: joint Generator and DAE ● Further Experiments ○ Image Generation: Captioning ○ Image Generation: Multifaceted Feature Visualization ○ Image inpainting ● Conclusions
  • 8. Probabilistic Interpretation of the method Metropolis-adjusted Langevin algorithm (MALA) which is a MCMC algorithm for iteratively producing random samples from a distribution p(x):
  • 9. Probabilistic Interpretation of the method Metropolis-adjusted Langevin algorithm (MALA) which is a MCMC algorithm for iteratively producing random samples: Current state
  • 10. Probabilistic Interpretation of the method Metropolis-adjusted Langevin algorithm (MALA) which is a MCMC algorithm for iteratively producing random samples: Future State Current state
  • 11. Probabilistic Interpretation of the method Metropolis-adjusted Langevin algorithm (MALA) which is a MCMC algorithm for iteratively producing random samples: Future State Current state Gradient to the natural manifold of p(x)
  • 12. Probabilistic Interpretation of the method Metropolis-adjusted Langevin algorithm (MALA) which is a MCMC algorithm for iteratively producing random samples: Gradient to the natural manifold of p(x) NoiseFuture State Current state
  • 13. Probabilistic Interpretation of the method Future State Current state Gradient to the natural manifold of p(x) Noise
  • 15. Probabilistic Interpretation of the method p(x) Step towards an image that causes the classifier to produce a higher score for class C Step towards a more generic image Noise
  • 16. Probabilistic Interpretation of the method xt Rough example
  • 17. Probabilistic Interpretation of the method y_co = Content activations y_st = Style activations Rough example
  • 18. Probabilistic Interpretation of the method xt+i Rough example
  • 19. Index ● Introduction ● Probabilistic Interpretation of the method ● Methods and Experiments ○ PPGN-x: DAE model of p(x) ○ DGN-AM: sampling without a learned prior ○ PPGN-h: Generator and DAE model of p(h) ○ Joint PPGN-h: joint Generator and DAE ● Further Experiments ○ Image Generation: Captioning ○ Image Generation: Multifaceted Feature Visualization ○ Image inpainting ● Conclusions
  • 21. Index ● Introduction ● Probabilistic Interpretation of the method ● Methods and Experiments ○ PPGN-x: DAE model of p(x) ○ DGN-AM: sampling without a learned prior ○ PPGN-h: Generator and DAE model of p(h) ○ Joint PPGN-h: joint Generator and DAE ● Further Experiments ○ Image Generation: Captioning ○ Image Generation: Multifaceted Feature Visualization ○ Image inpainting ● Conclusions
  • 22. Method | PPGN-x: DAE model of p(x) What a Denoising Autoencoder is? x h(x) R(x)
  • 23. Method | PPGN-x: DAE model of p(x) What a Denoising Autoencoder is? x_noise h(x) x N(0,σ^2) R(x)
  • 24. Method | PPGN-x: DAE model of p(x) What a Denoising Autoencoder is? x_noise h(x) x N(0,σ^2) R(x)
  • 25. Method | PPGN-x: DAE model of p(x)
  • 26. Method | PPGN-x: DAE model of p(x)
  • 27. Method | PPGN-x: DAE model of p(x) 1) Poorly modeled data, blurry 2) Slow changes
  • 28. Index ● Introduction ● Probabilistic Interpretation of the method ● Methods and Experiments ○ PPGN-x: DAE model of p(x) ○ DGN-AM: sampling without a learned prior ○ PPGN-h: Generator and DAE model of p(h) ○ Joint PPGN-h: joint Generator and DAE ● Further Experiments ○ Image Generation: Captioning ○ Image Generation: Multifaceted Feature Visualization ○ Image inpainting ● Conclusions
  • 29. Method | DGN-AM: sampling without a learned prior Deep Generator Network-based Activation Maximization It is faster if we move over h subspace instead of the x fc6 AlexNet
  • 30. Method | DGN-AM: sampling without a learned prior Deep Generator Network-based Activation Maximization Discriminator 1/0 AlexNet fc6
  • 31. Method | DGN-AM: sampling without a learned prior Once we trained the network G we find the equation for the MALA algorithm
  • 32. Method | DGN-AM: sampling without a learned prior Once we trained the network G we find the equation for the MALA algorithm
  • 33. Method | DGN-AM: sampling without a learned prior Once we trained the network G we find the equation for the MALA algorithm
  • 34. Method | DGN-AM: sampling without a learned prior Once we trained the network G we find the equation for the MALA algorithm No learned prior No noise
  • 35. Method | DGN-AM: sampling without a learned prior + Different modes from different starts - Same image after many steps - Low mixing speed
  • 36. Index ● Introduction ● Probabilistic Interpretation of the method ● Methods and Experiments ○ PPGN-x: DAE model of p(x) ○ DGN-AM: sampling without a learned prior ○ PPGN-h: Generator and DAE model of p(h) ○ Joint PPGN-h: joint Generator and DAE ● Further Experiments ○ Image Generation: Captioning ○ Image Generation: Multifaceted Feature Visualization ○ Image inpainting ● Conclusions
  • 37. Method | PPGN-h: Generator and DAE model of p(h) A 7 layers DAE is added to model the prior p(h) in order to increase the mixing speed
  • 38. Method | PPGN-h: Generator and DAE model of p(h) The equation is the following: Prior p(h) Conditioned Gradient Noise
  • 39. Method | PPGN-h: Generator and DAE model of p(h) - Similar to the last case. Low diversity - p(h) model learned by DAE is too simple
  • 40. Index ● Introduction ● Probabilistic Interpretation of the method ● Methods and Experiments ○ PPGN-x: DAE model of p(x) ○ DGN-AM: sampling without a learned prior ○ PPGN-h: Generator and DAE model of p(h) ○ Joint PPGN-h: joint Generator and DAE ● Further Experiments ○ Image Generation: Captioning ○ Image Generation: Multifaceted Feature Visualization ○ Image inpainting ● Conclusions
  • 41. Method | Joint PPGN-h: joint Generator and DAE In order to model p(h) in a more complex way DAE: h/fc6 → ? → h/fc6
  • 42. Method | Joint PPGN-h: joint Generator and DAE In order to model p(h) in a more complex way DAE: h/fc6 → ? → h/fc6 Joint Generator and DAE: h/fc6 x h/fc6 G E
  • 43. Method | Joint PPGN-h: joint Generator and DAE In order to model p(h) in a more complex way DAE: h/fc6 → ? → h/fc6 Joint Generator and DAE: h/fc6 x h/fc6 G E With the same existing network we train the Generator G to act as a DAE in conjunction with the E network
  • 44. Method | Joint PPGN-h: joint Generator and DAE AlexNet Equation is the same than before
  • 45. Method | Joint PPGN-h: joint Generator and DAE - Faster mixing - Better quality
  • 46. Method | Joint PPGN-h: joint Generator and AE AlexNet Equation is the same than before
  • 47. Method | Joint PPGN-h: joint Generator and AE - Faster mixing - Better quality
  • 48. Method | Joint PPGN-h: joint Generator and DAE Noise sweeps For the last model we test the reconstruction of different h/fc6 vectors when adding different noise levels: fc6 N(0, ) +
  • 49. Method | Joint PPGN-h: joint Generator and AE Noise sweeps For the last model we test the reconstruction of different h/fc6 vectors when adding different noise levels:
  • 50. Method | Joint PPGN-h: joint Generator and AE Noise sweeps
  • 51. Method | Joint PPGN-h: joint Generator and AE Noise sweeps We can still recover large information from the image when mapping with a lot of noise. Many → one.
  • 52. Method | Joint PPGN-h: joint Generator and DAE Combination of Losses Comparison of Losses: ● Real Images ● ● ● ●
  • 53. Method | Joint PPGN-h: joint Generator and DAE Combination of Losses
  • 54. Method | Joint PPGN-h: joint Generator and DAE Combination of Losses
  • 55. Method | Joint PPGN-h: joint Generator and DAE Evaluating: Qualitatively
  • 56. Method | Joint PPGN-h: joint Generator and DAE Evaluating: Qualitatively
  • 57. Method | Joint PPGN-h: joint Generator and DAE Evaluating: Qualitatively
  • 58. Index ● Introduction ● Probabilistic Interpretation of the method ● Methods and Experiments ○ PPGN-x: DAE model of p(x) ○ DGN-AM: sampling without a learned prior ○ PPGN-h: Generator and DAE model of p(h) ○ Joint PPGN-h: joint Generator and DAE ● Further Experiments ○ Image Generation: Captioning ○ Image Generation: Multifaceted Feature Visualization ○ Image inpainting ● Conclusions
  • 59. Further Experiments | Captioning MS-COCO Dataset
  • 60. Further Experiments | Captioning
  • 61. Index ● Introduction ● Probabilistic Interpretation of the method ● Methods and Experiments ○ PPGN-x: DAE model of p(x) ○ DGN-AM: sampling without a learned prior ○ PPGN-h: Generator and DAE model of p(h) ○ Joint PPGN-h: joint Generator and DAE ● Further Experiments ○ Image Generation: Captioning ○ Image Generation: Multifaceted Feature Visualization ○ Image inpainting ● Conclusions
  • 62. Further Experiments | MFV Multifaceted Feature Visualization
  • 64. Index ● Introduction ● Probabilistic Interpretation of the method ● Methods and Experiments ○ PPGN-x: DAE model of p(x) ○ DGN-AM: sampling without a learned prior ○ PPGN-h: Generator and DAE model of p(h) ○ Joint PPGN-h: joint Generator and DAE ● Further Experiments ○ Image Generation: Captioning ○ Image Generation: Multifaceted Feature Visualization ○ Image inpainting ● Conclusions
  • 65. Further Experiments | Inpainting Multifaceted Feature Visualization
  • 66. Further Experiments | Inpainting Multifaceted Feature Visualization
  • 67. Further Experiments | Inpainting Multifaceted Feature Visualization
  • 68. Further Experiments | Inpainting Multifaceted Feature Visualization
  • 69. Further Experiments | Inpainting Multifaceted Feature Visualization
  • 70. Conclusions ● Only using GANs for the reconstruction, GANs collapse into fewer modes, far from the original p(x). ● Using extra Losses it is possible to better reconstruct the images even for 1000 classes and for higher resolution. Mapping one-to-one helps to prevent typical latent → missing modes. ● It would be great to generate also the embedding space for this super-resolution multi-class images instead of using a supervised learned space.