Want to create interactive content? It’s easy in Genially!
ml -2 cw -2
Sonali Nandagopalan
Created on April 30, 2023
Start designing with a free template
Discover more than 1500 professional designs like these:
Transcript
Analyzing and Improving the Image Quality of StyleGAN
Authors:Tero Karras, NVIDIA Samuli Laine, NVIDIA Miika Aittala, NVIDIA Janne Hellsten, NVIDIA Jaakko Lehtinen, NVIDIA and Aalto UniversityTimo Aila, NVIDIA
2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)
Index
Aim
Image quality and generator smoothness
Conclusion
GAN
Phase Artifact
References
Resolution Usage
StyleGAN
Droplet Artifact
Project of Latent Space to Images
Aim
"Fixing characteristic artifacts in StyleGAN and improving the quality further."
01
Generative Adversial Networks
70%
02
StyleGAN (2019)
- Data driven unconditional generative modeling
- Unconventional generator architecture
- Incorporates stochastic variation
- Generates high-resolution images
03
"But the SOTA has visible characteristic artifacts"
Distortion
Bluriness
Visual Anomalies
04
Issue - 1
Droplet Artifact
70%
+ info
Discriminator should have been able to detect the "Droplet Effect"
05
Problem Description
Problem: AdaIN Operation
Normalises Mean and Variance
Destroying information in the magnitude of features
Why?
Generator sneaks signal strength information
Hypothesis: Removal of Normalisation will remove the droplet artifact
06
Solution
Removing normalisation impacts scale-specific controls
07
Issue - 1
Droplet Artifact
Removes droplet effect and retains full controllability
70%
Remove the effect of s: Aim of Instance Normalisation
+ info
11
StyleGAN2 Improvement
Weight demodulation
11
Image quality and generator smoothness
PPL is inversely proportional to the Image Quality
12
Issue - 2
What conclusions can we draw from the relationship?
70%
- Generator stretches the region of latent space to penalise broken images and improve the quality of images.
- We cannot encourage low PPL as it would return a zero recall.
13
The Fix
Path Length Regularisation
Adapts the concept of "lazy regulariser". Uses the concepts of Jacobian matrix and Euclidean Norm to provide more effective generator models
Regularisation terms are computed less frequently. that the main loss function. Lowers computational cost
Lazy Regularisation
14
Issue - 3
Phase Artifact
70%
+ info
15
What is the potential cause?
Progressive Growing
- Resolution of images are gradually increased at each step.
- At each step, high frequency details are generated for the current resolution, which helps capture fine grain details.
- This leads to the trained network having high frequency details.
- The network becomes sensitive to small changes.
- This compromises network shift invariance, which is the network's ability to identify shift in location or position.
Rises the need for change in Arhcitecture!
16
Comparison of Architectures
Progressive Growing
Using "skip" generator and a "residual" discriminator without progressive growing
17
StyleGAN2 Improvement
Without Progressive Growing
18
Resolution Usage
Start of the training the learning is similar to progressive growing. But at the end of training high resolution fails to dominate.
Solution: Doubling the number of feature maps.
Generator has to focus on "low-resolution" features and then slowly start focussing on finer details (key aspect of progressive growing).
The cause of the problem: Capacity problem
Works in line with "progressive growing" and the hypothesised "expectation"
The architectures discussed allow it
INFO
19
Projection of Images to Latent Space
- Use ramped-down noise during optimization to explore the latent space thoroughly.
- Optimize the stochastic noise inputs of StyleGAN generator.
- Regulate the stochastic noise inputs to avoid carrying coherent signals.
Detection of the generated image back to the source is important
Conclusions
- StyleGAN2 addressed image quality issues in StyleGAN and improved image quality in various datasets, especially in motion.
- It's easier to attribute a generated image to its source using StyleGAN2.
- Training performance was improved, with faster training time and energy usage comparable to the original StyleGAN.
- The project consumed a significant amount of electricity and computation resources.
- Future improvements in path length regularization could be explored, and finding ways to reduce training data requirements is important for practical deployment of GANs, especially with datasets with intrinsic variation.
21
References
- Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., & Aila, T. (2019), Analyzing and improving the image quality of StyleGAN, in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 8110-8119).
- Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., & Aila, T. (2020). Supplemental Material: Analyzing and Improving the Image Quality of StyleGAN. NVIDIA
- https://github.com/NVlabs/stylegan2
70%
" Thank you!"