1
2
3
4
Program
JDLS
12 mai
5
1
2
3
5
6
Introduction
➔ Reminder on generative model
➔ Diffusion Model vs VAE
Denoising Diffusion Probabilistic Models
➔ Principle
➔ Forward and Reverse Diffusion
➔ Training and Sampling
Example: Fashion MNIST
➔ Generation of Fashion MNIST
DDPM improvements
➔ Beta scheduling and Variance learning
➔ Fast sampling
➔ Latent diffusion
DDPM applications
➔ Text-to-image
➔ Other task : inpainting / outpainting / super-resolution
Rappel - VAE - 1
Source https://lilianweng.github.io/posts/2021-07-11-diffusion-models/
6
7
Rappel - VAE - 2
VAE COST FUNCTION
Sampling process
Training process
metric
Decoder
- Gaussian
- Mixture of gaussian
Rappel - GAN - 1
2 networks in opposition :
- Generator
- Discriminator
Ideal solution :
- Generator ~ P(x|z)
- Discriminator = ½
Adding supervision concept in an
unsupervised task !
8
Data to train a
classification
network
Source
https://www.kdnuggets.com/2017/01/generative-adversari
al-networks-hot-topic-machine-learning.html
9
Rappel - GAN - 2
Sampling process
Training process
Generator
Train
discriminator
Train
generator
GAN COST FUNCTION
- Uniform
Rappel - GAN - 3 GAN convergence problems !
Vanishing gradient due to
discriminant being too
perfect, generator can’t train
anymore
Mode collapse due to
generator learning only some
good examples instead of the
whole data distribution
10
True Data
Generated Data
Source
https://lilianweng.github.io/posts/2017-08-20-gan/
No convergence due to the
nature of the problem
(MinMax)
True data
Generated data
11
VAE vs GAN
VAE
GAN
- Generate high
quality data
- Hard to train
- Have more
diversity
How to compare generative models ?
VAE vs DPM
12
VAE
Low dimensional
representation of the input.
DPM
Fixed
encoder
High dimensional
representation of the input.
Z
Decoder
DPM - Landscape
13
Source : https://github.com/bentoml/stable-diffusion-bentoml
Dhariwal & Nichol, 2021
Source : Dall-E 2
DDPM - Principle - 1
14
After the training the Diffusion Model will generate images from Gaussian noise:
DDPM - Principle - 2
15
There are three processes that characterize Diffusion Models:
1. Forward Diffusion Process
2. Reverse Diffusion Process
3. Sampling Process
DDPM - Principle - 3
16
Forward Diffusion Process
This process will add noise to any image gradually
0 ≤ t ≤ T ; T is a hyperparameter
DDPM - Principle - 4
17
Forward Diffusion Process
Examples of images at different times t
Here we choose T=1000, but it can be different values (it’s an hyperparameter)
18
Reverse Diffusion Process
We train a model to predict xt-1
from xt
x0
is any image from the dataset
DDPM - Principle - 5
a bit less
noisy than xt
a bit more
noisy than xt-1
19
Reverse Diffusion Process
The same model must predict every xt-1
from xt
DDPM - Principle - 6
20
Sampling Process
From a random noise we can generate an image
DDPM - Principle - 7
DDPM - Forward Diffusion - 1
21
DDPM - Forward Diffusion - 2
22
DDPM - Forward Diffusion - 3
23
DDPM - Forward Diffusion - 4
24
DDPM - Forward Diffusion - 5
25
DDPM - Forward Diffusion - 6
26
So we can sample a noised image at any time step directly from original image
Question break #1
27
DDPM - Reverse Diffusion - 1
28
We predict only the mean, we know the rest.
DDPM - Reverse Diffusion - 2
29
We can predict xt-1
by predicting zt
DDPM - Reverse Diffusion - 3
30
A little bit of explanation (a tiny bit):
https://lilianweng.github.io/posts/2021-07-11-diffusion-models/
DDPM - Reverse Diffusion - 4
31
DDPM - Reverse Diffusion - 5
32
Question break #2
33
DDPM - Training - 0
34
https://arxiv.org/abs/2006.11239
35
Dataset
DDPM - Training - 1
x0
36
DDPM - Training - 3
Uniform
distribution
Between 1 and T
t = 50
x0
37
DDPM - Training - 4
Gaussian
distribution
Same shape than x0
zt
( = ϵ )
x0
t = 50
38
DDPM - Training - 5
x0
t = 50
zt
39
DDPM - Training - 5
xt
t = 50
zθ
( = ϵθ
)
x0
t = 50
zt
xt
40
DDPM - Training - 6
zθ
Loss
x0
t = 50
zt
zθ
zt
2
xt
41
DDPM - Training - 7
Loss
Backward
∇θ
Weight update
42
DDPM - Training - 8
and repeat !
It was just 1 iteration.
43
DDPM - Sampling - 0
https://arxiv.org/abs/2006.11239
44
DDPM - Sampling - 1
Gaussian
distribution
Same shape than training
dataset images
xT
45
DDPM - Sampling - 2
xT
T
zθ
(xT
,T) ≈ zT
Reminder:
46
DDPM - Sampling - 3
Gaussian
distribution
Same shape than training
dataset images
z (noise)
47
DDPM - Sampling - 4
48
DDPM - Sampling - 5
and repeat !
Don’t generate xT
, replace it by xT-1
and T by T-1… and do it again T time.
Question break #3
49
DDPM
50
DDPM vs VAE vs GAN
VAE
GAN
- Generate high
quality data
- Hard to train
- Have more
diversity
- Long sampling
process
51
DDPM improvements
Improving the log-likelihood
metrics (Improved DDPM, 2021)
Improving image synthesis
Dhariwal & Nichol, 2021
Faster sampling
(Denoising Diffusion Implicit Model,
2021 / Latent Diffusion Model, 2021)
Beta scheduling - Cosine scheduling (IDDPM, 2021)
52
0.02 0.0001
1000
Almost pure noise
Strong noising
Variance learning (IDDPM, 2021)
53
Limitations :
“(...) learning reverse
process variances
(...) leads to
unstable training and
poorer sample
quality compared to
fixed variance.”
(DDPM, 2020)
0.001
DDPM
IDDPM
the early stage of diffusion are very important
54
Diffusion & reverse process (DDIM, 2021)
Mr. Gaussian
Mr. Data
Limitations :
“For example, it
takes around 20
hours to sample
50k images of size
32 x 32 from a
DDPM, but less
than a minute to do
so from a GAN on a
Nvidia 2080 Ti
GPU.” (DDIM,
2021)
55
Extension of DDPM (DDIM, 2021)
Generalisation to a bigger class of
inverse process (non-Markovian)
Important :
Same network and
training as a DDPM
56
Generation process (DDIM, 2021)
DDPM
DDIM
can be computed !!!
57
Finding a better inverse process (DDIM, 2021)
=> DDIM
FID
58
Noise interpolation (DDIM, 2021)
+
Question break #4
59
Latent diffusion : Concept of latent space (reminder)
60
Latent space of MNIST database for AE and VAE
Source https://thilospinner.com/towards-an-interpretable-latent-space/
● Similar objects are close to one another in the latent space
● Usually lower dimension than original data (therefore does compression as well)
● Usually impossible to visualize by a human
Latent diffusion : Concept of latent space (example)
61
Example : Word embedding
Actor - Pierre Curie + Marie Curie ≈ Actress
Turn sparse data (for instance words) into
vectors
DPM
62
+
n
o
i
s
e
+
d
i
f
f
u
s
i
o
n
Latent diffusion model
63
Latent space
Encoder Decoder
Image space
+
n
o
i
s
e
+
d
i
f
f
u
s
i
o
n
Conditional diffusion
A picture of GENCI’s supercomputer Jean Zay. On the storage bays, a picture
of the eponymous minister, with a background representing a simulation of a
turbulent flow of liquid sodium, and a quote from Jean Zay’s memoirs.
Alongside the bays, the cooling equipment with the logo of the manufacturer
and the owner of the supercomputer.
How to control the output of a diffusion model
and make sure it generates what we want ?
64
Conditional diffusion : text → image
A picture of GENCI’s supercomputer Jean Zay. On
the storage bays, a picture of the eponymous
minister, with a background representing a
simulation of a turbulent flow of liquid sodium, and
a quote from Jean Zay’s memoirs. Alongside the
bays, the cooling equipment with the logo of the
manufacturer and the owner of the
supercomputer.
Latent space
65
Conditional diffusion: text → image
66
Conditional diffusion : cross-attention
Stable Diffusion uses cross-attention to make the denoising process consistent with the
provided sentence embedding
67
Source Rombach, Robin, et al. "High-resolution image synthesis with latent diffusion models."
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022.
Conditional diffusion : other method
Spatial Self
Attention
Dense layer
Standard Unet layer
(conv, maxpool,
upsample, …)
68
➔ Inpainting
➔ Super-resolution
➔ Outpainting
Other tasks
Diffusion models can solve a variety of tasks. We already know about image generation, as well as
conditional image generation (for instance with a short paragraph describing the picture)
Other tasks:
69
Source Lugmayr, Andreas, et al. "Repaint: Inpainting using denoising diffusion probabilistic models." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022.
➔ Thin mask
➔ Right side mask
for halving the
image
➔ Every second row
of pixels for
alternating lines
Inpainting through masking
➔ Wide mask
➔ Outer mask for
expanding the image
➔ Every second pixel in
both directions for
super-resolution
We can solve many of these tasks through the usage of a mask
70
Source Lugmayr, Andreas, et al. "Repaint: Inpainting using denoising diffusion probabilistic models." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022.
Inpainting through masking
71
Step t :
Source Lugmayr, Andreas, et al. "Repaint: Inpainting using denoising diffusion probabilistic models." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022.
Inpainting through masking: step t
72
+
n
o
i
s
e
+
d
i
f
f
u
s
i
o
n
New artifacts added (in this coarse example, our diffusion
model drew a sun), so we force the known background again!
xt
x0
xt-1
Inpainting through masking: step t
73
+
n
o
i
s
e
+
d
i
f
f
u
s
i
o
n
xt
x0
xt-1
This operation does not take into
account the generated information
Inpainting deharmonization
74
Picture deharmonization: the generated image has a satisfying texture but is wrong semantically.
The suggested solution is to resample.
Source Lugmayr, Andreas, et al. "Repaint: Inpainting using denoising diffusion probabilistic models." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022.
Inpainting resampling
75
xt
x0 noising
diffusion
× mask
× (1 - mask)
+ xt-1
resampling
This loop is performed several times (a hyperparameter) before moving on the next step
Inpainting resampling
76
n is the number of times the resampling loop was performed
Disadvantage: the number of required denoising steps is much higher
Advantage: it produces much more satisfying results
Lugmayr, Andreas, et al. "Repaint: Inpainting using denoising diffusion probabilistic models." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022.
77
Sources
Papers:
- Deep Unsupervised Learning using Nonequilibrium Thermodynamics (DPM) (https://arxiv.org/abs/1503.03585)
- Denoising Diffusion Probabilistic Models (DDPM) (https://arxiv.org/abs/2006.11239)
- Improved Denoising Diffusion Probabilistic Models (IDDPM) (https://arxiv.org/abs/2102.09672)
- Denoising Diffusion Implicit Models (DDIM) (https://arxiv.org/abs/2010.02502)
- Diffusion Models Beat GANs on Image Synthesis (https://arxiv.org/abs/2105.05233)
- High-Resolution Image Synthesis with Latent Diffusion Models (LDM) (https://arxiv.org/abs/2112.10752)
- Repaint: Inpainting using denoising diffusion probabilistic models (https://arxiv.org/pdf/2201.09865)
- Diffusion Models in Vision: A Survey (https://arxiv.org/abs/2209.04747)
- Diffusion Models: A Comprehensive Survey of Methods and Applications(https://arxiv.org/abs/2209.00796)
Other ressources:
- Lilian Weng’s article (https://lilianweng.github.io/posts/2021-07-11-diffusion-models)
- Yang Song’s article (https://yang-song.net/blog/2021/score)
- Outlier video (https://www.youtube.com/watch?v=HoKDTa5jHvg)
Question break #5 & Practice
78
Épisode 15 :
AI, droit, société et éthique
● Interprétabilité, reproductibilité, biais
● Cadre légal
● Privacy
● Session interactive
Durée : 2h
Next, on Fidle: Jeudi 23 mars, 14h00
To be continued...
Next on Fidle :
Attribution-NonCommercial-NoDerivatives 4.0 International (CC BY-NC-ND 4.0)
https://creativecommons.org/licenses/by-nc-nd/4.0/
Séquence 15 :
AI, droit, société et éthique
Jeudi 23 mars,
https://fidle.cnrs.fr
Contact@fidle.cnrs.fr
https://fidle.cnrs.fr/youtube
Merci !

Introduction to Diffusion Models on deep learning

  • 1.
  • 2.
  • 3.
  • 4.
  • 5.
    5 1 2 3 5 6 Introduction ➔ Reminder ongenerative model ➔ Diffusion Model vs VAE Denoising Diffusion Probabilistic Models ➔ Principle ➔ Forward and Reverse Diffusion ➔ Training and Sampling Example: Fashion MNIST ➔ Generation of Fashion MNIST DDPM improvements ➔ Beta scheduling and Variance learning ➔ Fast sampling ➔ Latent diffusion DDPM applications ➔ Text-to-image ➔ Other task : inpainting / outpainting / super-resolution
  • 6.
    Rappel - VAE- 1 Source https://lilianweng.github.io/posts/2021-07-11-diffusion-models/ 6
  • 7.
    7 Rappel - VAE- 2 VAE COST FUNCTION Sampling process Training process metric Decoder - Gaussian - Mixture of gaussian
  • 8.
    Rappel - GAN- 1 2 networks in opposition : - Generator - Discriminator Ideal solution : - Generator ~ P(x|z) - Discriminator = ½ Adding supervision concept in an unsupervised task ! 8 Data to train a classification network Source https://www.kdnuggets.com/2017/01/generative-adversari al-networks-hot-topic-machine-learning.html
  • 9.
    9 Rappel - GAN- 2 Sampling process Training process Generator Train discriminator Train generator GAN COST FUNCTION - Uniform
  • 10.
    Rappel - GAN- 3 GAN convergence problems ! Vanishing gradient due to discriminant being too perfect, generator can’t train anymore Mode collapse due to generator learning only some good examples instead of the whole data distribution 10 True Data Generated Data Source https://lilianweng.github.io/posts/2017-08-20-gan/ No convergence due to the nature of the problem (MinMax) True data Generated data
  • 11.
    11 VAE vs GAN VAE GAN -Generate high quality data - Hard to train - Have more diversity How to compare generative models ?
  • 12.
    VAE vs DPM 12 VAE Lowdimensional representation of the input. DPM Fixed encoder High dimensional representation of the input. Z Decoder
  • 13.
    DPM - Landscape 13 Source: https://github.com/bentoml/stable-diffusion-bentoml Dhariwal & Nichol, 2021 Source : Dall-E 2
  • 14.
    DDPM - Principle- 1 14 After the training the Diffusion Model will generate images from Gaussian noise:
  • 15.
    DDPM - Principle- 2 15 There are three processes that characterize Diffusion Models: 1. Forward Diffusion Process 2. Reverse Diffusion Process 3. Sampling Process
  • 16.
    DDPM - Principle- 3 16 Forward Diffusion Process This process will add noise to any image gradually 0 ≤ t ≤ T ; T is a hyperparameter
  • 17.
    DDPM - Principle- 4 17 Forward Diffusion Process Examples of images at different times t Here we choose T=1000, but it can be different values (it’s an hyperparameter)
  • 18.
    18 Reverse Diffusion Process Wetrain a model to predict xt-1 from xt x0 is any image from the dataset DDPM - Principle - 5 a bit less noisy than xt a bit more noisy than xt-1
  • 19.
    19 Reverse Diffusion Process Thesame model must predict every xt-1 from xt DDPM - Principle - 6
  • 20.
    20 Sampling Process From arandom noise we can generate an image DDPM - Principle - 7
  • 21.
    DDPM - ForwardDiffusion - 1 21
  • 22.
    DDPM - ForwardDiffusion - 2 22
  • 23.
    DDPM - ForwardDiffusion - 3 23
  • 24.
    DDPM - ForwardDiffusion - 4 24
  • 25.
    DDPM - ForwardDiffusion - 5 25
  • 26.
    DDPM - ForwardDiffusion - 6 26 So we can sample a noised image at any time step directly from original image
  • 27.
  • 28.
    DDPM - ReverseDiffusion - 1 28 We predict only the mean, we know the rest.
  • 29.
    DDPM - ReverseDiffusion - 2 29 We can predict xt-1 by predicting zt
  • 30.
    DDPM - ReverseDiffusion - 3 30 A little bit of explanation (a tiny bit): https://lilianweng.github.io/posts/2021-07-11-diffusion-models/
  • 31.
    DDPM - ReverseDiffusion - 4 31
  • 32.
    DDPM - ReverseDiffusion - 5 32
  • 33.
  • 34.
    DDPM - Training- 0 34 https://arxiv.org/abs/2006.11239
  • 35.
  • 36.
    36 DDPM - Training- 3 Uniform distribution Between 1 and T t = 50 x0
  • 37.
    37 DDPM - Training- 4 Gaussian distribution Same shape than x0 zt ( = ϵ ) x0 t = 50
  • 38.
    38 DDPM - Training- 5 x0 t = 50 zt
  • 39.
    39 DDPM - Training- 5 xt t = 50 zθ ( = ϵθ ) x0 t = 50 zt xt
  • 40.
    40 DDPM - Training- 6 zθ Loss x0 t = 50 zt zθ zt 2 xt
  • 41.
    41 DDPM - Training- 7 Loss Backward ∇θ Weight update
  • 42.
    42 DDPM - Training- 8 and repeat ! It was just 1 iteration.
  • 43.
    43 DDPM - Sampling- 0 https://arxiv.org/abs/2006.11239
  • 44.
    44 DDPM - Sampling- 1 Gaussian distribution Same shape than training dataset images xT
  • 45.
    45 DDPM - Sampling- 2 xT T zθ (xT ,T) ≈ zT Reminder:
  • 46.
    46 DDPM - Sampling- 3 Gaussian distribution Same shape than training dataset images z (noise)
  • 47.
  • 48.
    48 DDPM - Sampling- 5 and repeat ! Don’t generate xT , replace it by xT-1 and T by T-1… and do it again T time.
  • 49.
  • 50.
    DDPM 50 DDPM vs VAEvs GAN VAE GAN - Generate high quality data - Hard to train - Have more diversity - Long sampling process
  • 51.
    51 DDPM improvements Improving thelog-likelihood metrics (Improved DDPM, 2021) Improving image synthesis Dhariwal & Nichol, 2021 Faster sampling (Denoising Diffusion Implicit Model, 2021 / Latent Diffusion Model, 2021)
  • 52.
    Beta scheduling -Cosine scheduling (IDDPM, 2021) 52 0.02 0.0001 1000 Almost pure noise Strong noising
  • 53.
    Variance learning (IDDPM,2021) 53 Limitations : “(...) learning reverse process variances (...) leads to unstable training and poorer sample quality compared to fixed variance.” (DDPM, 2020) 0.001 DDPM IDDPM the early stage of diffusion are very important
  • 54.
    54 Diffusion & reverseprocess (DDIM, 2021) Mr. Gaussian Mr. Data Limitations : “For example, it takes around 20 hours to sample 50k images of size 32 x 32 from a DDPM, but less than a minute to do so from a GAN on a Nvidia 2080 Ti GPU.” (DDIM, 2021)
  • 55.
    55 Extension of DDPM(DDIM, 2021) Generalisation to a bigger class of inverse process (non-Markovian) Important : Same network and training as a DDPM
  • 56.
    56 Generation process (DDIM,2021) DDPM DDIM can be computed !!!
  • 57.
    57 Finding a betterinverse process (DDIM, 2021) => DDIM FID
  • 58.
  • 59.
  • 60.
    Latent diffusion :Concept of latent space (reminder) 60 Latent space of MNIST database for AE and VAE Source https://thilospinner.com/towards-an-interpretable-latent-space/ ● Similar objects are close to one another in the latent space ● Usually lower dimension than original data (therefore does compression as well) ● Usually impossible to visualize by a human
  • 61.
    Latent diffusion :Concept of latent space (example) 61 Example : Word embedding Actor - Pierre Curie + Marie Curie ≈ Actress Turn sparse data (for instance words) into vectors
  • 62.
  • 63.
    Latent diffusion model 63 Latentspace Encoder Decoder Image space + n o i s e + d i f f u s i o n
  • 64.
    Conditional diffusion A pictureof GENCI’s supercomputer Jean Zay. On the storage bays, a picture of the eponymous minister, with a background representing a simulation of a turbulent flow of liquid sodium, and a quote from Jean Zay’s memoirs. Alongside the bays, the cooling equipment with the logo of the manufacturer and the owner of the supercomputer. How to control the output of a diffusion model and make sure it generates what we want ? 64
  • 65.
    Conditional diffusion :text → image A picture of GENCI’s supercomputer Jean Zay. On the storage bays, a picture of the eponymous minister, with a background representing a simulation of a turbulent flow of liquid sodium, and a quote from Jean Zay’s memoirs. Alongside the bays, the cooling equipment with the logo of the manufacturer and the owner of the supercomputer. Latent space 65
  • 66.
  • 67.
    Conditional diffusion :cross-attention Stable Diffusion uses cross-attention to make the denoising process consistent with the provided sentence embedding 67 Source Rombach, Robin, et al. "High-resolution image synthesis with latent diffusion models." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022.
  • 68.
    Conditional diffusion :other method Spatial Self Attention Dense layer Standard Unet layer (conv, maxpool, upsample, …) 68
  • 69.
    ➔ Inpainting ➔ Super-resolution ➔Outpainting Other tasks Diffusion models can solve a variety of tasks. We already know about image generation, as well as conditional image generation (for instance with a short paragraph describing the picture) Other tasks: 69 Source Lugmayr, Andreas, et al. "Repaint: Inpainting using denoising diffusion probabilistic models." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022.
  • 70.
    ➔ Thin mask ➔Right side mask for halving the image ➔ Every second row of pixels for alternating lines Inpainting through masking ➔ Wide mask ➔ Outer mask for expanding the image ➔ Every second pixel in both directions for super-resolution We can solve many of these tasks through the usage of a mask 70 Source Lugmayr, Andreas, et al. "Repaint: Inpainting using denoising diffusion probabilistic models." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022.
  • 71.
    Inpainting through masking 71 Stept : Source Lugmayr, Andreas, et al. "Repaint: Inpainting using denoising diffusion probabilistic models." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022.
  • 72.
    Inpainting through masking:step t 72 + n o i s e + d i f f u s i o n New artifacts added (in this coarse example, our diffusion model drew a sun), so we force the known background again! xt x0 xt-1
  • 73.
    Inpainting through masking:step t 73 + n o i s e + d i f f u s i o n xt x0 xt-1 This operation does not take into account the generated information
  • 74.
    Inpainting deharmonization 74 Picture deharmonization:the generated image has a satisfying texture but is wrong semantically. The suggested solution is to resample. Source Lugmayr, Andreas, et al. "Repaint: Inpainting using denoising diffusion probabilistic models." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022.
  • 75.
    Inpainting resampling 75 xt x0 noising diffusion ×mask × (1 - mask) + xt-1 resampling This loop is performed several times (a hyperparameter) before moving on the next step
  • 76.
    Inpainting resampling 76 n isthe number of times the resampling loop was performed Disadvantage: the number of required denoising steps is much higher Advantage: it produces much more satisfying results Lugmayr, Andreas, et al. "Repaint: Inpainting using denoising diffusion probabilistic models." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2022.
  • 77.
    77 Sources Papers: - Deep UnsupervisedLearning using Nonequilibrium Thermodynamics (DPM) (https://arxiv.org/abs/1503.03585) - Denoising Diffusion Probabilistic Models (DDPM) (https://arxiv.org/abs/2006.11239) - Improved Denoising Diffusion Probabilistic Models (IDDPM) (https://arxiv.org/abs/2102.09672) - Denoising Diffusion Implicit Models (DDIM) (https://arxiv.org/abs/2010.02502) - Diffusion Models Beat GANs on Image Synthesis (https://arxiv.org/abs/2105.05233) - High-Resolution Image Synthesis with Latent Diffusion Models (LDM) (https://arxiv.org/abs/2112.10752) - Repaint: Inpainting using denoising diffusion probabilistic models (https://arxiv.org/pdf/2201.09865) - Diffusion Models in Vision: A Survey (https://arxiv.org/abs/2209.04747) - Diffusion Models: A Comprehensive Survey of Methods and Applications(https://arxiv.org/abs/2209.00796) Other ressources: - Lilian Weng’s article (https://lilianweng.github.io/posts/2021-07-11-diffusion-models) - Yang Song’s article (https://yang-song.net/blog/2021/score) - Outlier video (https://www.youtube.com/watch?v=HoKDTa5jHvg)
  • 78.
    Question break #5& Practice 78
  • 79.
    Épisode 15 : AI,droit, société et éthique ● Interprétabilité, reproductibilité, biais ● Cadre légal ● Privacy ● Session interactive Durée : 2h Next, on Fidle: Jeudi 23 mars, 14h00
  • 80.
    To be continued... Nexton Fidle : Attribution-NonCommercial-NoDerivatives 4.0 International (CC BY-NC-ND 4.0) https://creativecommons.org/licenses/by-nc-nd/4.0/ Séquence 15 : AI, droit, société et éthique Jeudi 23 mars, https://fidle.cnrs.fr Contact@fidle.cnrs.fr https://fidle.cnrs.fr/youtube Merci !