|
124 | 124 | None, |
125 | 125 | 'code-examples-using-keras'), |
126 | 126 | ('Code in PyTorch for VAEs', 2, None, 'code-in-pytorch-for-vaes'), |
| 127 | + ('Pros and cons about VAEs', 2, None, 'pros-and-cons-about-vaes'), |
| 128 | + ('Pros and cons about VAEs', 2, None, 'pros-and-cons-about-vaes'), |
127 | 129 | ('Diffusion models, basics', 2, None, 'diffusion-models-basics'), |
128 | 130 | ('Problems with probabilistic models', |
129 | 131 | 2, |
|
252 | 254 | <!-- navigation toc: --> <li><a href="#computing-the-gradients" style="font-size: 80%;">Computing the gradients</a></li> |
253 | 255 | <!-- navigation toc: --> <li><a href="#code-examples-using-keras" style="font-size: 80%;">Code examples using Keras</a></li> |
254 | 256 | <!-- navigation toc: --> <li><a href="#code-in-pytorch-for-vaes" style="font-size: 80%;">Code in PyTorch for VAEs</a></li> |
| 257 | + <!-- navigation toc: --> <li><a href="#pros-and-cons-about-vaes" style="font-size: 80%;">Pros and cons about VAEs</a></li> |
| 258 | + <!-- navigation toc: --> <li><a href="#pros-and-cons-about-vaes" style="font-size: 80%;">Pros and cons about VAEs</a></li> |
255 | 259 | <!-- navigation toc: --> <li><a href="#diffusion-models-basics" style="font-size: 80%;">Diffusion models, basics</a></li> |
256 | 260 | <!-- navigation toc: --> <li><a href="#problems-with-probabilistic-models" style="font-size: 80%;">Problems with probabilistic models</a></li> |
257 | 261 | <!-- navigation toc: --> <li><a href="#diffusion-models" style="font-size: 80%;">Diffusion models</a></li> |
@@ -1293,6 +1297,26 @@ <h2 id="code-in-pytorch-for-vaes" class="anchor">Code in PyTorch for VAEs </h2> |
1293 | 1297 | </div> |
1294 | 1298 |
|
1295 | 1299 |
|
| 1300 | +<!-- !split --> |
| 1301 | +<h2 id="pros-and-cons-about-vaes" class="anchor">Pros and cons about VAEs </h2> |
| 1302 | + |
| 1303 | +<ol> |
| 1304 | +<li> Generative Capability: VAEs are powerful generative models that can produce new data similar to the training data.</li> |
| 1305 | +<li> Latent Space Structure: The latent space learned by VAEs is continuous, structured, and often interpretable, which makes it useful for interpolation, clustering, and downstream tasks.</li> |
| 1306 | +<li> Theoretical Foundation: VAEs are grounded in probability theory and Bayesian inference, offering a principled approach to modeling uncertainty.</li> |
| 1307 | +<li> Efficient Training: Unlike GANs (discussed later), VAEs typically train more stably and don’t suffer from mode collapse.</li> |
| 1308 | +<li> Regularization: The KL divergence term in the loss encourages the latent space to follow a standard normal distribution, aiding generalization. |
| 1309 | +<!-- o Semi-supervised Learning: VAEs can be adapted for semi-supervised tasks due to their probabilistic framework. --></li> |
| 1310 | +</ol> |
| 1311 | +<!-- !split --> |
| 1312 | +<h2 id="pros-and-cons-about-vaes" class="anchor">Pros and cons about VAEs </h2> |
| 1313 | +<ol> |
| 1314 | +<li> Blurry Reconstructions: VAEs often produce blurrier outputs compared to GANs due to the use of a probabilistic decoder and the Gaussian likelihood assumption.</li> |
| 1315 | +<li> Posterior Collapse: The model might ignore the latent variables entirely (especially with powerful decoders), leading to <b>posterior collapse</b>.</li> |
| 1316 | +<li> Trade-off in Loss Function: Balancing the reconstruction term and the KL divergence can be tricky, and tuning this balance is crucial.</li> |
| 1317 | +<li> Assumed Distribution Limitations: The prior (usually Gaussian) and the approximated posterior may not capture the true data distribution well.</li> |
| 1318 | +<li> Less Sharp Compared to GANs: In image generation tasks, the samples are usually less sharp and detailed than those generated by GANs.</li> |
| 1319 | +</ol> |
1296 | 1320 | <!-- !split --> |
1297 | 1321 | <h2 id="diffusion-models-basics" class="anchor">Diffusion models, basics </h2> |
1298 | 1322 |
|
|
0 commit comments