speyside42
speyside42 t1_jb425d4 wrote
Reply to comment by PassionatePossum in [R] [N] Dropout Reduces Underfitting - Liu et al. by radi-cho
A good mixture is key. Independent applied research will show whether the claims of slight improvements hold in general. A counter example where "this kind of research" has failed us are novel optimizers.
speyside42 t1_jc44rbn wrote
Reply to comment by currentscurrents in [D]: Generalisation ability of autoencoders by Blutorangensaft
> Vanilla autoencoders don't generalize well, but variational autoencoders have a much better structured latent space and generalize much better.
For toy problems yes, but not generally. For a generalizing Image Autoencoder, check for example ConvNextv2: https://arxiv.org/pdf/2301.00808.pdf
As a side note: The VQ-VAE from the blog post has actually really little to do with variational inference. You have basically no prior at all (uniform over all discrete latents) therefore the KL-divergence term can also be dropped. It's basically just a glorified quantized Autoencoder that could be interpreted in the language of variational models.