brendenc commited on
Commit
162a915
·
1 Parent(s): 8c1fc94

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -0
README.md CHANGED
@@ -14,6 +14,8 @@ Learning latent space representations of data remains to be an important task in
14
 
15
  To learn discrete latent representations, ideas from vector quantisation (VQ) are used. Using the VQ method allows the model to avoid issues of ["posterior collapse"](https://datascience.stackexchange.com/questions/48962/what-is-posterior-collapse-phenomenon). By pairing these representations with an autoregressive prior, VQ-VAE models can generate high quality images, videos, speech as well as doing high quality speaker conversion and unsupervised learning of phonemes, providing further evidence of the utility of the learnt representations.
16
 
 
 
17
  ### Further learning
18
  This model has been trained using code from this [example](https://keras.io/examples/generative/vq_vae/), and a result of this [paper.](https://arxiv.org/pdf/1711.00937.pdf)
19
 
 
14
 
15
  To learn discrete latent representations, ideas from vector quantisation (VQ) are used. Using the VQ method allows the model to avoid issues of ["posterior collapse"](https://datascience.stackexchange.com/questions/48962/what-is-posterior-collapse-phenomenon). By pairing these representations with an autoregressive prior, VQ-VAE models can generate high quality images, videos, speech as well as doing high quality speaker conversion and unsupervised learning of phonemes, providing further evidence of the utility of the learnt representations.
16
 
17
+ Full Credits for this example go to [Sayak Paul](https://twitter.com/RisingSayak)
18
+
19
  ### Further learning
20
  This model has been trained using code from this [example](https://keras.io/examples/generative/vq_vae/), and a result of this [paper.](https://arxiv.org/pdf/1711.00937.pdf)
21