FastDiff / README.md
Rongjiehuang's picture
Create README.md
3fbe667 verified
metadata
license: other
tags:
  - text-to-speech
  - neural-vocoder
  - diffusion probabilistic model
inference: false
datasets:
  - LJSpeech
extra_gated_prompt: >-
  One more step before getting this model.

  This model is open access and available to all, with a license further
  specifying rights and usage.

  Any organization or individual is prohibited from using any technology
  mentioned in this paper to generate someone's speech without his/her consent,
  including but not limited to government leaders, political figures, and
  celebrities. If you do not comply with this item, you could be in violation of
  copyright laws.



  By clicking on "Access repository" below, you accept that your *contact
  information* (email address and username) can be shared with the model authors
  as well.
    
extra_gated_fields:
  I have read the License and agree with its terms: checkbox

FastDiff Model Card

Model Details

  • Model type: Diffusion-based text-to-speech generation model

  • Language(s): English

  • Model Description: A conditional diffusion probabilistic model capable of generating high fidelity speech efficiently.

  • Resources for more information: FastDiff GitHub Repository, FastDiff Paper.

  • Cite as:

    @inproceedings{huang2022fastdiff,
       title={FastDiff: A Fast Conditional Diffusion Model for High-Quality Speech Synthesis},
       author={Huang, Rongjie and Lam, Max WY and Wang, Jun and Su, Dan and Yu, Dong and Ren, Yi and Zhao, Zhou},
       booktitle = {Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, {IJCAI-22}},
       year={2022}
    

This model card was written based on the DALL-E Mini model card.