Safetensors
ASLP-lab commited on
Commit
21b1b27
·
verified ·
1 Parent(s): 87e2128

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +108 -3
README.md CHANGED
@@ -1,3 +1,108 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <p align="center">
2
+ <img src="src/diffrhythm2_logo.png" width="400"/>
3
+ <p>
4
+
5
+ <p align="center">
6
+ <h1>Di♪♪Rhythm 2: Efficient And High Fidelity Song Generation Via Block Flow Matching</h1>
7
+ </p>
8
+
9
+ <div style='display:flex; gap: 0.25rem; '>
10
+ <a href='https://arxiv.org/pdf/2510.22950'><img src='https://img.shields.io/badge/Paper-PDF-red'></a>
11
+ <a href='https://aslp-lab.github.io/DiffRhythm2.github.io'><img src='https://img.shields.io/badge/Project-Demo-green'></a>
12
+ <a href='https://huggingface.co/ASLP-lab/DiffRhythm2'><img src='https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-Models-yellow'></a>
13
+ <a href='https://huggingface.co/spaces/ASLP-lab/DiffRhythm2'><img src='https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-Space-yellow'></a>
14
+ <a href='https://www.modelscope.cn/models/ASLPlab/DiffRhythm2'><img src='https://img.shields.io/badge/ModelScope-Models-6149ff'></a>
15
+ <a href="https://github.com/xiaomi-research/diffrhythm2/blob/main/LICENSE"><img alt="License" src="https://img.shields.io/badge/License-Apache%202.0-blue?&color=blue"/></a>
16
+ </div>
17
+
18
+ <a href="https://discord.gg/vUD4zgTpJa"><img src="https://img.shields.io/badge/Contact-Discord-5865f2"/></a>
19
+ <a href="https://github.com/ASLP-lab/DiffRhythm2/blob/main/src/contact.md"><img src="https://img.shields.io/badge/Contact-WeChat-17d56b"/></a>
20
+
21
+ Yuepeng Jiang, Huakang Chen, Ziqian Ning, Jixun Yao, Zerui Han, Di Wu, Meng Meng, Jian Luan, Zhonghua Fu, Lei Xie†
22
+
23
+ <!-- <p align="center">
24
+ <b>DiffRhythm 2</b> &nbsp;&nbsp;|&nbsp;&nbsp;
25
+ 📑 <a href="https://arxiv.org/abs/2503.01183">Paper</a> &nbsp;&nbsp;|&nbsp;&nbsp;
26
+ 🎵 <a href="https://aslp-lab.github.io/DiffRhythm.github.io/">Demo</a>
27
+ </p> -->
28
+
29
+ DiffRhythm 2 (Chinese: 谛韵, Dì Yùn) is the next-generation open-sourced music generation framework that advances the original DiffRhythm with a semi-autoregressive diffusion architecture. It is capable of generating full-length songs with precise lyric alignment and coherent musical structures. The name inherits the essence of DiffRhythm — “Diff” reflects its diffusion-based generative backbone, while “Rhythm” emphasizes its dedication to musicality and temporal flow. The Chinese name 谛韵 (Dì Yùn) continues this dual symbolism: “谛” (attentive listening) represents perceptual awareness, and “韵” (melodic charm) captures the expressive beauty of music.
30
+
31
+ <p align="center">
32
+ <img src="src/model2.png" width="80%"/>
33
+ <p>
34
+
35
+ <p align="center">
36
+ <img src="src/model1.png" width="80%"/>
37
+ <p>
38
+
39
+ ## Demo Video
40
+
41
+
42
+ https://github.com/user-attachments/assets/95bac874-82b2-4c92-950e-3489a9c03ab0
43
+
44
+
45
+ ## 📢 News and Updates
46
+
47
+ * **2025.10.30** 🚀 We released the [DiffRhythm2 paper](https://arxiv.org/pdf/2510.22950), demo code, and [model weights](https://huggingface.co/ASLP-lab/DiffRhythm2).
48
+
49
+ ## 📋 TODOs
50
+ - [ ] Support Colab.
51
+ - [ ] Gradio support.
52
+ - [ ] Song extension.
53
+ - [ ] Instrumental music generation.
54
+ - [x] Release code and weights.
55
+ - [x] Release paper to Arxiv.
56
+
57
+ ## 🔨 Inference
58
+
59
+ Following the steps below to clone the repository and install the environment.
60
+
61
+ ```bash
62
+ # clone and enter the repositry
63
+ git clone https://github.com/ASLP-lab/DiffRhythm2.git
64
+ cd DiffRhythm2
65
+
66
+ # install the environment
67
+ ## espeak-ng
68
+ # For Debian-like distribution (e.g. Ubuntu, Mint, etc.)
69
+ sudo apt-get install espeak-ng
70
+ # For RedHat-like distribution (e.g. CentOS, Fedora, etc.)
71
+ sudo yum install espeak-ng
72
+ # For MacOS
73
+ brew install espeak-ng
74
+ # For Windows
75
+ # Please visit https://github.com/espeak-ng/espeak-ng/releases to download .msi installer
76
+
77
+ ## install requirements
78
+ pip install -r requirements.txt
79
+ ```
80
+
81
+ On Linux you can now simply use the inference script:
82
+ ```bash
83
+ # For inference using a reference WAV file
84
+ bash inference.sh
85
+ ```
86
+
87
+ Weights will be automatically downloaded from Hugging Face upon the first run.
88
+
89
+ Example files of lyrics and reference audio can be found in `example`.
90
+
91
+
92
+ ## 📜 License & Disclaimer
93
+
94
+ DiffRhythm 2 (code and weights) is released under the [Apache License 2.0](https://www.apache.org/licenses/LICENSE-2.0). This open-source license allows you to freely use, modify, and distribute the model, as long as you include the appropriate copyright notice and disclaimer.
95
+
96
+ We do not make any profit from this model. Our goal is to provide a high-quality base model for music generation, fostering innovation in AI music and contributing to the advancement of human creativity. We hope that DiffRhythm 2 will serve as a foundation for further research and development in the field of AI-generated music.
97
+
98
+ DiffRhythm 2 enables the creation of original music across diverse genres, supporting applications in artistic creation, education, and entertainment. While designed for positive use cases, potential risks include unintentional copyright infringement through stylistic similarities, inappropriate blending of cultural musical elements, and misuse for generating harmful content. To ensure responsible deployment, users must implement verification mechanisms to confirm musical originality, disclose AI involvement in generated works, and obtain permissions when adapting protected styles.
99
+
100
+ ## Citation
101
+ ```
102
+ @article{diffrhythm2,
103
+ title={DiffRhythm 2: Efficient and High Fidelity Song Generation via Block Flow Matching},
104
+ author={Jiang, Yuepeng and Chen, Huakang and Ning, Ziqian and Yao, Jixun and Han, Zerui and Wu, Di and Meng, Meng and Luan, Jian and Fu, Zhonghua and Xie, Lei},
105
+ journal={arXiv preprint arXiv:2510.22950},
106
+ year={2025}
107
+ }
108
+ ```