update read,e
Browse files
README.md
CHANGED
@@ -15,8 +15,18 @@ tags:
|
|
15 |
|
16 |
# Japanese DialoGPT trained with Aozora
|
17 |
|
18 |
-
|
19 |
-
|
20 |
|
21 |
-
##
|
22 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
15 |
|
16 |
# Japanese DialoGPT trained with Aozora
|
17 |
|
18 |
+
**(ja) 青空文庫のセリフで学習した日本語のDialoGPT Smallです**
|
19 |
+
**(en) Japanese DialoGPT Small trained on Aozora Bunko.**
|
20 |
|
21 |
+
## [Demo](https://huggingface.co/spaces/akiFQC/Japanese_DialoGPT_small_Aozora)
|
22 |
+
Demo in this page is not working so well. I recommend you to try it on [Hugging Face Spaces Version](https://huggingface.co/spaces/akiFQC/Japanese_DialoGPT_small_Aozora).
|
23 |
+
|
24 |
+
|
25 |
+
## Reference
|
26 |
+
- [Aozora-bunko](https://www.aozora.gr.jp/)
|
27 |
+
- Japanese public domain books.
|
28 |
+
- I extracted the dialogue part from the books and used it as the training data.
|
29 |
+
- [japanese-gpt2-small](https://huggingface.co/rinna/japanese-gpt2-small)
|
30 |
+
- Novel Japanese GPT2. I used a small model because of the limitation of GPU memory of my desktop PC(with RTX3060x1) 😢.
|
31 |
+
- I used this model as a pre-trained model.
|
32 |
+
- [DialoGPT: Large-Scale Generative Pre-training for Conversational Response Generation](https://arxiv.org/abs/1911.00536)
|