Update README.md
Browse files
README.md
CHANGED
@@ -62,5 +62,23 @@ Here are a few samples generated with and without the toy prefix weights, respec
|
|
62 |
# Inference with FasterTransformer
|
63 |
After version 5.1, [NVIDIA FasterTransformer](https://github.com/NVIDIA/FasterTransformer) now supports both inference for GPT-NeoX and a variety of soft prompts (including prefix-tuning). The released pretrained model and prefix weights in this repo have been verified to work with FasterTransformer 5.1.
|
64 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
65 |
# Licenese
|
66 |
[The MIT license](https://opensource.org/licenses/MIT)
|
|
|
62 |
# Inference with FasterTransformer
|
63 |
After version 5.1, [NVIDIA FasterTransformer](https://github.com/NVIDIA/FasterTransformer) now supports both inference for GPT-NeoX and a variety of soft prompts (including prefix-tuning). The released pretrained model and prefix weights in this repo have been verified to work with FasterTransformer 5.1.
|
64 |
|
65 |
+
# How to cite
|
66 |
+
~~~
|
67 |
+
@misc{rinna-japanese-gpt-neox-small,
|
68 |
+
title = {rinna/japanese-gpt-neox-small},
|
69 |
+
author = {Zhao, Tianyu and Sawada, Kei}
|
70 |
+
url = {https://huggingface.co/rinna/japanese-gpt-neox-small},
|
71 |
+
}
|
72 |
+
|
73 |
+
@inproceedings{sawada2024release,
|
74 |
+
title = {Release of Pre-Trained Models for the {J}apanese Language},
|
75 |
+
author = {Sawada, Kei and Zhao, Tianyu and Shing, Makoto and Mitsui, Kentaro and Kaga, Akio and Hono, Yukiya and Wakatsuki, Toshiaki and Mitsuda, Koh},
|
76 |
+
booktitle = {Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)},
|
77 |
+
month = {5},
|
78 |
+
year = {2024},
|
79 |
+
url = {https://arxiv.org/abs/2404.01657},
|
80 |
+
}
|
81 |
+
~~~
|
82 |
+
|
83 |
# Licenese
|
84 |
[The MIT license](https://opensource.org/licenses/MIT)
|