keisawada commited on
Commit
b167c16
1 Parent(s): f33d445

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -0
README.md CHANGED
@@ -62,5 +62,23 @@ Here are a few samples generated with and without the toy prefix weights, respec
62
  # Inference with FasterTransformer
63
  After version 5.1, [NVIDIA FasterTransformer](https://github.com/NVIDIA/FasterTransformer) now supports both inference for GPT-NeoX and a variety of soft prompts (including prefix-tuning). The released pretrained model and prefix weights in this repo have been verified to work with FasterTransformer 5.1.
64
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
65
  # Licenese
66
  [The MIT license](https://opensource.org/licenses/MIT)
62
  # Inference with FasterTransformer
63
  After version 5.1, [NVIDIA FasterTransformer](https://github.com/NVIDIA/FasterTransformer) now supports both inference for GPT-NeoX and a variety of soft prompts (including prefix-tuning). The released pretrained model and prefix weights in this repo have been verified to work with FasterTransformer 5.1.
64
 
65
+ # How to cite
66
+ ~~~
67
+ @misc{rinna-japanese-gpt-neox-small,
68
+ title = {rinna/japanese-gpt-neox-small},
69
+ author = {Zhao, Tianyu and Sawada, Kei}
70
+ url = {https://huggingface.co/rinna/japanese-gpt-neox-small},
71
+ }
72
+
73
+ @inproceedings{sawada2024release,
74
+ title = {Release of Pre-Trained Models for the {J}apanese Language},
75
+ author = {Sawada, Kei and Zhao, Tianyu and Shing, Makoto and Mitsui, Kentaro and Kaga, Akio and Hono, Yukiya and Wakatsuki, Toshiaki and Mitsuda, Koh},
76
+ booktitle = {Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)},
77
+ month = {5},
78
+ year = {2024},
79
+ url = {https://arxiv.org/abs/2404.01657},
80
+ }
81
+ ~~~
82
+
83
  # Licenese
84
  [The MIT license](https://opensource.org/licenses/MIT)