RioShiina commited on
Commit
dbaf5eb
1 Parent(s): c5b3ffe

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +71 -3
README.md CHANGED
@@ -1,3 +1,71 @@
1
- ---
2
- license: llama3.1
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: llama3.1
3
+ language:
4
+ - ja
5
+ - en
6
+ ---
7
+
8
+ **[2.2bpw](https://huggingface.co/rioshiina/Llama-3.1-Swallow-70B-Instruct-v0.1-exl2/tree/2.2bpw)** (high quality loss, only for 24GB vRAM test.)
9
+ **[4.0bpw](https://huggingface.co/rioshiina/Llama-3.1-Swallow-70B-Instruct-v0.1-exl2/tree/4.0bpw)**
10
+ **[6.0bpw](https://huggingface.co/rioshiina/Llama-3.1-Swallow-70B-Instruct-v0.1-exl2/tree/6.0bpw)**
11
+ **[8.0bpw](https://huggingface.co/rioshiina/Llama-3.1-Swallow-70B-Instruct-v0.1-exl2/tree/8.0bpw)**
12
+
13
+ # Llama-3.1-Swallow-70B-Instruct-v0.1-exl2
14
+ - Model creator: [tokyotech-llm](https://huggingface.co/tokyotech-llm)
15
+ - Original model: [Llama-3.1-Swallow-70B-Instruct-v0.1](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-70B-Instruct-v0.1)
16
+
17
+ ### License
18
+
19
+ [META LLAMA 3.1 COMMUNITY LICENSE](https://www.llama.com/llama3_1/license/)
20
+
21
+ ## Prompt template
22
+
23
+ ```
24
+ <|begin_of_text|><|start_header_id|>system<|end_header_id|>
25
+
26
+ あなたは誠実で優秀な日本人のアシスタントです。<|eot_id|><|start_header_id|>user<|end_header_id|>
27
+
28
+ 東京の紅葉した公園で、東京タワーと高層ビルを背景に、空を舞うツバメと草地に佇むラマが出会う温かな物語を書いてください。<|eot_id|><|start_header_id|>assistant<|end_header_id|>
29
+
30
+ ```
31
+
32
+ ### Citations
33
+
34
+ ```tex
35
+ @inproceedings{Fujii:COLM2024,
36
+ title={Continual Pre-Training for Cross-Lingual LLM Adaptation:
37
+ Enhancing Japanese Language Capabilities},
38
+ author={Kazuki Fujii and Taishi Nakamura and Mengsay Loem and Hiroki
39
+ Iida and Masanari Ohi and Kakeru Hattori and Hirai Shota and Sakae
40
+ Mizuki and Rio Yokota and Naoaki Okazaki},
41
+ booktitle="Proceedings of the First Conference on Language Modeling",
42
+ series={COLM},
43
+ pages="(to appear)",
44
+ year="2024",
45
+ month=oct,
46
+ address={University of Pennsylvania, USA},
47
+ }
48
+
49
+ @inproceedings{Okazaki:COLM2024,
50
+ title={Building a Large Japanese Web Corpus for Large Language Models},
51
+ author={Naoaki Okazaki and Kakeru Hattori and Hirai Shota and Hiroki
52
+ Iida and Masanari Ohi and Kazuki Fujii and Taishi Nakamura and Mengsay
53
+ Loem and Rio Yokota and Sakae Mizuki},
54
+ booktitle="Proceedings of the First Conference on Language Modeling",
55
+ series={COLM},
56
+ pages="(to appear)",
57
+ year="2024",
58
+ month=oct,
59
+ address={University of Pennsylvania, USA},
60
+ }
61
+
62
+ @misc{dubey2024llama3herdmodels,
63
+ title={The Llama 3 Herd of Models},
64
+ author={Abhimanyu Dubey and Abhinav Jauhri and Abhinav Pandey and Abhishek Kadian and Ahmad Al-Dahle and Aiesha Letman and Akhil Mathur and Alan Schelten and Amy Yang and Angela Fan et al.},
65
+ year={2024},
66
+ eprint={2407.21783},
67
+ archivePrefix={arXiv},
68
+ primaryClass={cs.AI},
69
+ url={https://arxiv.org/abs/2407.21783},
70
+ }
71
+ ```