SO0529 commited on
Commit
960ce5b
1 Parent(s): adf7edf

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +47 -0
README.md CHANGED
@@ -1,3 +1,50 @@
1
  ---
2
  license: apache-2.0
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ language:
4
+ - ja
5
  ---
6
+
7
+ # Mixtral-8x7B-v0.1-japanese
8
+
9
+ Mixtral-8x7B-v0.1-japaneseは[Mixtral-8x7B-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-v0.1)をベースに語彙拡張継続事前学習したモデルです。
10
+ 詳細は[ABEJAのテックブログ](https://tech-blog.abeja.asia/)を参照してください。
11
+
12
+
13
+ # 使い方
14
+ ``` python
15
+ import torch
16
+ from transformers import AutoModelForCausalLM, AutoTokenizer
17
+
18
+ model_name = "abeja/Mixtral-8x7B-v0.1-japanese"
19
+
20
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
21
+ model = AutoModelForCausalLM.from_pretrained(
22
+ model_name,
23
+ torch_dtype=torch.float16,
24
+ use_cache=True,
25
+ device_map="auto",
26
+ )
27
+ model.eval()
28
+
29
+ text = "人とAIが協調するためには、"
30
+ input_ids = tokenizer.encode(input_text, return_tensors="pt")
31
+
32
+ with torch.no_grad():
33
+ output_ids = model.generate(
34
+ token_ids.to(model.device),
35
+ max_new_tokens=256,
36
+ pad_token_id=tokenizer.pad_token_id,
37
+ eos_token_id=tokenizer.eos_token_id,
38
+ )
39
+ output = tokenizer.decode(output_ids.tolist()[0], skip_special_tokens=True)
40
+ print(output)
41
+ ```
42
+
43
+ # 開発者
44
+ Kentaro Nakanishi
45
+ Keisuke Fujimoto
46
+ Kyo Hattori
47
+ Shinya Otani
48
+ Shogo Muranushi
49
+ (*)アルファベット順
50
+