Update README.md
Browse files
README.md
CHANGED
@@ -36,7 +36,7 @@ from tqdm import tqdm
|
|
36 |
import json
|
37 |
|
38 |
HF_TOKEN = "your-token"
|
39 |
-
model_name = "kky84176/llm-jp-3-13b-
|
40 |
|
41 |
#
|
42 |
bnb_config = BitsAndBytesConfig(
|
@@ -84,7 +84,7 @@ for data in tqdm(datasets):
|
|
84 |
outputs = model.generate(
|
85 |
tokenized_input,
|
86 |
attention_mask=attention_mask,
|
87 |
-
max_new_tokens=
|
88 |
do_sample=False,
|
89 |
repetition_penalty=1.2,
|
90 |
pad_token_id=tokenizer.eos_token_id
|
@@ -95,7 +95,7 @@ for data in tqdm(datasets):
|
|
95 |
|
96 |
# jsonl への出力
|
97 |
import re
|
98 |
-
new_model_id = "llm-jp-3-13b-
|
99 |
jsonl_id = re.sub(".*/", "", new_model_id)
|
100 |
with open(f"./{jsonl_id}-outputs.jsonl", 'w', encoding='utf-8') as f:
|
101 |
for result in results:
|
|
|
36 |
import json
|
37 |
|
38 |
HF_TOKEN = "your-token"
|
39 |
+
model_name = "kky84176/llm-jp-3-13b-instruct-it03"
|
40 |
|
41 |
#
|
42 |
bnb_config = BitsAndBytesConfig(
|
|
|
84 |
outputs = model.generate(
|
85 |
tokenized_input,
|
86 |
attention_mask=attention_mask,
|
87 |
+
max_new_tokens=512,
|
88 |
do_sample=False,
|
89 |
repetition_penalty=1.2,
|
90 |
pad_token_id=tokenizer.eos_token_id
|
|
|
95 |
|
96 |
# jsonl への出力
|
97 |
import re
|
98 |
+
new_model_id = "llm-jp-3-13b-instruct-it03"
|
99 |
jsonl_id = re.sub(".*/", "", new_model_id)
|
100 |
with open(f"./{jsonl_id}-outputs.jsonl", 'w', encoding='utf-8') as f:
|
101 |
for result in results:
|