Update README.md
Browse files
README.md
CHANGED
@@ -57,7 +57,6 @@ from transformers import AutoTokenizer
|
|
57 |
from auto_gptq import AutoGPTQForCausalLM
|
58 |
|
59 |
quantized_model_dir = "dahara1/ELYZA-japanese-Llama-2-7b-fast-instruct-GPTQ"
|
60 |
-
quantized_model_dir = "tmo/ELYZA-japanese-Llama-2-7b-fast-instruct-GPTQ"
|
61 |
|
62 |
model_basename = "gptq_model-4bit-128g"
|
63 |
tokenizer = AutoTokenizer.from_pretrained(quantized_model_dir)
|
|
|
57 |
from auto_gptq import AutoGPTQForCausalLM
|
58 |
|
59 |
quantized_model_dir = "dahara1/ELYZA-japanese-Llama-2-7b-fast-instruct-GPTQ"
|
|
|
60 |
|
61 |
model_basename = "gptq_model-4bit-128g"
|
62 |
tokenizer = AutoTokenizer.from_pretrained(quantized_model_dir)
|