Update README.md
Browse files
README.md
CHANGED
@@ -57,7 +57,7 @@ repo_id = "ruslanmv/Medical-Llama3-8B-GPTQ"
|
|
57 |
|
58 |
# download quantized model from Hugging Face Hub and load to the first GPU
|
59 |
model = AutoGPTQForCausalLM.from_quantized(repo_id,
|
60 |
-
device=
|
61 |
use_safetensors=True,
|
62 |
use_triton=False)
|
63 |
tokenizer = AutoTokenizer.from_pretrained(quantized_model_dir)
|
|
|
57 |
|
58 |
# download quantized model from Hugging Face Hub and load to the first GPU
|
59 |
model = AutoGPTQForCausalLM.from_quantized(repo_id,
|
60 |
+
device=device,
|
61 |
use_safetensors=True,
|
62 |
use_triton=False)
|
63 |
tokenizer = AutoTokenizer.from_pretrained(quantized_model_dir)
|