Update README.md
Browse files
README.md
CHANGED
@@ -27,14 +27,14 @@ from transformers import AutoModelForCausalLM
|
|
27 |
|
28 |
MODEL_ID = "LoftQ/Llama-2-7b-hf-fp16-64rank-gsm8k"
|
29 |
|
30 |
-
|
31 |
MODEL_ID,
|
32 |
torch_dtype=torch.bfloat16, # you may change it with different models
|
33 |
token=YOUR_HF_TOKEN,
|
34 |
|
35 |
)
|
36 |
|
37 |
-
# Do inference with
|
38 |
```
|
39 |
|
40 |
See full evaluation on GSM8K on [Github](https://github.com/yxli2123/LoftQ/blob/main/test_gsm8k.py).
|
|
|
27 |
|
28 |
MODEL_ID = "LoftQ/Llama-2-7b-hf-fp16-64rank-gsm8k"
|
29 |
|
30 |
+
model = AutoModelForCausalLM.from_pretrained(
|
31 |
MODEL_ID,
|
32 |
torch_dtype=torch.bfloat16, # you may change it with different models
|
33 |
token=YOUR_HF_TOKEN,
|
34 |
|
35 |
)
|
36 |
|
37 |
+
# Do inference with `model` ...
|
38 |
```
|
39 |
|
40 |
See full evaluation on GSM8K on [Github](https://github.com/yxli2123/LoftQ/blob/main/test_gsm8k.py).
|