Update app.py (#40)
Browse files- Update app.py (891d9b976d4a7a50cee129679e3223544a29d9a4)
app.py
CHANGED
@@ -8,8 +8,7 @@ from smolagents import CodeAgent, DuckDuckGoSearchTool, HfApiModel, InferenceCli
|
|
8 |
from huggingface_hub import login
|
9 |
from transformers import (
|
10 |
AutoTokenizer,
|
11 |
-
AutoModelForCausalLM
|
12 |
-
BitsAndBytesConfig
|
13 |
)
|
14 |
import os
|
15 |
|
@@ -34,17 +33,10 @@ class BasicAgent:
|
|
34 |
|
35 |
# Load tokenizer and inject a minimal chat_template
|
36 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
37 |
-
|
38 |
-
quant_config = BitsAndBytesConfig(
|
39 |
-
load_in_4bit=True,
|
40 |
-
bnb_4bit_quant_type="nf4",
|
41 |
-
bnb_4bit_use_double_quant=True,
|
42 |
-
bnb_4bit_compute_dtype=torch.float32
|
43 |
-
)
|
44 |
# Load the GPT-2 model (FP16 if you like, but default is fine)
|
45 |
model = AutoModelForCausalLM.from_pretrained(
|
46 |
model_name,
|
47 |
-
quantization_config=quant_config,
|
48 |
device_map="auto"
|
49 |
)
|
50 |
|
|
|
8 |
from huggingface_hub import login
|
9 |
from transformers import (
|
10 |
AutoTokenizer,
|
11 |
+
AutoModelForCausalLM
|
|
|
12 |
)
|
13 |
import os
|
14 |
|
|
|
33 |
|
34 |
# Load tokenizer and inject a minimal chat_template
|
35 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
36 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
37 |
# Load the GPT-2 model (FP16 if you like, but default is fine)
|
38 |
model = AutoModelForCausalLM.from_pretrained(
|
39 |
model_name,
|
|
|
40 |
device_map="auto"
|
41 |
)
|
42 |
|