Spaces:
Sleeping
Sleeping
seapoe1809
commited on
Update darnabot.py
Browse files- darnabot.py +6 -5
darnabot.py
CHANGED
@@ -72,15 +72,16 @@ import spaces
|
|
72 |
MODEL_LIST = ["mistralai/Mistral-Nemo-Instruct-2407"]
|
73 |
HF_TOKEN = os.environ.get("HF_TOKEN", None)
|
74 |
MODEL = os.environ.get("MODEL_ID")
|
75 |
-
|
|
|
76 |
|
77 |
tokenizer = AutoTokenizer.from_pretrained(MODEL)
|
78 |
model = AutoModelForCausalLM.from_pretrained(
|
79 |
MODEL,
|
80 |
-
torch_dtype=torch.bfloat16,
|
81 |
-
device_map="auto",
|
82 |
-
ignore_mismatched_sizes=True
|
83 |
-
|
84 |
|
85 |
async def chat(messages):
|
86 |
# Convert messages to the format required for the model
|
|
|
72 |
MODEL_LIST = ["mistralai/Mistral-Nemo-Instruct-2407"]
|
73 |
HF_TOKEN = os.environ.get("HF_TOKEN", None)
|
74 |
MODEL = os.environ.get("MODEL_ID")
|
75 |
+
|
76 |
+
device = "cuda"
|
77 |
|
78 |
tokenizer = AutoTokenizer.from_pretrained(MODEL)
|
79 |
model = AutoModelForCausalLM.from_pretrained(
|
80 |
MODEL,
|
81 |
+
torch_dtype=torch.bfloat16,
|
82 |
+
device_map="auto",
|
83 |
+
ignore_mismatched_sizes=True)
|
84 |
+
|
85 |
|
86 |
async def chat(messages):
|
87 |
# Convert messages to the format required for the model
|