Spaces:
Runtime error
Runtime error
Commit
•
ddaaa7d
1
Parent(s):
79ecda9
Update app.py
Browse files
app.py
CHANGED
@@ -2,19 +2,17 @@ import os
|
|
2 |
HF_TOKEN = os.getenv('HF_TOKEN')
|
3 |
print("Token loaded")
|
4 |
|
5 |
-
from transformers import GemmaTokenizer, AutoModelForCausalLM
|
6 |
-
from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
|
7 |
-
from threading import Thread
|
8 |
import transformers
|
9 |
import torch
|
10 |
|
11 |
-
|
12 |
-
|
13 |
-
|
14 |
-
|
15 |
-
|
16 |
-
|
17 |
-
|
|
|
18 |
|
19 |
print("llama download successfully")
|
20 |
|
|
|
2 |
HF_TOKEN = os.getenv('HF_TOKEN')
|
3 |
print("Token loaded")
|
4 |
|
|
|
|
|
|
|
5 |
import transformers
|
6 |
import torch
|
7 |
|
8 |
+
model_id = "meta-llama/Meta-Llama-3-8B-Instruct"
|
9 |
+
|
10 |
+
pipeline = transformers.pipeline(
|
11 |
+
"text-generation",
|
12 |
+
model="meta-llama/Meta-Llama-3-8B-Instruct",
|
13 |
+
model_kwargs={"torch_dtype": torch.bfloat16},
|
14 |
+
device="cuda",
|
15 |
+
)
|
16 |
|
17 |
print("llama download successfully")
|
18 |
|