Update funcs.py
Browse files
funcs.py
CHANGED
|
@@ -133,7 +133,7 @@ def summarize_and_recommend(therapy_session_conversation):
|
|
| 133 |
import torch
|
| 134 |
|
| 135 |
# Falcon-180B model name
|
| 136 |
-
model_name = "tiiuae/falcon-
|
| 137 |
|
| 138 |
# Load tokenizer and model
|
| 139 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
|
@@ -142,7 +142,7 @@ def summarize_and_recommend(therapy_session_conversation):
|
|
| 142 |
device_map="auto", # Automatically maps layers to available GPUs
|
| 143 |
torch_dtype=torch.bfloat16, # Use bf16 if supported
|
| 144 |
trust_remote_code=True, # Falcon uses custom model code
|
| 145 |
-
token=huggingface_token
|
| 146 |
)
|
| 147 |
|
| 148 |
# Create conversation prompt (chat-style)
|
|
|
|
| 133 |
import torch
|
| 134 |
|
| 135 |
# Falcon-180B model name
|
| 136 |
+
model_name = "tiiuae/falcon-7b-instruct"
|
| 137 |
|
| 138 |
# Load tokenizer and model
|
| 139 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
|
|
|
| 142 |
device_map="auto", # Automatically maps layers to available GPUs
|
| 143 |
torch_dtype=torch.bfloat16, # Use bf16 if supported
|
| 144 |
trust_remote_code=True, # Falcon uses custom model code
|
| 145 |
+
# token=huggingface_token
|
| 146 |
)
|
| 147 |
|
| 148 |
# Create conversation prompt (chat-style)
|