Update README.md
Browse files
README.md
CHANGED
@@ -147,7 +147,7 @@ import torch
|
|
147 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
148 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
149 |
model = AutoModelForCausalLM.from_pretrained(
|
150 |
-
"OuteAI/Lite-Oute-2-Mamba2Attn-
|
151 |
# To allow custom modeling files
|
152 |
trust_remote_code=True,
|
153 |
|
@@ -155,7 +155,7 @@ model = AutoModelForCausalLM.from_pretrained(
|
|
155 |
# attn_implementation="flash_attention_2",
|
156 |
)
|
157 |
model.to(device)
|
158 |
-
tokenizer = AutoTokenizer.from_pretrained("OuteAI/Lite-Oute-2-Mamba2Attn-
|
159 |
|
160 |
def generate_response(message: str, temperature: float = 0.2, repetition_penalty: float = 1.12) -> str:
|
161 |
# Convert message to PyTorch tensors
|
|
|
147 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
148 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
149 |
model = AutoModelForCausalLM.from_pretrained(
|
150 |
+
"OuteAI/Lite-Oute-2-Mamba2Attn-Base",
|
151 |
# To allow custom modeling files
|
152 |
trust_remote_code=True,
|
153 |
|
|
|
155 |
# attn_implementation="flash_attention_2",
|
156 |
)
|
157 |
model.to(device)
|
158 |
+
tokenizer = AutoTokenizer.from_pretrained("OuteAI/Lite-Oute-2-Mamba2Attn-Base")
|
159 |
|
160 |
def generate_response(message: str, temperature: float = 0.2, repetition_penalty: float = 1.12) -> str:
|
161 |
# Convert message to PyTorch tensors
|