edwko commited on
Commit
fecf524
·
verified ·
1 Parent(s): 1ab68a9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -147,7 +147,7 @@ import torch
147
  from transformers import AutoModelForCausalLM, AutoTokenizer
148
  device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
149
  model = AutoModelForCausalLM.from_pretrained(
150
- "OuteAI/Lite-Oute-2-Mamba2Attn-Instruct",
151
  # To allow custom modeling files
152
  trust_remote_code=True,
153
 
@@ -155,7 +155,7 @@ model = AutoModelForCausalLM.from_pretrained(
155
  # attn_implementation="flash_attention_2",
156
  )
157
  model.to(device)
158
- tokenizer = AutoTokenizer.from_pretrained("OuteAI/Lite-Oute-2-Mamba2Attn-Instruct")
159
 
160
  def generate_response(message: str, temperature: float = 0.2, repetition_penalty: float = 1.12) -> str:
161
  # Convert message to PyTorch tensors
 
147
  from transformers import AutoModelForCausalLM, AutoTokenizer
148
  device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
149
  model = AutoModelForCausalLM.from_pretrained(
150
+ "OuteAI/Lite-Oute-2-Mamba2Attn-Base",
151
  # To allow custom modeling files
152
  trust_remote_code=True,
153
 
 
155
  # attn_implementation="flash_attention_2",
156
  )
157
  model.to(device)
158
+ tokenizer = AutoTokenizer.from_pretrained("OuteAI/Lite-Oute-2-Mamba2Attn-Base")
159
 
160
  def generate_response(message: str, temperature: float = 0.2, repetition_penalty: float = 1.12) -> str:
161
  # Convert message to PyTorch tensors