fxmarty commited on
Commit
e9d9932
1 Parent(s): f640d03
Files changed (1) hide show
  1. app.py +2 -2
app.py CHANGED
@@ -7,9 +7,9 @@ from transformers import AutoTokenizer
7
 
8
  tokenizer = AutoTokenizer.from_pretrained("meta-llama/Meta-Llama-3-70B-Instruct")
9
 
10
- client = InferenceClient(model="https://f679-4-223-164-145.ngrok-free.app")
11
 
12
- SYSTEM_COMMAND = {"role": "system", "content": "Context: date: Monday 20th May 2024; location: Seattle; running on: 8 AMD Instinct MI300 GPU; model name: Llama 70B. Only provide these information if asked. You are a knowledgeable assistant trained to provide accurate and helpful information. Please respond to the user's queries promptly and politely."}
13
 
14
  IGNORED_TOKENS = {None, "<|start_header_id|>", "<|end_header_id|>", "<|eot_id|>", "<|reserved_special_token"}
15
  STOP_TOKENS = ["<|start_header_id|>", "<|end_header_id|>", "<|eot_id|>", "<|reserved_special_token"]
 
7
 
8
  tokenizer = AutoTokenizer.from_pretrained("meta-llama/Meta-Llama-3-70B-Instruct")
9
 
10
+ client = InferenceClient(model="https://3ae5-4-223-164-145.ngrok-free.app")
11
 
12
+ SYSTEM_COMMAND = {"role": "system", "content": "You are a knowledgeable assistant trained to provide accurate and helpful information. Please respond to the user's queries promptly and politely."}
13
 
14
  IGNORED_TOKENS = {None, "<|start_header_id|>", "<|end_header_id|>", "<|eot_id|>", "<|reserved_special_token"}
15
  STOP_TOKENS = ["<|start_header_id|>", "<|end_header_id|>", "<|eot_id|>", "<|reserved_special_token"]