yentinglin commited on
Commit
a2ab71e
1 Parent(s): 2413fcb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -214,7 +214,7 @@ class EosListStoppingCriteria(StoppingCriteria):
214
  return self.eos_sequence in last_ids
215
 
216
  # Initialize the model with automatic device mapping
217
- llm = pipeline("text-generation", model="yentinglin/Llama-3-Taiwan-70B-Instruct-rc1", device_map="auto")
218
  tokenizer = llm.tokenizer
219
 
220
  # Define a conversation example
@@ -258,7 +258,7 @@ docker run \
258
  -p "${PORT}:8000" \
259
  --ipc=host \
260
  vllm/vllm-openai:v0.4.0.post1 \
261
- --model "yentinglin/Llama-3-Taiwan-70B-Instruct-rc1" \
262
  -tp "${NUM_GPUS}"
263
  ```
264
 
@@ -277,7 +277,7 @@ client = OpenAI(
277
  )
278
 
279
  chat_response = client.chat.completions.create(
280
- model="yentinglin/Llama-3-Taiwan-70B-Instruct-rc1",
281
  messages=[
282
  {"role": "system", "content": "You are a helpful assistant."},
283
  {"role": "user", "content": "Tell me a joke."},
 
214
  return self.eos_sequence in last_ids
215
 
216
  # Initialize the model with automatic device mapping
217
+ llm = pipeline("text-generation", model="yentinglin/Llama-3-Taiwan-70B-Instruct", device_map="auto")
218
  tokenizer = llm.tokenizer
219
 
220
  # Define a conversation example
 
258
  -p "${PORT}:8000" \
259
  --ipc=host \
260
  vllm/vllm-openai:v0.4.0.post1 \
261
+ --model "yentinglin/Llama-3-Taiwan-70B-Instruct" \
262
  -tp "${NUM_GPUS}"
263
  ```
264
 
 
277
  )
278
 
279
  chat_response = client.chat.completions.create(
280
+ model="yentinglin/Llama-3-Taiwan-70B-Instruct",
281
  messages=[
282
  {"role": "system", "content": "You are a helpful assistant."},
283
  {"role": "user", "content": "Tell me a joke."},