nm-research commited on
Commit
73bd426
·
verified ·
1 Parent(s): 577d996

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -8,7 +8,7 @@ base_model: deepseek-ai/DeepSeek-R1-Distill-Llama-8B
8
  library_name: transformers
9
  ---
10
 
11
- # DeepSeek-R1-Distill-Llama-8B-FP8-Dynamic
12
 
13
  ## Model Overview
14
  - **Model Architecture:** LlamaForCausalLM
@@ -42,7 +42,7 @@ from transformers import AutoTokenizer
42
  from vllm import LLM, SamplingParams
43
 
44
  number_gpus = 1
45
- model_name = "neuralmagic/DeepSeek-R1-Distill-Llama-8B-FP8-Dynamic"
46
 
47
  tokenizer = AutoTokenizer.from_pretrained(model_name)
48
  sampling_params = SamplingParams(temperature=0.6, max_tokens=256, stop_token_ids=[tokenizer.eos_token_id])
@@ -112,7 +112,7 @@ OpenLLM Leaderboard V1:
112
  ```
113
  lm_eval \
114
  --model vllm \
115
- --model_args pretrained="neuralmagic/DeepSeek-R1-Distill-Llama-8B-FP8-Dynamic",dtype=auto,max_model_len=4096,enable_chunked_prefill=True \
116
  --tasks openllm \
117
  --write_out \
118
  --batch_size auto \
@@ -124,7 +124,7 @@ OpenLLM Leaderboard V2:
124
  ```
125
  lm_eval \
126
  --model vllm \
127
- --model_args pretrained="neuralmagic/DeepSeek-R1-Distill-Llama-8B-FP8-Dynamic",dtype=auto,max_model_len=4096,enable_chunked_prefill=True \
128
  --apply_chat_template \
129
  --fewshot_as_multiturn \
130
  --tasks leaderboard \
@@ -142,7 +142,7 @@ lm_eval \
142
  <th>Category</th>
143
  <th>Metric</th>
144
  <th>deepseek-ai/DeepSeek-R1-Distill-Llama-8B</th>
145
- <th>neuralmagic/DeepSeek-R1-Distill-Llama-8B-FP8-Dynamic</th>
146
  <th>Recovery</th>
147
  </tr>
148
  </thead>
 
8
  library_name: transformers
9
  ---
10
 
11
+ # DeepSeek-R1-Distill-Llama-8B-FP8-dynamic
12
 
13
  ## Model Overview
14
  - **Model Architecture:** LlamaForCausalLM
 
42
  from vllm import LLM, SamplingParams
43
 
44
  number_gpus = 1
45
+ model_name = "neuralmagic/DeepSeek-R1-Distill-Llama-8B-FP8-dynamic"
46
 
47
  tokenizer = AutoTokenizer.from_pretrained(model_name)
48
  sampling_params = SamplingParams(temperature=0.6, max_tokens=256, stop_token_ids=[tokenizer.eos_token_id])
 
112
  ```
113
  lm_eval \
114
  --model vllm \
115
+ --model_args pretrained="neuralmagic/DeepSeek-R1-Distill-Llama-8B-FP8-dynamic",dtype=auto,max_model_len=4096,enable_chunked_prefill=True \
116
  --tasks openllm \
117
  --write_out \
118
  --batch_size auto \
 
124
  ```
125
  lm_eval \
126
  --model vllm \
127
+ --model_args pretrained="neuralmagic/DeepSeek-R1-Distill-Llama-8B-FP8-dynamic",dtype=auto,max_model_len=4096,enable_chunked_prefill=True \
128
  --apply_chat_template \
129
  --fewshot_as_multiturn \
130
  --tasks leaderboard \
 
142
  <th>Category</th>
143
  <th>Metric</th>
144
  <th>deepseek-ai/DeepSeek-R1-Distill-Llama-8B</th>
145
+ <th>neuralmagic/DeepSeek-R1-Distill-Llama-8B-FP8-dynamic</th>
146
  <th>Recovery</th>
147
  </tr>
148
  </thead>