OzTianlu commited on
Commit
f54f89f
·
verified ·
1 Parent(s): 253c71b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -8
README.md CHANGED
@@ -23,16 +23,15 @@ A 30B-parameter instruction-tuned language model optimized for reasoning, math,
23
  | | |
24
  |---|---|
25
  | **Model** | Kai-30B-Instruct |
26
- | **Architecture** | LlamaForCausalLM |
27
  | **Parameters** | ~30B |
28
- | **Hidden size** | 7168 |
29
- | **Intermediate size** | 20480 |
30
- | **Layers** | 60 |
31
- | **Attention heads** | 56 (8 KV heads, GQA) |
32
- | **Head dim** | 128 |
33
- | **Context length** | 4096 |
34
  | **Precision** | bfloat16 |
35
- | **Vocab size** | 64,000 |
36
  | **Chat template** | ChatML (`<\|im_start\|>` / `<\|im_end\|>`) |
37
 
38
  ## Benchmark Results (5-shot, acc_norm)
 
23
  | | |
24
  |---|---|
25
  | **Model** | Kai-30B-Instruct |
26
+ | **Architecture** | Qwen2ForCausalLM |
27
  | **Parameters** | ~30B |
28
+ | **Hidden size** | 5120 |
29
+ | **Intermediate size** | 27648 |
30
+ | **Layers** | 64 |
31
+ | **Attention heads** | 40 (8 KV heads, GQA) |
32
+ | **Context length** | 32768 |
 
33
  | **Precision** | bfloat16 |
34
+ | **Vocab size** | 152064 |
35
  | **Chat template** | ChatML (`<\|im_start\|>` / `<\|im_end\|>`) |
36
 
37
  ## Benchmark Results (5-shot, acc_norm)