YC-Chen commited on
Commit
47300a8
1 Parent(s): b2afa7e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -5
README.md CHANGED
@@ -84,16 +84,19 @@ and is comparable with Mistral-7B-Instruct-v0.1 on MMLU and MT-Bench in English.
84
 
85
  \* Taiwan-LLM models responds to multi-turn questions (English) in Traditional Chinese.
86
 
87
- | Category ACC of TMMLU+ | STEM | Social Science | Humanities | Other |
88
  |-----------------------------------------------------|--------------|----------------|------------|------------|
 
 
 
89
  | Yi-6B-Chat | 26.28 | 33.48 | 29.48 | 27.62 |
90
  | Breeze-7B-Instruct-v0.1 | 37.45 | 48.35 | 40.26 | 40.44 |
91
- | Taiwan-LLM-7B-v2.1-chat | 26.53 | 29.47 | 26.11 | 26.90 |
92
  | Qwen-7B-Chat | 32.89 | 44.26 | 38.21 | 37.83 |
93
  | Taiwan-LLM-13B-v2.0-chat | 29.68 | 37.13 | 30.31 | 30.55 |
94
- | Qwen-14B-Chat | 46.51 | 58.20 | 51.12 | 49.38 |
95
- | Yi-34B-Chat | 46.36 | 65.02 | 52.84 | 52.21 |
96
- | gpt-3.5-turbo-1106 | | | | |
97
 
98
  ## Examples
99
 
 
84
 
85
  \* Taiwan-LLM models responds to multi-turn questions (English) in Traditional Chinese.
86
 
87
+ | Category ACC of TMMLU+ (0 shot) | STEM | Social Science | Humanities | Other |
88
  |-----------------------------------------------------|--------------|----------------|------------|------------|
89
+ | gpt-3.5-turbo-1106 | | | | |
90
+ | Yi-34B-Chat | 46.36 | 65.02 | 52.84 | 52.21 |
91
+ | Qwen-14B-Chat | 46.51 | 58.20 | 51.12 | 49.38 |
92
  | Yi-6B-Chat | 26.28 | 33.48 | 29.48 | 27.62 |
93
  | Breeze-7B-Instruct-v0.1 | 37.45 | 48.35 | 40.26 | 40.44 |
94
+ | Breeze-7B-Instruct-64k-v0.1 | 37.45 | 48.35 | 40.26 | 40.44 |
95
  | Qwen-7B-Chat | 32.89 | 44.26 | 38.21 | 37.83 |
96
  | Taiwan-LLM-13B-v2.0-chat | 29.68 | 37.13 | 30.31 | 30.55 |
97
+ | Taiwan-LLM-7B-v2.1-chat | 26.53 | 29.47 | 26.11 | 26.90 |
98
+
99
+
100
 
101
  ## Examples
102