Update README.md
Browse files
README.md
CHANGED
@@ -90,18 +90,17 @@ and is comparable with Mistral-7B-Instruct-v0.1 on MMLU and MT-Bench in English.
|
|
90 |
|
91 |
\* Taiwan-LLM models responds to multi-turn questions (English) in Traditional Chinese.
|
92 |
|
93 |
-
| Category ACC of TMMLU+ (0 shot)
|
94 |
|-----------------------------------------------------|--------------|----------------|------------|------------|
|
95 |
| gpt-3.5-turbo-1106 | | | | |
|
96 |
-
| Yi-34B-Chat
|
97 |
-
| Qwen-14B-Chat
|
98 |
-
| Yi-6B-Chat
|
99 |
-
| Breeze-7B-Instruct-v0.1
|
100 |
-
| Breeze-7B-Instruct-64k-v0.1
|
101 |
-
| Qwen-7B-Chat
|
102 |
-
| Taiwan-LLM-13B-v2.0-chat
|
103 |
-
| Taiwan-LLM-7B-v2.1-chat
|
104 |
-
|
105 |
|
106 |
|
107 |
## Examples
|
|
|
90 |
|
91 |
\* Taiwan-LLM models responds to multi-turn questions (English) in Traditional Chinese.
|
92 |
|
93 |
+
| Category ACC of TMMLU+ (0 shot) | STEM | Social Science | Humanities | Other |
|
94 |
|-----------------------------------------------------|--------------|----------------|------------|------------|
|
95 |
| gpt-3.5-turbo-1106 | | | | |
|
96 |
+
| Yi-34B-Chat | 47.65 | 64.25 | 52.73 | 54.91 |
|
97 |
+
| Qwen-14B-Chat | 43.83 | 55.00 | 48.55 | 46.22 |
|
98 |
+
| Yi-6B-Chat | 37.80 | 51.74 | 45.36 | 44.25 |
|
99 |
+
| Breeze-7B-Instruct-v0.1 | 37.41 | 46.81 | 42.06 | 40.16 |
|
100 |
+
| Breeze-7B-Instruct-64k-v0.1 | 37.88 | 46.35 | 40.31 | 39.40 |
|
101 |
+
| Qwen-7B-Chat | 35.44 | 46.22 | 38.35 | 40.06 |
|
102 |
+
| Taiwan-LLM-13B-v2.0-chat | 27.74 | 33.69 | 27.03 | 29.43 |
|
103 |
+
| Taiwan-LLM-7B-v2.1-chat | 25.58 | 31.76 | 27.36 | 27.61 |
|
|
|
104 |
|
105 |
|
106 |
## Examples
|