Update README.md
Browse files
README.md
CHANGED
@@ -84,7 +84,7 @@ We evaluate the pretrained MobileLLM models on Zero-shot Common Sense Reasoning
|
|
84 |
|
85 |
## MobileLLM-125M
|
86 |
|
87 |
-
| model |
|
88 |
| --- | --- | --- | --- | --- | --- | --- | --- | --- | --- |
|
89 |
| OPT-125M | 41.3 | 25.2 | 57.5 | 62.0 | 41.9 | 31.1 | 31.2 | 50.8 | 42.6 |
|
90 |
| GPT-neo-125M | 40.7 | 24.8 | 61.3 | 62.5 | 41.9 | 29.7 | 31.6 | 50.7 | 42.9 |
|
@@ -94,7 +94,7 @@ We evaluate the pretrained MobileLLM models on Zero-shot Common Sense Reasoning
|
|
94 |
|
95 |
## MobileLLM-350M
|
96 |
|
97 |
-
| model |
|
98 |
| --- | --- | --- | --- | --- | --- | --- | --- | --- | --- |
|
99 |
| OPT-350M | 41.9 | 25.7 | 54.0 | 64.8 | 42.6 | 36.2 | 33.3 | 52.4 | 43.9 |
|
100 |
| Pythia-410M | 47.1 | 30.3 | 55.3 | 67.2 | 43.1 | 40.1 | 36.2 | 53.4 | 46.6 |
|
@@ -103,7 +103,7 @@ We evaluate the pretrained MobileLLM models on Zero-shot Common Sense Reasoning
|
|
103 |
|
104 |
## MobileLLM-600M
|
105 |
|
106 |
-
| model |
|
107 |
| --- | --- | --- | --- | --- | --- | --- | --- | --- | --- |
|
108 |
| Qwen1.5-500M | 54.7 | 32.1 | 46.9 | 68.9 | 46.0 | 48.8 | 37.7 | 55.0 | 48.8 |
|
109 |
| BLOOM-560M | 43.7 | 27.5 | 53.7 | 65.1 | 42.5 | 36.5 | 32.6 | 52.2 | 44.2 |
|
@@ -112,7 +112,7 @@ We evaluate the pretrained MobileLLM models on Zero-shot Common Sense Reasoning
|
|
112 |
|
113 |
## MobileLLM-1B
|
114 |
|
115 |
-
| model |
|
116 |
| --- | --- | --- | --- | --- | --- | --- | --- | --- | --- |
|
117 |
| Pythia-1B | 49.9 | 30.4 | 58.7 | 69.2 | 43.3 | 47.4 | 38.6 | 52.2 | 48.7 |
|
118 |
| MobiLlama-1B | 59.7 | 38.4 | 59.2 | 74.5 | 44.9 | 62.0 | 43.7 | 59.0 | 55.2 |
|
@@ -123,7 +123,7 @@ We evaluate the pretrained MobileLLM models on Zero-shot Common Sense Reasoning
|
|
123 |
|
124 |
## MobileLLM-1.5B
|
125 |
|
126 |
-
| model |
|
127 |
| --- | --- | --- | --- | --- | --- | --- | --- | --- | --- |
|
128 |
| GPT-neo-1.3B | 51.3 | 33.0 | 61.8 | 70.9 | 43.7 | 48.6 | 41.2 | 54.5 | 50.6 |
|
129 |
| OPT-1.3B | 54.4 | 31.7 | 58.4 | 71.5 | 44.7 | 53.7 | 44.6 | 59.1 | 52.3 |
|
|
|
84 |
|
85 |
## MobileLLM-125M
|
86 |
|
87 |
+
| model | arc_easy | arc_challenge | boolq | piqa | siqa | hellaswag | obqa | winogrande | avg. |
|
88 |
| --- | --- | --- | --- | --- | --- | --- | --- | --- | --- |
|
89 |
| OPT-125M | 41.3 | 25.2 | 57.5 | 62.0 | 41.9 | 31.1 | 31.2 | 50.8 | 42.6 |
|
90 |
| GPT-neo-125M | 40.7 | 24.8 | 61.3 | 62.5 | 41.9 | 29.7 | 31.6 | 50.7 | 42.9 |
|
|
|
94 |
|
95 |
## MobileLLM-350M
|
96 |
|
97 |
+
| model | arc_easy | arc_challenge | boolq | piqa | siqa | hellaswag | obqa | winogrande | avg. |
|
98 |
| --- | --- | --- | --- | --- | --- | --- | --- | --- | --- |
|
99 |
| OPT-350M | 41.9 | 25.7 | 54.0 | 64.8 | 42.6 | 36.2 | 33.3 | 52.4 | 43.9 |
|
100 |
| Pythia-410M | 47.1 | 30.3 | 55.3 | 67.2 | 43.1 | 40.1 | 36.2 | 53.4 | 46.6 |
|
|
|
103 |
|
104 |
## MobileLLM-600M
|
105 |
|
106 |
+
| model | arc_easy | arc_challenge | boolq | piqa | siqa | hellaswag | obqa | winogrande | avg. |
|
107 |
| --- | --- | --- | --- | --- | --- | --- | --- | --- | --- |
|
108 |
| Qwen1.5-500M | 54.7 | 32.1 | 46.9 | 68.9 | 46.0 | 48.8 | 37.7 | 55.0 | 48.8 |
|
109 |
| BLOOM-560M | 43.7 | 27.5 | 53.7 | 65.1 | 42.5 | 36.5 | 32.6 | 52.2 | 44.2 |
|
|
|
112 |
|
113 |
## MobileLLM-1B
|
114 |
|
115 |
+
| model | arc_easy | arc_challenge | boolq | piqa | siqa | hellaswag | obqa | winogrande | avg. |
|
116 |
| --- | --- | --- | --- | --- | --- | --- | --- | --- | --- |
|
117 |
| Pythia-1B | 49.9 | 30.4 | 58.7 | 69.2 | 43.3 | 47.4 | 38.6 | 52.2 | 48.7 |
|
118 |
| MobiLlama-1B | 59.7 | 38.4 | 59.2 | 74.5 | 44.9 | 62.0 | 43.7 | 59.0 | 55.2 |
|
|
|
123 |
|
124 |
## MobileLLM-1.5B
|
125 |
|
126 |
+
| model | arc_easy | arc_challenge | boolq | piqa | siqa | hellaswag | obqa | winogrande | avg. |
|
127 |
| --- | --- | --- | --- | --- | --- | --- | --- | --- | --- |
|
128 |
| GPT-neo-1.3B | 51.3 | 33.0 | 61.8 | 70.9 | 43.7 | 48.6 | 41.2 | 54.5 | 50.6 |
|
129 |
| OPT-1.3B | 54.4 | 31.7 | 58.4 | 71.5 | 44.7 | 53.7 | 44.6 | 59.1 | 52.3 |
|