picocreator commited on
Commit
a8d3efc
1 Parent(s): 3bddf42
compile-results.ipynb CHANGED
@@ -2,7 +2,7 @@
2
  "cells": [
3
  {
4
  "cell_type": "code",
5
- "execution_count": 62,
6
  "metadata": {},
7
  "outputs": [
8
  {
@@ -11,10 +11,10 @@
11
  "text": [
12
  "Defaulting to user installation because normal site-packages is not writeable\n",
13
  "Requirement already satisfied: pandas in /Users/picocreator/Library/Python/3.9/lib/python/site-packages (2.2.0)\n",
 
 
14
  "Requirement already satisfied: pytz>=2020.1 in /Users/picocreator/Library/Python/3.9/lib/python/site-packages (from pandas) (2024.1)\n",
15
  "Requirement already satisfied: numpy<2,>=1.22.4 in /Users/picocreator/Library/Python/3.9/lib/python/site-packages (from pandas) (1.26.1)\n",
16
- "Requirement already satisfied: tzdata>=2022.7 in /Users/picocreator/Library/Python/3.9/lib/python/site-packages (from pandas) (2024.1)\n",
17
- "Requirement already satisfied: python-dateutil>=2.8.2 in /Users/picocreator/Library/Python/3.9/lib/python/site-packages (from pandas) (2.8.2)\n",
18
  "Requirement already satisfied: six>=1.5 in /Library/Developer/CommandLineTools/Library/Frameworks/Python3.framework/Versions/3.9/lib/python3.9/site-packages (from python-dateutil>=2.8.2->pandas) (1.15.0)\n",
19
  "\u001b[33mWARNING: You are using pip version 21.2.4; however, version 24.0 is available.\n",
20
  "You should consider upgrading via the '/Library/Developer/CommandLineTools/usr/bin/python3 -m pip install --upgrade pip' command.\u001b[0m\n"
@@ -36,14 +36,14 @@
36
  },
37
  {
38
  "cell_type": "code",
39
- "execution_count": 63,
40
  "metadata": {},
41
  "outputs": [
42
  {
43
  "name": "stdout",
44
  "output_type": "stream",
45
  "text": [
46
- "Found 2560 results.json files\n"
47
  ]
48
  }
49
  ],
@@ -71,7 +71,7 @@
71
  },
72
  {
73
  "cell_type": "code",
74
- "execution_count": 64,
75
  "metadata": {},
76
  "outputs": [
77
  {
@@ -156,16 +156,16 @@
156
  },
157
  {
158
  "cell_type": "code",
159
- "execution_count": 65,
160
  "metadata": {},
161
  "outputs": [
162
  {
163
  "name": "stdout",
164
  "output_type": "stream",
165
  "text": [
166
- "Found 44 models\n",
167
  "Models: \n",
168
- "['mistralai/Mistral-7B-v0.1', 'mosaicml/mpt-7b-instruct', 'mosaicml/mpt-7b', 'mosaicml/mpt-7b-chat', 'bigscience/bloom-7b1', 'bigscience/bloomz-7b1-mt', 'bigscience/bloomz-7b1', 'EleutherAI/pythia-2.8b', 'EleutherAI/pythia-1.4b', 'EleutherAI/gpt-j-6b', 'EleutherAI/pythia-6.9b', 'microsoft/phi-1_5', 'microsoft/phi-2', 'microsoft/phi-1', 'allenai/OLMo-7B', 'TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T', 'TinyLlama/TinyLlama-1.1B-Chat-v1.0', 'RWKV/rwkv-5-world-1b5', 'RWKV/rwkv-5-world-3b', 'RWKV/rwkv-4-world-3b', 'RWKV/rwkv-4-world-1b5', 'RWKV/v5-Eagle-7B-HF', 'RWKV/rwkv-4-world-7b', './rwkv-x-dev/chunk4-0_85_pth', './rwkv-x-dev/chunk0-0_8_pth', './rwkv-x-dev/RWKV-5-World-1B5-v2-20231025-ctx4096', './rwkv-x-dev/RWKV-5-World-3B-v2-20231118-ctx16k', './rwkv-x-dev/RWKV-5-World-7B-v2-20240128-ctx4096', './rwkv-x-dev/chunk6-0_85_pth', './rwkv-x-dev/chunk7-1-0_85_pth', './rwkv-x-dev/Hermes-RWKV-v5-7B_pth', 'togethercomputer/RedPajama-INCITE-7B-Base', 'togethercomputer/RedPajama-INCITE-7B-Instruct', 'togethercomputer/RedPajama-INCITE-7B-Chat', 'facebook/opt-2.7b', 'facebook/opt-6.7b', 'facebook/opt-1.3b', 'tiiuae/falcon-7b-instruct', 'tiiuae/falcon-rw-1b', 'tiiuae/falcon-rw-7b', 'tiiuae/falcon-7b', 'huggyllama/llama-7b', 'meta-llama/Llama-2-7b-chat-hf', 'meta-llama/Llama-2-7b-hf']\n",
169
  "Saved to compiled-lm-eval-results.json\n"
170
  ]
171
  }
@@ -199,7 +199,7 @@
199
  },
200
  {
201
  "cell_type": "code",
202
- "execution_count": 66,
203
  "metadata": {},
204
  "outputs": [
205
  {
@@ -321,6 +321,54 @@
321
  " </tr>\n",
322
  " <tr>\n",
323
  " <th>11</th>\n",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
324
  " <td>microsoft/phi-1_5</td>\n",
325
  " <td>0.521636</td>\n",
326
  " <td>0.026198</td>\n",
@@ -328,7 +376,7 @@
328
  " <td>0.026198</td>\n",
329
  " </tr>\n",
330
  " <tr>\n",
331
- " <th>12</th>\n",
332
  " <td>microsoft/phi-2</td>\n",
333
  " <td>0.512182</td>\n",
334
  " <td>0.029742</td>\n",
@@ -336,7 +384,7 @@
336
  " <td>0.029742</td>\n",
337
  " </tr>\n",
338
  " <tr>\n",
339
- " <th>13</th>\n",
340
  " <td>microsoft/phi-1</td>\n",
341
  " <td>0.517636</td>\n",
342
  " <td>0.029612</td>\n",
@@ -344,7 +392,7 @@
344
  " <td>0.029612</td>\n",
345
  " </tr>\n",
346
  " <tr>\n",
347
- " <th>14</th>\n",
348
  " <td>allenai/OLMo-7B</td>\n",
349
  " <td>0.537818</td>\n",
350
  " <td>0.034147</td>\n",
@@ -352,7 +400,7 @@
352
  " <td>0.034147</td>\n",
353
  " </tr>\n",
354
  " <tr>\n",
355
- " <th>15</th>\n",
356
  " <td>TinyLlama/TinyLlama-1.1B-intermediate-step-143...</td>\n",
357
  " <td>0.529273</td>\n",
358
  " <td>0.029316</td>\n",
@@ -360,7 +408,7 @@
360
  " <td>0.029316</td>\n",
361
  " </tr>\n",
362
  " <tr>\n",
363
- " <th>16</th>\n",
364
  " <td>TinyLlama/TinyLlama-1.1B-Chat-v1.0</td>\n",
365
  " <td>0.528909</td>\n",
366
  " <td>0.031702</td>\n",
@@ -368,23 +416,23 @@
368
  " <td>0.031702</td>\n",
369
  " </tr>\n",
370
  " <tr>\n",
371
- " <th>17</th>\n",
372
  " <td>RWKV/rwkv-5-world-1b5</td>\n",
373
- " <td>0.578909</td>\n",
374
- " <td>0.045103</td>\n",
375
- " <td>0.578909</td>\n",
376
- " <td>0.045103</td>\n",
377
  " </tr>\n",
378
  " <tr>\n",
379
- " <th>18</th>\n",
380
  " <td>RWKV/rwkv-5-world-3b</td>\n",
381
- " <td>0.590182</td>\n",
382
- " <td>0.056241</td>\n",
383
- " <td>0.590182</td>\n",
384
- " <td>0.056241</td>\n",
385
  " </tr>\n",
386
  " <tr>\n",
387
- " <th>19</th>\n",
388
  " <td>RWKV/rwkv-4-world-3b</td>\n",
389
  " <td>0.575455</td>\n",
390
  " <td>0.040977</td>\n",
@@ -392,7 +440,7 @@
392
  " <td>0.040977</td>\n",
393
  " </tr>\n",
394
  " <tr>\n",
395
- " <th>20</th>\n",
396
  " <td>RWKV/rwkv-4-world-1b5</td>\n",
397
  " <td>0.554000</td>\n",
398
  " <td>0.039406</td>\n",
@@ -400,15 +448,15 @@
400
  " <td>0.039406</td>\n",
401
  " </tr>\n",
402
  " <tr>\n",
403
- " <th>21</th>\n",
404
  " <td>RWKV/v5-Eagle-7B-HF</td>\n",
405
- " <td>0.621818</td>\n",
406
- " <td>0.068986</td>\n",
407
- " <td>0.621818</td>\n",
408
- " <td>0.068986</td>\n",
409
  " </tr>\n",
410
  " <tr>\n",
411
- " <th>22</th>\n",
412
  " <td>RWKV/rwkv-4-world-7b</td>\n",
413
  " <td>0.601455</td>\n",
414
  " <td>0.053116</td>\n",
@@ -416,7 +464,7 @@
416
  " <td>0.053116</td>\n",
417
  " </tr>\n",
418
  " <tr>\n",
419
- " <th>23</th>\n",
420
  " <td>togethercomputer/RedPajama-INCITE-7B-Base</td>\n",
421
  " <td>0.525455</td>\n",
422
  " <td>0.036407</td>\n",
@@ -424,7 +472,7 @@
424
  " <td>0.036407</td>\n",
425
  " </tr>\n",
426
  " <tr>\n",
427
- " <th>24</th>\n",
428
  " <td>togethercomputer/RedPajama-INCITE-7B-Instruct</td>\n",
429
  " <td>0.528545</td>\n",
430
  " <td>0.036470</td>\n",
@@ -432,7 +480,7 @@
432
  " <td>0.036470</td>\n",
433
  " </tr>\n",
434
  " <tr>\n",
435
- " <th>25</th>\n",
436
  " <td>togethercomputer/RedPajama-INCITE-7B-Chat</td>\n",
437
  " <td>0.535455</td>\n",
438
  " <td>0.038723</td>\n",
@@ -440,7 +488,7 @@
440
  " <td>0.038723</td>\n",
441
  " </tr>\n",
442
  " <tr>\n",
443
- " <th>26</th>\n",
444
  " <td>facebook/opt-2.7b</td>\n",
445
  " <td>0.521818</td>\n",
446
  " <td>0.029821</td>\n",
@@ -448,7 +496,7 @@
448
  " <td>0.029821</td>\n",
449
  " </tr>\n",
450
  " <tr>\n",
451
- " <th>27</th>\n",
452
  " <td>facebook/opt-6.7b</td>\n",
453
  " <td>0.522909</td>\n",
454
  " <td>0.027216</td>\n",
@@ -456,7 +504,7 @@
456
  " <td>0.027216</td>\n",
457
  " </tr>\n",
458
  " <tr>\n",
459
- " <th>28</th>\n",
460
  " <td>facebook/opt-1.3b</td>\n",
461
  " <td>0.521818</td>\n",
462
  " <td>0.029112</td>\n",
@@ -464,7 +512,7 @@
464
  " <td>0.029112</td>\n",
465
  " </tr>\n",
466
  " <tr>\n",
467
- " <th>29</th>\n",
468
  " <td>tiiuae/falcon-7b-instruct</td>\n",
469
  " <td>0.536727</td>\n",
470
  " <td>0.053430</td>\n",
@@ -472,7 +520,7 @@
472
  " <td>0.053430</td>\n",
473
  " </tr>\n",
474
  " <tr>\n",
475
- " <th>30</th>\n",
476
  " <td>tiiuae/falcon-rw-1b</td>\n",
477
  " <td>0.522545</td>\n",
478
  " <td>0.029446</td>\n",
@@ -480,7 +528,7 @@
480
  " <td>0.029446</td>\n",
481
  " </tr>\n",
482
  " <tr>\n",
483
- " <th>31</th>\n",
484
  " <td>tiiuae/falcon-rw-7b</td>\n",
485
  " <td>0.535818</td>\n",
486
  " <td>0.033185</td>\n",
@@ -488,7 +536,7 @@
488
  " <td>0.033185</td>\n",
489
  " </tr>\n",
490
  " <tr>\n",
491
- " <th>32</th>\n",
492
  " <td>tiiuae/falcon-7b</td>\n",
493
  " <td>0.559636</td>\n",
494
  " <td>0.071650</td>\n",
@@ -496,7 +544,7 @@
496
  " <td>0.071650</td>\n",
497
  " </tr>\n",
498
  " <tr>\n",
499
- " <th>33</th>\n",
500
  " <td>huggyllama/llama-7b</td>\n",
501
  " <td>0.541818</td>\n",
502
  " <td>0.040718</td>\n",
@@ -504,7 +552,7 @@
504
  " <td>0.040718</td>\n",
505
  " </tr>\n",
506
  " <tr>\n",
507
- " <th>34</th>\n",
508
  " <td>meta-llama/Llama-2-7b-chat-hf</td>\n",
509
  " <td>0.559818</td>\n",
510
  " <td>0.054954</td>\n",
@@ -512,7 +560,7 @@
512
  " <td>0.054954</td>\n",
513
  " </tr>\n",
514
  " <tr>\n",
515
- " <th>35</th>\n",
516
  " <td>meta-llama/Llama-2-7b-hf</td>\n",
517
  " <td>0.566727</td>\n",
518
  " <td>0.052515</td>\n",
@@ -536,31 +584,37 @@
536
  "8 EleutherAI/pythia-1.4b 0.526545 \n",
537
  "9 EleutherAI/gpt-j-6b 0.544182 \n",
538
  "10 EleutherAI/pythia-6.9b 0.540545 \n",
539
- "11 microsoft/phi-1_5 0.521636 \n",
540
- "12 microsoft/phi-2 0.512182 \n",
541
- "13 microsoft/phi-1 0.517636 \n",
542
- "14 allenai/OLMo-7B 0.537818 \n",
543
- "15 TinyLlama/TinyLlama-1.1B-intermediate-step-143... 0.529273 \n",
544
- "16 TinyLlama/TinyLlama-1.1B-Chat-v1.0 0.528909 \n",
545
- "17 RWKV/rwkv-5-world-1b5 0.578909 \n",
546
- "18 RWKV/rwkv-5-world-3b 0.590182 \n",
547
- "19 RWKV/rwkv-4-world-3b 0.575455 \n",
548
- "20 RWKV/rwkv-4-world-1b5 0.554000 \n",
549
- "21 RWKV/v5-Eagle-7B-HF 0.621818 \n",
550
- "22 RWKV/rwkv-4-world-7b 0.601455 \n",
551
- "23 togethercomputer/RedPajama-INCITE-7B-Base 0.525455 \n",
552
- "24 togethercomputer/RedPajama-INCITE-7B-Instruct 0.528545 \n",
553
- "25 togethercomputer/RedPajama-INCITE-7B-Chat 0.535455 \n",
554
- "26 facebook/opt-2.7b 0.521818 \n",
555
- "27 facebook/opt-6.7b 0.522909 \n",
556
- "28 facebook/opt-1.3b 0.521818 \n",
557
- "29 tiiuae/falcon-7b-instruct 0.536727 \n",
558
- "30 tiiuae/falcon-rw-1b 0.522545 \n",
559
- "31 tiiuae/falcon-rw-7b 0.535818 \n",
560
- "32 tiiuae/falcon-7b 0.559636 \n",
561
- "33 huggyllama/llama-7b 0.541818 \n",
562
- "34 meta-llama/Llama-2-7b-chat-hf 0.559818 \n",
563
- "35 meta-llama/Llama-2-7b-hf 0.566727 \n",
 
 
 
 
 
 
564
  "\n",
565
  " avg_acc_stderr xcopa (acc) xcopa (acc_stderr) \n",
566
  "0 0.053879 0.559455 0.053879 \n",
@@ -574,34 +628,40 @@
574
  "8 0.027441 0.526545 0.027441 \n",
575
  "9 0.034404 0.544182 0.034404 \n",
576
  "10 0.029689 0.540545 0.029689 \n",
577
- "11 0.026198 0.521636 0.026198 \n",
578
- "12 0.029742 0.512182 0.029742 \n",
579
- "13 0.029612 0.517636 0.029612 \n",
580
- "14 0.034147 0.537818 0.034147 \n",
581
- "15 0.029316 0.529273 0.029316 \n",
582
- "16 0.031702 0.528909 0.031702 \n",
583
- "17 0.045103 0.578909 0.045103 \n",
584
- "18 0.056241 0.590182 0.056241 \n",
585
- "19 0.040977 0.575455 0.040977 \n",
586
- "20 0.039406 0.554000 0.039406 \n",
587
- "21 0.068986 0.621818 0.068986 \n",
588
- "22 0.053116 0.601455 0.053116 \n",
589
- "23 0.036407 0.525455 0.036407 \n",
590
- "24 0.036470 0.528545 0.036470 \n",
591
- "25 0.038723 0.535455 0.038723 \n",
592
- "26 0.029821 0.521818 0.029821 \n",
593
- "27 0.027216 0.522909 0.027216 \n",
594
- "28 0.029112 0.521818 0.029112 \n",
595
- "29 0.053430 0.536727 0.053430 \n",
596
- "30 0.029446 0.522545 0.029446 \n",
597
- "31 0.033185 0.535818 0.033185 \n",
598
- "32 0.071650 0.559636 0.071650 \n",
599
- "33 0.040718 0.541818 0.040718 \n",
600
- "34 0.054954 0.559818 0.054954 \n",
601
- "35 0.052515 0.566727 0.052515 "
 
 
 
 
 
 
602
  ]
603
  },
604
- "execution_count": 66,
605
  "metadata": {},
606
  "output_type": "execute_result"
607
  }
@@ -791,27 +851,27 @@
791
  },
792
  {
793
  "cell_type": "code",
794
- "execution_count": 67,
795
  "metadata": {},
796
  "outputs": [
797
  {
798
  "name": "stdout",
799
  "output_type": "stream",
800
  "text": [
801
- "total 14936\n",
802
- "-rw-r--r--@ 1 picocreator staff 930K Feb 26 01:25 bf16-all-results-and-groups.csv\n",
803
- "-rw-r--r--@ 1 picocreator staff 60K Feb 26 01:25 bf16-eng-focus.csv\n",
804
- "-rw-r--r--@ 1 picocreator staff 847K Feb 26 01:25 bf16-eng-results.csv\n",
805
- "-rw-r--r--@ 1 picocreator staff 72K Feb 26 01:25 bf16-eng-summary.csv\n",
806
- "-rw-r--r--@ 1 picocreator staff 86K Feb 26 01:25 bf16-multilang-results.csv\n",
807
- "-rw-r--r--@ 1 picocreator staff 12K Feb 26 01:25 bf16-multilang-summary.csv\n",
808
- "-rw-r--r--@ 1 picocreator staff 60K Feb 26 01:25 bf16-sorted-eng-focus.csv\n",
809
- "-rw-r--r--@ 1 picocreator staff 847K Feb 26 01:25 bf16-sorted-eng-results.csv\n",
810
- "-rw-r--r--@ 1 picocreator staff 72K Feb 26 01:25 bf16-sorted-eng-summary.csv\n",
811
- "-rw-r--r--@ 1 picocreator staff 12K Feb 26 01:25 bf16-sorted-multilang-summary.csv\n",
812
- "-rw-r--r-- 1 picocreator staff 3.7M Feb 26 01:25 compiled-lm-eval-results.json\n",
813
- "-rw-r--r-- 1 picocreator staff 13K Feb 26 01:25 rwkv-x-dev-bf16-sorted-eng-focus.csv\n",
814
- "-rw-r--r--@ 1 picocreator staff 3.8K Feb 26 01:25 rwkv-x-dev-bf16-sorted-multilang-summary.csv\n"
815
  ]
816
  }
817
  ],
@@ -856,8 +916,8 @@
856
  "eng_grp_sorted.to_csv('summary/bf16-sorted-eng-summary.csv', index=False)\n",
857
  "\n",
858
  "# English focused subset\n",
859
- "eng_focus_tGrps=[\"anli\", \"glue\", \"truthfulqa\", \"lambada\", \"cmmlu\", \"pythia\", \"mmlu\", \"blimp\", \"trivaqa\", \"record\", \"np_open\", \"piqa\", \"copa\", \"sciq\"]\n",
860
- "eng_focus_tTest=[\"blimp\", \"arc_*\", \"logiqa\", \"winogrande\", \"openbookqa\", \"hellaswag\", \"blimp\", \"trivaqa\", \"record\", \"np_open\", \"piqa\", \"copa\", \"sciq\"]\n",
861
  "eng_focus = generate_result_table( inConfig = { \"dtype\": \"bfloat16\" }, inGroups=eng_focus_tGrps, inResults=eng_focus_tTest )\n",
862
  "eng_focus_sorted = generate_result_table( inConfig = { \"dtype\": \"bfloat16\" }, inGroups=eng_focus_tGrps, inResults=eng_focus_tTest, sort=True )\n",
863
  "eng_focus.to_csv('summary/bf16-eng-focus.csv', index=False)\n",
 
2
  "cells": [
3
  {
4
  "cell_type": "code",
5
+ "execution_count": 86,
6
  "metadata": {},
7
  "outputs": [
8
  {
 
11
  "text": [
12
  "Defaulting to user installation because normal site-packages is not writeable\n",
13
  "Requirement already satisfied: pandas in /Users/picocreator/Library/Python/3.9/lib/python/site-packages (2.2.0)\n",
14
+ "Requirement already satisfied: python-dateutil>=2.8.2 in /Users/picocreator/Library/Python/3.9/lib/python/site-packages (from pandas) (2.8.2)\n",
15
+ "Requirement already satisfied: tzdata>=2022.7 in /Users/picocreator/Library/Python/3.9/lib/python/site-packages (from pandas) (2024.1)\n",
16
  "Requirement already satisfied: pytz>=2020.1 in /Users/picocreator/Library/Python/3.9/lib/python/site-packages (from pandas) (2024.1)\n",
17
  "Requirement already satisfied: numpy<2,>=1.22.4 in /Users/picocreator/Library/Python/3.9/lib/python/site-packages (from pandas) (1.26.1)\n",
 
 
18
  "Requirement already satisfied: six>=1.5 in /Library/Developer/CommandLineTools/Library/Frameworks/Python3.framework/Versions/3.9/lib/python3.9/site-packages (from python-dateutil>=2.8.2->pandas) (1.15.0)\n",
19
  "\u001b[33mWARNING: You are using pip version 21.2.4; however, version 24.0 is available.\n",
20
  "You should consider upgrading via the '/Library/Developer/CommandLineTools/usr/bin/python3 -m pip install --upgrade pip' command.\u001b[0m\n"
 
36
  },
37
  {
38
  "cell_type": "code",
39
+ "execution_count": 87,
40
  "metadata": {},
41
  "outputs": [
42
  {
43
  "name": "stdout",
44
  "output_type": "stream",
45
  "text": [
46
+ "Found 2860 results.json files\n"
47
  ]
48
  }
49
  ],
 
71
  },
72
  {
73
  "cell_type": "code",
74
+ "execution_count": 88,
75
  "metadata": {},
76
  "outputs": [
77
  {
 
156
  },
157
  {
158
  "cell_type": "code",
159
+ "execution_count": 89,
160
  "metadata": {},
161
  "outputs": [
162
  {
163
  "name": "stdout",
164
  "output_type": "stream",
165
  "text": [
166
+ "Found 50 models\n",
167
  "Models: \n",
168
+ "['mistralai/Mistral-7B-v0.1', 'mosaicml/mpt-7b-instruct', 'mosaicml/mpt-7b', 'mosaicml/mpt-7b-chat', 'bigscience/bloom-7b1', 'bigscience/bloomz-7b1-mt', 'bigscience/bloomz-7b1', 'EleutherAI/pythia-2.8b', 'EleutherAI/pythia-1.4b', 'EleutherAI/gpt-j-6b', 'EleutherAI/pythia-6.9b', 'google/flan-t5-base', 'google/gemma-2b', 'google/gemma-2b-it', 'google/gemma-7b', 'google/gemma-7b-it', 'google/flan-t5-large', 'microsoft/phi-1_5', 'microsoft/phi-2', 'microsoft/phi-1', 'allenai/OLMo-7B', 'TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T', 'TinyLlama/TinyLlama-1.1B-Chat-v1.0', 'RWKV/rwkv-5-world-1b5', 'RWKV/rwkv-5-world-3b', 'RWKV/rwkv-4-world-3b', 'RWKV/rwkv-4-world-1b5', 'RWKV/v5-Eagle-7B-HF', 'RWKV/rwkv-4-world-7b', './rwkv-x-dev/chunk4-0_85_pth', './rwkv-x-dev/chunk0-0_8_pth', './rwkv-x-dev/RWKV-5-World-1B5-v2-20231025-ctx4096', './rwkv-x-dev/RWKV-5-World-3B-v2-20231118-ctx16k', './rwkv-x-dev/RWKV-5-World-7B-v2-20240128-ctx4096', './rwkv-x-dev/chunk6-0_85_pth', './rwkv-x-dev/chunk7-1-0_85_pth', './rwkv-x-dev/Hermes-RWKV-v5-7B_pth', 'togethercomputer/RedPajama-INCITE-7B-Base', 'togethercomputer/RedPajama-INCITE-7B-Instruct', 'togethercomputer/RedPajama-INCITE-7B-Chat', 'facebook/opt-2.7b', 'facebook/opt-6.7b', 'facebook/opt-1.3b', 'tiiuae/falcon-7b-instruct', 'tiiuae/falcon-rw-1b', 'tiiuae/falcon-rw-7b', 'tiiuae/falcon-7b', 'huggyllama/llama-7b', 'meta-llama/Llama-2-7b-chat-hf', 'meta-llama/Llama-2-7b-hf']\n",
169
  "Saved to compiled-lm-eval-results.json\n"
170
  ]
171
  }
 
199
  },
200
  {
201
  "cell_type": "code",
202
+ "execution_count": 90,
203
  "metadata": {},
204
  "outputs": [
205
  {
 
321
  " </tr>\n",
322
  " <tr>\n",
323
  " <th>11</th>\n",
324
+ " <td>google/flan-t5-base</td>\n",
325
+ " <td>0.510909</td>\n",
326
+ " <td>0.006743</td>\n",
327
+ " <td>0.510909</td>\n",
328
+ " <td>0.006743</td>\n",
329
+ " </tr>\n",
330
+ " <tr>\n",
331
+ " <th>12</th>\n",
332
+ " <td>google/gemma-2b</td>\n",
333
+ " <td>0.000000</td>\n",
334
+ " <td>0.000000</td>\n",
335
+ " <td>NaN</td>\n",
336
+ " <td>NaN</td>\n",
337
+ " </tr>\n",
338
+ " <tr>\n",
339
+ " <th>13</th>\n",
340
+ " <td>google/gemma-2b-it</td>\n",
341
+ " <td>0.000000</td>\n",
342
+ " <td>0.000000</td>\n",
343
+ " <td>NaN</td>\n",
344
+ " <td>NaN</td>\n",
345
+ " </tr>\n",
346
+ " <tr>\n",
347
+ " <th>14</th>\n",
348
+ " <td>google/gemma-7b</td>\n",
349
+ " <td>0.517636</td>\n",
350
+ " <td>0.006740</td>\n",
351
+ " <td>0.517636</td>\n",
352
+ " <td>0.006740</td>\n",
353
+ " </tr>\n",
354
+ " <tr>\n",
355
+ " <th>15</th>\n",
356
+ " <td>google/gemma-7b-it</td>\n",
357
+ " <td>0.517455</td>\n",
358
+ " <td>0.006742</td>\n",
359
+ " <td>0.517455</td>\n",
360
+ " <td>0.006742</td>\n",
361
+ " </tr>\n",
362
+ " <tr>\n",
363
+ " <th>16</th>\n",
364
+ " <td>google/flan-t5-large</td>\n",
365
+ " <td>0.510545</td>\n",
366
+ " <td>0.006743</td>\n",
367
+ " <td>0.510545</td>\n",
368
+ " <td>0.006743</td>\n",
369
+ " </tr>\n",
370
+ " <tr>\n",
371
+ " <th>17</th>\n",
372
  " <td>microsoft/phi-1_5</td>\n",
373
  " <td>0.521636</td>\n",
374
  " <td>0.026198</td>\n",
 
376
  " <td>0.026198</td>\n",
377
  " </tr>\n",
378
  " <tr>\n",
379
+ " <th>18</th>\n",
380
  " <td>microsoft/phi-2</td>\n",
381
  " <td>0.512182</td>\n",
382
  " <td>0.029742</td>\n",
 
384
  " <td>0.029742</td>\n",
385
  " </tr>\n",
386
  " <tr>\n",
387
+ " <th>19</th>\n",
388
  " <td>microsoft/phi-1</td>\n",
389
  " <td>0.517636</td>\n",
390
  " <td>0.029612</td>\n",
 
392
  " <td>0.029612</td>\n",
393
  " </tr>\n",
394
  " <tr>\n",
395
+ " <th>20</th>\n",
396
  " <td>allenai/OLMo-7B</td>\n",
397
  " <td>0.537818</td>\n",
398
  " <td>0.034147</td>\n",
 
400
  " <td>0.034147</td>\n",
401
  " </tr>\n",
402
  " <tr>\n",
403
+ " <th>21</th>\n",
404
  " <td>TinyLlama/TinyLlama-1.1B-intermediate-step-143...</td>\n",
405
  " <td>0.529273</td>\n",
406
  " <td>0.029316</td>\n",
 
408
  " <td>0.029316</td>\n",
409
  " </tr>\n",
410
  " <tr>\n",
411
+ " <th>22</th>\n",
412
  " <td>TinyLlama/TinyLlama-1.1B-Chat-v1.0</td>\n",
413
  " <td>0.528909</td>\n",
414
  " <td>0.031702</td>\n",
 
416
  " <td>0.031702</td>\n",
417
  " </tr>\n",
418
  " <tr>\n",
419
+ " <th>23</th>\n",
420
  " <td>RWKV/rwkv-5-world-1b5</td>\n",
421
+ " <td>0.579636</td>\n",
422
+ " <td>0.044978</td>\n",
423
+ " <td>0.579636</td>\n",
424
+ " <td>0.044978</td>\n",
425
  " </tr>\n",
426
  " <tr>\n",
427
+ " <th>24</th>\n",
428
  " <td>RWKV/rwkv-5-world-3b</td>\n",
429
+ " <td>0.590000</td>\n",
430
+ " <td>0.057252</td>\n",
431
+ " <td>0.590000</td>\n",
432
+ " <td>0.057252</td>\n",
433
  " </tr>\n",
434
  " <tr>\n",
435
+ " <th>25</th>\n",
436
  " <td>RWKV/rwkv-4-world-3b</td>\n",
437
  " <td>0.575455</td>\n",
438
  " <td>0.040977</td>\n",
 
440
  " <td>0.040977</td>\n",
441
  " </tr>\n",
442
  " <tr>\n",
443
+ " <th>26</th>\n",
444
  " <td>RWKV/rwkv-4-world-1b5</td>\n",
445
  " <td>0.554000</td>\n",
446
  " <td>0.039406</td>\n",
 
448
  " <td>0.039406</td>\n",
449
  " </tr>\n",
450
  " <tr>\n",
451
+ " <th>27</th>\n",
452
  " <td>RWKV/v5-Eagle-7B-HF</td>\n",
453
+ " <td>0.622364</td>\n",
454
+ " <td>0.070563</td>\n",
455
+ " <td>0.622364</td>\n",
456
+ " <td>0.070563</td>\n",
457
  " </tr>\n",
458
  " <tr>\n",
459
+ " <th>28</th>\n",
460
  " <td>RWKV/rwkv-4-world-7b</td>\n",
461
  " <td>0.601455</td>\n",
462
  " <td>0.053116</td>\n",
 
464
  " <td>0.053116</td>\n",
465
  " </tr>\n",
466
  " <tr>\n",
467
+ " <th>29</th>\n",
468
  " <td>togethercomputer/RedPajama-INCITE-7B-Base</td>\n",
469
  " <td>0.525455</td>\n",
470
  " <td>0.036407</td>\n",
 
472
  " <td>0.036407</td>\n",
473
  " </tr>\n",
474
  " <tr>\n",
475
+ " <th>30</th>\n",
476
  " <td>togethercomputer/RedPajama-INCITE-7B-Instruct</td>\n",
477
  " <td>0.528545</td>\n",
478
  " <td>0.036470</td>\n",
 
480
  " <td>0.036470</td>\n",
481
  " </tr>\n",
482
  " <tr>\n",
483
+ " <th>31</th>\n",
484
  " <td>togethercomputer/RedPajama-INCITE-7B-Chat</td>\n",
485
  " <td>0.535455</td>\n",
486
  " <td>0.038723</td>\n",
 
488
  " <td>0.038723</td>\n",
489
  " </tr>\n",
490
  " <tr>\n",
491
+ " <th>32</th>\n",
492
  " <td>facebook/opt-2.7b</td>\n",
493
  " <td>0.521818</td>\n",
494
  " <td>0.029821</td>\n",
 
496
  " <td>0.029821</td>\n",
497
  " </tr>\n",
498
  " <tr>\n",
499
+ " <th>33</th>\n",
500
  " <td>facebook/opt-6.7b</td>\n",
501
  " <td>0.522909</td>\n",
502
  " <td>0.027216</td>\n",
 
504
  " <td>0.027216</td>\n",
505
  " </tr>\n",
506
  " <tr>\n",
507
+ " <th>34</th>\n",
508
  " <td>facebook/opt-1.3b</td>\n",
509
  " <td>0.521818</td>\n",
510
  " <td>0.029112</td>\n",
 
512
  " <td>0.029112</td>\n",
513
  " </tr>\n",
514
  " <tr>\n",
515
+ " <th>35</th>\n",
516
  " <td>tiiuae/falcon-7b-instruct</td>\n",
517
  " <td>0.536727</td>\n",
518
  " <td>0.053430</td>\n",
 
520
  " <td>0.053430</td>\n",
521
  " </tr>\n",
522
  " <tr>\n",
523
+ " <th>36</th>\n",
524
  " <td>tiiuae/falcon-rw-1b</td>\n",
525
  " <td>0.522545</td>\n",
526
  " <td>0.029446</td>\n",
 
528
  " <td>0.029446</td>\n",
529
  " </tr>\n",
530
  " <tr>\n",
531
+ " <th>37</th>\n",
532
  " <td>tiiuae/falcon-rw-7b</td>\n",
533
  " <td>0.535818</td>\n",
534
  " <td>0.033185</td>\n",
 
536
  " <td>0.033185</td>\n",
537
  " </tr>\n",
538
  " <tr>\n",
539
+ " <th>38</th>\n",
540
  " <td>tiiuae/falcon-7b</td>\n",
541
  " <td>0.559636</td>\n",
542
  " <td>0.071650</td>\n",
 
544
  " <td>0.071650</td>\n",
545
  " </tr>\n",
546
  " <tr>\n",
547
+ " <th>39</th>\n",
548
  " <td>huggyllama/llama-7b</td>\n",
549
  " <td>0.541818</td>\n",
550
  " <td>0.040718</td>\n",
 
552
  " <td>0.040718</td>\n",
553
  " </tr>\n",
554
  " <tr>\n",
555
+ " <th>40</th>\n",
556
  " <td>meta-llama/Llama-2-7b-chat-hf</td>\n",
557
  " <td>0.559818</td>\n",
558
  " <td>0.054954</td>\n",
 
560
  " <td>0.054954</td>\n",
561
  " </tr>\n",
562
  " <tr>\n",
563
+ " <th>41</th>\n",
564
  " <td>meta-llama/Llama-2-7b-hf</td>\n",
565
  " <td>0.566727</td>\n",
566
  " <td>0.052515</td>\n",
 
584
  "8 EleutherAI/pythia-1.4b 0.526545 \n",
585
  "9 EleutherAI/gpt-j-6b 0.544182 \n",
586
  "10 EleutherAI/pythia-6.9b 0.540545 \n",
587
+ "11 google/flan-t5-base 0.510909 \n",
588
+ "12 google/gemma-2b 0.000000 \n",
589
+ "13 google/gemma-2b-it 0.000000 \n",
590
+ "14 google/gemma-7b 0.517636 \n",
591
+ "15 google/gemma-7b-it 0.517455 \n",
592
+ "16 google/flan-t5-large 0.510545 \n",
593
+ "17 microsoft/phi-1_5 0.521636 \n",
594
+ "18 microsoft/phi-2 0.512182 \n",
595
+ "19 microsoft/phi-1 0.517636 \n",
596
+ "20 allenai/OLMo-7B 0.537818 \n",
597
+ "21 TinyLlama/TinyLlama-1.1B-intermediate-step-143... 0.529273 \n",
598
+ "22 TinyLlama/TinyLlama-1.1B-Chat-v1.0 0.528909 \n",
599
+ "23 RWKV/rwkv-5-world-1b5 0.579636 \n",
600
+ "24 RWKV/rwkv-5-world-3b 0.590000 \n",
601
+ "25 RWKV/rwkv-4-world-3b 0.575455 \n",
602
+ "26 RWKV/rwkv-4-world-1b5 0.554000 \n",
603
+ "27 RWKV/v5-Eagle-7B-HF 0.622364 \n",
604
+ "28 RWKV/rwkv-4-world-7b 0.601455 \n",
605
+ "29 togethercomputer/RedPajama-INCITE-7B-Base 0.525455 \n",
606
+ "30 togethercomputer/RedPajama-INCITE-7B-Instruct 0.528545 \n",
607
+ "31 togethercomputer/RedPajama-INCITE-7B-Chat 0.535455 \n",
608
+ "32 facebook/opt-2.7b 0.521818 \n",
609
+ "33 facebook/opt-6.7b 0.522909 \n",
610
+ "34 facebook/opt-1.3b 0.521818 \n",
611
+ "35 tiiuae/falcon-7b-instruct 0.536727 \n",
612
+ "36 tiiuae/falcon-rw-1b 0.522545 \n",
613
+ "37 tiiuae/falcon-rw-7b 0.535818 \n",
614
+ "38 tiiuae/falcon-7b 0.559636 \n",
615
+ "39 huggyllama/llama-7b 0.541818 \n",
616
+ "40 meta-llama/Llama-2-7b-chat-hf 0.559818 \n",
617
+ "41 meta-llama/Llama-2-7b-hf 0.566727 \n",
618
  "\n",
619
  " avg_acc_stderr xcopa (acc) xcopa (acc_stderr) \n",
620
  "0 0.053879 0.559455 0.053879 \n",
 
628
  "8 0.027441 0.526545 0.027441 \n",
629
  "9 0.034404 0.544182 0.034404 \n",
630
  "10 0.029689 0.540545 0.029689 \n",
631
+ "11 0.006743 0.510909 0.006743 \n",
632
+ "12 0.000000 NaN NaN \n",
633
+ "13 0.000000 NaN NaN \n",
634
+ "14 0.006740 0.517636 0.006740 \n",
635
+ "15 0.006742 0.517455 0.006742 \n",
636
+ "16 0.006743 0.510545 0.006743 \n",
637
+ "17 0.026198 0.521636 0.026198 \n",
638
+ "18 0.029742 0.512182 0.029742 \n",
639
+ "19 0.029612 0.517636 0.029612 \n",
640
+ "20 0.034147 0.537818 0.034147 \n",
641
+ "21 0.029316 0.529273 0.029316 \n",
642
+ "22 0.031702 0.528909 0.031702 \n",
643
+ "23 0.044978 0.579636 0.044978 \n",
644
+ "24 0.057252 0.590000 0.057252 \n",
645
+ "25 0.040977 0.575455 0.040977 \n",
646
+ "26 0.039406 0.554000 0.039406 \n",
647
+ "27 0.070563 0.622364 0.070563 \n",
648
+ "28 0.053116 0.601455 0.053116 \n",
649
+ "29 0.036407 0.525455 0.036407 \n",
650
+ "30 0.036470 0.528545 0.036470 \n",
651
+ "31 0.038723 0.535455 0.038723 \n",
652
+ "32 0.029821 0.521818 0.029821 \n",
653
+ "33 0.027216 0.522909 0.027216 \n",
654
+ "34 0.029112 0.521818 0.029112 \n",
655
+ "35 0.053430 0.536727 0.053430 \n",
656
+ "36 0.029446 0.522545 0.029446 \n",
657
+ "37 0.033185 0.535818 0.033185 \n",
658
+ "38 0.071650 0.559636 0.071650 \n",
659
+ "39 0.040718 0.541818 0.040718 \n",
660
+ "40 0.054954 0.559818 0.054954 \n",
661
+ "41 0.052515 0.566727 0.052515 "
662
  ]
663
  },
664
+ "execution_count": 90,
665
  "metadata": {},
666
  "output_type": "execute_result"
667
  }
 
851
  },
852
  {
853
  "cell_type": "code",
854
+ "execution_count": 91,
855
  "metadata": {},
856
  "outputs": [
857
  {
858
  "name": "stdout",
859
  "output_type": "stream",
860
  "text": [
861
+ "total 17392\n",
862
+ "-rw-r--r--@ 1 picocreator staff 1.0M Feb 27 00:57 bf16-all-results-and-groups.csv\n",
863
+ "-rw-r--r--@ 1 picocreator staff 64K Feb 27 00:57 bf16-eng-focus.csv\n",
864
+ "-rw-r--r--@ 1 picocreator staff 920K Feb 27 00:57 bf16-eng-results.csv\n",
865
+ "-rw-r--r--@ 1 picocreator staff 77K Feb 27 00:57 bf16-eng-summary.csv\n",
866
+ "-rw-r--r--@ 1 picocreator staff 96K Feb 27 00:57 bf16-multilang-results.csv\n",
867
+ "-rw-r--r--@ 1 picocreator staff 14K Feb 27 00:57 bf16-multilang-summary.csv\n",
868
+ "-rw-r--r--@ 1 picocreator staff 64K Feb 27 00:57 bf16-sorted-eng-focus.csv\n",
869
+ "-rw-r--r--@ 1 picocreator staff 920K Feb 27 00:57 bf16-sorted-eng-results.csv\n",
870
+ "-rw-r--r--@ 1 picocreator staff 77K Feb 27 00:57 bf16-sorted-eng-summary.csv\n",
871
+ "-rw-r--r--@ 1 picocreator staff 14K Feb 27 00:57 bf16-sorted-multilang-summary.csv\n",
872
+ "-rw-r--r-- 1 picocreator staff 4.2M Feb 27 00:57 compiled-lm-eval-results.json\n",
873
+ "-rw-r--r--@ 1 picocreator staff 19K Feb 27 00:57 rwkv-x-dev-bf16-sorted-eng-focus.csv\n",
874
+ "-rw-r--r--@ 1 picocreator staff 3.8K Feb 27 00:57 rwkv-x-dev-bf16-sorted-multilang-summary.csv\n"
875
  ]
876
  }
877
  ],
 
916
  "eng_grp_sorted.to_csv('summary/bf16-sorted-eng-summary.csv', index=False)\n",
917
  "\n",
918
  "# English focused subset\n",
919
+ "eng_focus_tGrps=[\"anli\", \"glue\", \"truthfulqa\", \"lambada\", \"cmmlu\", \"pythia\", \"mmlu\", \"blimp\", \"record\", \"np_open\", \"piqa\", \"copa\", \"sciq\"]\n",
920
+ "eng_focus_tTest=[\"blimp\", \"arc_*\", \"logiqa\", \"winogrande\", \"openbookqa\", \"hellaswag\", \"blimp\", \"record\", \"np_open\", \"piqa\", \"copa\", \"sciq\"]\n",
921
  "eng_focus = generate_result_table( inConfig = { \"dtype\": \"bfloat16\" }, inGroups=eng_focus_tGrps, inResults=eng_focus_tTest )\n",
922
  "eng_focus_sorted = generate_result_table( inConfig = { \"dtype\": \"bfloat16\" }, inGroups=eng_focus_tGrps, inResults=eng_focus_tTest, sort=True )\n",
923
  "eng_focus.to_csv('summary/bf16-eng-focus.csv', index=False)\n",
summary/compiled-lm-eval-results.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d50a1f30815b2039a41a1636614a91fa82caab43d3fd5b49c3b17c55b9749d14
3
- size 4400826
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b42a2b7acf8133ba3099dd13da01aa0acba7bbc4d0e9ebca786caa5a29221cf
3
+ size 4780982
summary/rwkv-x-dev-bf16-sorted-eng-focus.csv CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:28dfd9bfb569b87880c5f7b1c526ffafc083c7fa29f9a8c4c646950e3e396569
3
- size 19955
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22168c9e4572c7a764a046b9cdedcf327172be5cecb4fd77159582ebdca8cd14
3
+ size 30506
summary/rwkv-x-dev-bf16-sorted-multilang-summary.csv CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fa3d5ee42037a49355ae0f4ab6c1c3c8f9e12e9191118ebf0634a9c80c0ebcb7
3
- size 3891
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:291794d3b95575a8036ab216c0727438add8899c0a59d284e24284884d940b9d
3
+ size 5904