open-ko-llm-bot commited on
Commit
7485c4c
1 Parent(s): 8584708

Add results for 2023-10-19 08:44:42

Browse files
choco9966/Llama-2-7b-instruct-tuning/result_2023-10-19 08:44:42.json CHANGED
@@ -7,10 +7,10 @@
7
  "acc_norm_stderr": 0.01378746032244138
8
  },
9
  "harness|ko_hellaswag|10": {
10
- "acc": 0.345947022505477,
11
- "acc_stderr": 0.004747038768172532,
12
- "acc_norm": 0.4251145190201155,
13
- "acc_norm_stderr": 0.004933500261683597
14
  },
15
  "harness|ko_mmlu_world_religions|5": {
16
  "acc": 0.4269005847953216,
@@ -355,10 +355,10 @@
355
  "acc_norm_stderr": 0.03793713171165634
356
  },
357
  "harness|ko_truthfulqa_mc|0": {
358
- "mc1": 0.33047735618115054,
359
- "mc1_stderr": 0.016466769613698293,
360
- "mc2": 0.5139753799906011,
361
- "mc2_stderr": 0.016082624616035393
362
  },
363
  "harness|ko_commongen_v2|2": {
364
  "acc": 0.2585596221959858,
 
7
  "acc_norm_stderr": 0.01378746032244138
8
  },
9
  "harness|ko_hellaswag|10": {
10
+ "acc": 0.3454491137223661,
11
+ "acc_stderr": 0.004745426656377564,
12
+ "acc_norm": 0.4245170284803824,
13
+ "acc_norm_stderr": 0.00493259334881363
14
  },
15
  "harness|ko_mmlu_world_religions|5": {
16
  "acc": 0.4269005847953216,
 
355
  "acc_norm_stderr": 0.03793713171165634
356
  },
357
  "harness|ko_truthfulqa_mc|0": {
358
+ "mc1": 0.3317013463892289,
359
+ "mc1_stderr": 0.01648214881024146,
360
+ "mc2": 0.5139810284158008,
361
+ "mc2_stderr": 0.01608284635540924
362
  },
363
  "harness|ko_commongen_v2|2": {
364
  "acc": 0.2585596221959858,