Update README.md
Browse files
README.md
CHANGED
@@ -14,4 +14,85 @@ tags:
|
|
14 |
|
15 |
# Uploaded model
|
16 |
|
17 |
-
base model pretrained on 2.5 epochs on no-prompt-oasst
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
14 |
|
15 |
# Uploaded model
|
16 |
|
17 |
+
base model pretrained on 2.5 epochs on no-prompt-oasst
|
18 |
+
|
19 |
+
hf (pretrained=appvoid/palmer-004-turbo-v1.2), gen_kwargs: (None), limit: None, num_fewshot: None, batch_size: 8
|
20 |
+
| Tasks |Version|Filter|n-shot| Metric | |Value | |Stderr|
|
21 |
+
|---------------------------------------|------:|------|-----:|--------|---|-----:|---|-----:|
|
22 |
+
|arc_challenge | 1|none | 0|acc |↑ |0.3097|± |0.0135|
|
23 |
+
| | |none | 0|acc_norm|↑ |0.3464|± |0.0139|
|
24 |
+
|hellaswag | 1|none | 0|acc |↑ |0.4660|± |0.0050|
|
25 |
+
| | |none | 0|acc_norm|↑ |0.6130|± |0.0049|
|
26 |
+
|mmlu | 1|none | |acc |↑ |0.2728|± |0.0037|
|
27 |
+
| - humanities | 1|none | |acc |↑ |0.2593|± |0.0064|
|
28 |
+
| - formal_logic | 0|none | 0|acc |↑ |0.3175|± |0.0416|
|
29 |
+
| - high_school_european_history | 0|none | 0|acc |↑ |0.2970|± |0.0357|
|
30 |
+
| - high_school_us_history | 0|none | 0|acc |↑ |0.2941|± |0.0320|
|
31 |
+
| - high_school_world_history | 0|none | 0|acc |↑ |0.2489|± |0.0281|
|
32 |
+
| - international_law | 0|none | 0|acc |↑ |0.3306|± |0.0429|
|
33 |
+
| - jurisprudence | 0|none | 0|acc |↑ |0.2685|± |0.0428|
|
34 |
+
| - logical_fallacies | 0|none | 0|acc |↑ |0.2515|± |0.0341|
|
35 |
+
| - moral_disputes | 0|none | 0|acc |↑ |0.2486|± |0.0233|
|
36 |
+
| - moral_scenarios | 0|none | 0|acc |↑ |0.2380|± |0.0142|
|
37 |
+
| - philosophy | 0|none | 0|acc |↑ |0.2797|± |0.0255|
|
38 |
+
| - prehistory | 0|none | 0|acc |↑ |0.2593|± |0.0244|
|
39 |
+
| - professional_law | 0|none | 0|acc |↑ |0.2555|± |0.0111|
|
40 |
+
| - world_religions | 0|none | 0|acc |↑ |0.2339|± |0.0325|
|
41 |
+
| - other | 1|none | |acc |↑ |0.2813|± |0.0080|
|
42 |
+
| - business_ethics | 0|none | 0|acc |↑ |0.1700|± |0.0378|
|
43 |
+
| - clinical_knowledge | 0|none | 0|acc |↑ |0.3434|± |0.0292|
|
44 |
+
| - college_medicine | 0|none | 0|acc |↑ |0.3006|± |0.0350|
|
45 |
+
| - global_facts | 0|none | 0|acc |↑ |0.3600|± |0.0482|
|
46 |
+
| - human_aging | 0|none | 0|acc |↑ |0.1659|± |0.0250|
|
47 |
+
| - management | 0|none | 0|acc |↑ |0.3883|± |0.0483|
|
48 |
+
| - marketing | 0|none | 0|acc |↑ |0.2137|± |0.0269|
|
49 |
+
| - medical_genetics | 0|none | 0|acc |↑ |0.2500|± |0.0435|
|
50 |
+
| - miscellaneous | 0|none | 0|acc |↑ |0.2784|± |0.0160|
|
51 |
+
| - nutrition | 0|none | 0|acc |↑ |0.2386|± |0.0244|
|
52 |
+
| - professional_accounting | 0|none | 0|acc |↑ |0.2376|± |0.0254|
|
53 |
+
| - professional_medicine | 0|none | 0|acc |↑ |0.4265|± |0.0300|
|
54 |
+
| - virology | 0|none | 0|acc |↑ |0.3133|± |0.0361|
|
55 |
+
| - social sciences | 1|none | |acc |↑ |0.2847|± |0.0081|
|
56 |
+
| - econometrics | 0|none | 0|acc |↑ |0.2719|± |0.0419|
|
57 |
+
| - high_school_geography | 0|none | 0|acc |↑ |0.3788|± |0.0346|
|
58 |
+
| - high_school_government_and_politics| 0|none | 0|acc |↑ |0.2850|± |0.0326|
|
59 |
+
| - high_school_macroeconomics | 0|none | 0|acc |↑ |0.3359|± |0.0239|
|
60 |
+
| - high_school_microeconomics | 0|none | 0|acc |↑ |0.3361|± |0.0307|
|
61 |
+
| - high_school_psychology | 0|none | 0|acc |↑ |0.3064|± |0.0198|
|
62 |
+
| - human_sexuality | 0|none | 0|acc |↑ |0.2519|± |0.0381|
|
63 |
+
| - professional_psychology | 0|none | 0|acc |↑ |0.2190|± |0.0167|
|
64 |
+
| - public_relations | 0|none | 0|acc |↑ |0.3091|± |0.0443|
|
65 |
+
| - security_studies | 0|none | 0|acc |↑ |0.2449|± |0.0275|
|
66 |
+
| - sociology | 0|none | 0|acc |↑ |0.2388|± |0.0301|
|
67 |
+
| - us_foreign_policy | 0|none | 0|acc |↑ |0.2800|± |0.0451|
|
68 |
+
| - stem | 1|none | |acc |↑ |0.2731|± |0.0079|
|
69 |
+
| - abstract_algebra | 0|none | 0|acc |↑ |0.1900|± |0.0394|
|
70 |
+
| - anatomy | 0|none | 0|acc |↑ |0.1926|± |0.0341|
|
71 |
+
| - astronomy | 0|none | 0|acc |↑ |0.2829|± |0.0367|
|
72 |
+
| - college_biology | 0|none | 0|acc |↑ |0.2083|± |0.0340|
|
73 |
+
| - college_chemistry | 0|none | 0|acc |↑ |0.3400|± |0.0476|
|
74 |
+
| - college_computer_science | 0|none | 0|acc |↑ |0.3000|± |0.0461|
|
75 |
+
| - college_mathematics | 0|none | 0|acc |↑ |0.3100|± |0.0465|
|
76 |
+
| - college_physics | 0|none | 0|acc |↑ |0.2941|± |0.0453|
|
77 |
+
| - computer_security | 0|none | 0|acc |↑ |0.2400|± |0.0429|
|
78 |
+
| - conceptual_physics | 0|none | 0|acc |↑ |0.2553|± |0.0285|
|
79 |
+
| - electrical_engineering | 0|none | 0|acc |↑ |0.2828|± |0.0375|
|
80 |
+
| - elementary_mathematics | 0|none | 0|acc |↑ |0.2513|± |0.0223|
|
81 |
+
| - high_school_biology | 0|none | 0|acc |↑ |0.2935|± |0.0259|
|
82 |
+
| - high_school_chemistry | 0|none | 0|acc |↑ |0.2808|± |0.0316|
|
83 |
+
| - high_school_computer_science | 0|none | 0|acc |↑ |0.2400|± |0.0429|
|
84 |
+
| - high_school_mathematics | 0|none | 0|acc |↑ |0.2741|± |0.0272|
|
85 |
+
| - high_school_physics | 0|none | 0|acc |↑ |0.3775|± |0.0396|
|
86 |
+
| - high_school_statistics | 0|none | 0|acc |↑ |0.3194|± |0.0318|
|
87 |
+
| - machine_learning | 0|none | 0|acc |↑ |0.2321|± |0.0401|
|
88 |
+
|piqa | 1|none | 0|acc |↑ |0.7269|± |0.0104|
|
89 |
+
| | |none | 0|acc_norm|↑ |0.7242|± |0.0104|
|
90 |
+
|winogrande | 1|none | 0|acc |↑ |0.5848|± |0.0138|
|
91 |
+
|
92 |
+
| Groups |Version|Filter|n-shot|Metric| |Value | |Stderr|
|
93 |
+
|------------------|------:|------|------|------|---|-----:|---|-----:|
|
94 |
+
|mmlu | 1|none | |acc |↑ |0.2728|± |0.0037|
|
95 |
+
| - humanities | 1|none | |acc |↑ |0.2593|± |0.0064|
|
96 |
+
| - other | 1|none | |acc |↑ |0.2813|± |0.0080|
|
97 |
+
| - social sciences| 1|none | |acc |↑ |0.2847|± |0.0081|
|
98 |
+
| - stem | 1|none | |acc |↑ |0.2731|± |0.0079|
|