csabakecskemeti commited on
Commit
219b2b6
1 Parent(s): ec98a4a

Upload 2 files

Browse files
results_2024-11-30T09-11-17.191351.json ADDED
@@ -0,0 +1,200 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "leaderboard_musr": {
4
+ " ": " ",
5
+ "alias": "leaderboard_musr"
6
+ },
7
+ "leaderboard_musr_murder_mysteries": {
8
+ "alias": " - leaderboard_musr_murder_mysteries",
9
+ "acc_norm,none": 0.54,
10
+ "acc_norm_stderr,none": 0.03158465389149902
11
+ },
12
+ "leaderboard_musr_object_placements": {
13
+ "alias": " - leaderboard_musr_object_placements",
14
+ "acc_norm,none": 0.234375,
15
+ "acc_norm_stderr,none": 0.02652733398834892
16
+ },
17
+ "leaderboard_musr_team_allocation": {
18
+ "alias": " - leaderboard_musr_team_allocation",
19
+ "acc_norm,none": 0.32,
20
+ "acc_norm_stderr,none": 0.029561724955241033
21
+ }
22
+ },
23
+ "group_subtasks": {
24
+ "leaderboard_musr": [
25
+ "leaderboard_musr_murder_mysteries",
26
+ "leaderboard_musr_object_placements",
27
+ "leaderboard_musr_team_allocation"
28
+ ]
29
+ },
30
+ "configs": {
31
+ "leaderboard_musr_murder_mysteries": {
32
+ "task": "leaderboard_musr_murder_mysteries",
33
+ "dataset_path": "TAUR-Lab/MuSR",
34
+ "test_split": "murder_mysteries",
35
+ "doc_to_text": "def doc_to_text(doc):\n \"\"\"\n Convert a doc to text.\n \"\"\"\n choices = \"\"\n for i, choice in enumerate(ast.literal_eval(doc[\"choices\"])):\n choices += f\"{i+1} - {choice}\\n\"\n\n text = DOC_TO_TEXT.format(\n narrative=doc[\"narrative\"], question=doc[\"question\"], choices=choices\n )\n\n return text\n",
36
+ "doc_to_target": "{{answer_choice}}",
37
+ "doc_to_choice": "{{choices}}",
38
+ "description": "",
39
+ "target_delimiter": " ",
40
+ "fewshot_delimiter": "\n\n",
41
+ "num_fewshot": 0,
42
+ "metric_list": [
43
+ {
44
+ "metric": "acc_norm",
45
+ "aggregation": "mean",
46
+ "higher_is_better": true
47
+ }
48
+ ],
49
+ "output_type": "multiple_choice",
50
+ "repeats": 1,
51
+ "should_decontaminate": false,
52
+ "metadata": {
53
+ "version": 1.0
54
+ }
55
+ },
56
+ "leaderboard_musr_object_placements": {
57
+ "task": "leaderboard_musr_object_placements",
58
+ "dataset_path": "TAUR-Lab/MuSR",
59
+ "test_split": "object_placements",
60
+ "doc_to_text": "def doc_to_text(doc):\n \"\"\"\n Convert a doc to text.\n \"\"\"\n choices = \"\"\n for i, choice in enumerate(ast.literal_eval(doc[\"choices\"])):\n choices += f\"{i+1} - {choice}\\n\"\n\n text = DOC_TO_TEXT.format(\n narrative=doc[\"narrative\"], question=doc[\"question\"], choices=choices\n )\n\n return text\n",
61
+ "doc_to_target": "{{answer_choice}}",
62
+ "doc_to_choice": "{{choices}}",
63
+ "description": "",
64
+ "target_delimiter": " ",
65
+ "fewshot_delimiter": "\n\n",
66
+ "num_fewshot": 0,
67
+ "metric_list": [
68
+ {
69
+ "metric": "acc_norm",
70
+ "aggregation": "mean",
71
+ "higher_is_better": true
72
+ }
73
+ ],
74
+ "output_type": "multiple_choice",
75
+ "repeats": 1,
76
+ "should_decontaminate": false,
77
+ "metadata": {
78
+ "version": 1.0
79
+ }
80
+ },
81
+ "leaderboard_musr_team_allocation": {
82
+ "task": "leaderboard_musr_team_allocation",
83
+ "dataset_path": "TAUR-Lab/MuSR",
84
+ "test_split": "team_allocation",
85
+ "doc_to_text": "def doc_to_text(doc):\n \"\"\"\n Convert a doc to text.\n \"\"\"\n choices = \"\"\n for i, choice in enumerate(ast.literal_eval(doc[\"choices\"])):\n choices += f\"{i+1} - {choice}\\n\"\n\n text = DOC_TO_TEXT.format(\n narrative=doc[\"narrative\"], question=doc[\"question\"], choices=choices\n )\n\n return text\n",
86
+ "doc_to_target": "{{answer_choice}}",
87
+ "doc_to_choice": "{{choices}}",
88
+ "description": "",
89
+ "target_delimiter": " ",
90
+ "fewshot_delimiter": "\n\n",
91
+ "num_fewshot": 0,
92
+ "metric_list": [
93
+ {
94
+ "metric": "acc_norm",
95
+ "aggregation": "mean",
96
+ "higher_is_better": true
97
+ }
98
+ ],
99
+ "output_type": "multiple_choice",
100
+ "repeats": 1,
101
+ "should_decontaminate": false,
102
+ "metadata": {
103
+ "version": 1.0
104
+ }
105
+ }
106
+ },
107
+ "versions": {
108
+ "leaderboard_musr_murder_mysteries": 1.0,
109
+ "leaderboard_musr_object_placements": 1.0,
110
+ "leaderboard_musr_team_allocation": 1.0
111
+ },
112
+ "n-shot": {
113
+ "leaderboard_musr_murder_mysteries": 0,
114
+ "leaderboard_musr_object_placements": 0,
115
+ "leaderboard_musr_team_allocation": 0
116
+ },
117
+ "higher_is_better": {
118
+ "leaderboard_musr": {
119
+ "acc_norm": true
120
+ },
121
+ "leaderboard_musr_murder_mysteries": {
122
+ "acc_norm": true
123
+ },
124
+ "leaderboard_musr_object_placements": {
125
+ "acc_norm": true
126
+ },
127
+ "leaderboard_musr_team_allocation": {
128
+ "acc_norm": true
129
+ }
130
+ },
131
+ "n-samples": {
132
+ "leaderboard_musr_murder_mysteries": {
133
+ "original": 250,
134
+ "effective": 250
135
+ },
136
+ "leaderboard_musr_object_placements": {
137
+ "original": 256,
138
+ "effective": 256
139
+ },
140
+ "leaderboard_musr_team_allocation": {
141
+ "original": 250,
142
+ "effective": 250
143
+ }
144
+ },
145
+ "config": {
146
+ "model": "hf",
147
+ "model_args": "pretrained=DevQuasar/analytical_reasoning_r16a32_unsloth-Llama-3.2-3B-Instruct-bnb-4bit",
148
+ "batch_size": "auto:4",
149
+ "batch_sizes": [
150
+ 16,
151
+ 16,
152
+ 16,
153
+ 32
154
+ ],
155
+ "device": null,
156
+ "use_cache": "eval_cache",
157
+ "limit": null,
158
+ "bootstrap_iters": 100000,
159
+ "gen_kwargs": null,
160
+ "random_seed": 0,
161
+ "numpy_seed": 1234,
162
+ "torch_seed": 1234,
163
+ "fewshot_seed": 1234
164
+ },
165
+ "git_hash": "0230356",
166
+ "date": 1732986471.4917576,
167
+ "pretty_env_info": "PyTorch version: 2.5.1+cu124\nIs debug build: False\nCUDA used to build PyTorch: 12.4\nROCM used to build PyTorch: N/A\n\nOS: Debian GNU/Linux 12 (bookworm) (x86_64)\nGCC version: (Debian 12.2.0-14) 12.2.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.36\n\nPython version: 3.11.10 (main, Oct 3 2024, 07:29:13) [GCC 11.2.0] (64-bit runtime)\nPython platform: Linux-6.1.0-26-amd64-x86_64-with-glibc2.36\nIs CUDA available: True\nCUDA runtime version: Could not collect\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: \nGPU 0: NVIDIA GeForce GTX 1050 Ti\nGPU 1: Tesla P40\nGPU 2: Tesla V100-PCIE-32GB\nGPU 3: Tesla V100-PCIE-32GB\n\nNvidia driver version: 535.183.01\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nAddress sizes: 43 bits physical, 48 bits virtual\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nVendor ID: AuthenticAMD\nModel name: AMD Ryzen Threadripper 1950X 16-Core Processor\nCPU family: 23\nModel: 1\nThread(s) per core: 2\nCore(s) per socket: 16\nSocket(s): 1\nStepping: 1\nFrequency boost: enabled\nCPU(s) scaling MHz: 66%\nCPU max MHz: 3400.0000\nCPU min MHz: 2200.0000\nBogoMIPS: 6786.43\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc rep_good nopl nonstop_tsc cpuid extd_apicid amd_dcm aperfmperf rapl pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_llc mwaitx cpb hw_pstate ssbd ibpb vmmcall fsgsbase bmi1 avx2 smep bmi2 rdseed adx smap clflushopt sha_ni xsaveopt xsavec xgetbv1 clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif overflow_recov succor smca sev\nVirtualization: AMD-V\nL1d cache: 512 KiB (16 instances)\nL1i cache: 1 MiB (16 instances)\nL2 cache: 8 MiB (16 instances)\nL3 cache: 32 MiB (4 instances)\nNUMA node(s): 1\nNUMA node0 CPU(s): 0-31\nVulnerability Gather data sampling: Not affected\nVulnerability Itlb multihit: Not affected\nVulnerability L1tf: Not affected\nVulnerability Mds: Not affected\nVulnerability Meltdown: Not affected\nVulnerability Mmio stale data: Not affected\nVulnerability Reg file data sampling: Not affected\nVulnerability Retbleed: Mitigation; untrained return thunk; SMT vulnerable\nVulnerability Spec rstack overflow: Mitigation; safe RET\nVulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl\nVulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization\nVulnerability Spectre v2: Mitigation; Retpolines; IBPB conditional; STIBP disabled; RSB filling; PBRSB-eIBRS Not affected; BHI Not affected\nVulnerability Srbds: Not affected\nVulnerability Tsx async abort: Not affected\n\nVersions of relevant libraries:\n[pip3] numpy==2.1.3\n[pip3] torch==2.5.1\n[pip3] triton==3.1.0\n[conda] numpy 2.1.3 pypi_0 pypi\n[conda] torch 2.5.1 pypi_0 pypi\n[conda] triton 3.1.0 pypi_0 pypi",
168
+ "transformers_version": "4.46.3",
169
+ "upper_git_hash": null,
170
+ "tokenizer_pad_token": [
171
+ "<|finetune_right_pad_id|>",
172
+ "128004"
173
+ ],
174
+ "tokenizer_eos_token": [
175
+ "<|eot_id|>",
176
+ "128009"
177
+ ],
178
+ "tokenizer_bos_token": [
179
+ "<|begin_of_text|>",
180
+ "128000"
181
+ ],
182
+ "eot_token_id": 128009,
183
+ "max_length": 131072,
184
+ "task_hashes": {
185
+ "leaderboard_musr_murder_mysteries": "a696259562ea5c5c09a2613e30526fae1de29f55da9e28e8d7e8a53027e6d330",
186
+ "leaderboard_musr_object_placements": "3aa8c5e5bc59cd6ba2326269b9f0bf3cee8cba1b4e9e1d1330cf5f1f59ea0dce",
187
+ "leaderboard_musr_team_allocation": "5a75f135c145ee861a1cf31b63346709ef41b9d542be6a61c5818c210a3797a5"
188
+ },
189
+ "model_source": "hf",
190
+ "model_name": "DevQuasar/analytical_reasoning_r16a32_unsloth-Llama-3.2-3B-Instruct-bnb-4bit",
191
+ "model_name_sanitized": "DevQuasar__analytical_reasoning_r16a32_unsloth-Llama-3.2-3B-Instruct-bnb-4bit",
192
+ "system_instruction": null,
193
+ "system_instruction_sha": null,
194
+ "fewshot_as_multiturn": false,
195
+ "chat_template": null,
196
+ "chat_template_sha": null,
197
+ "start_time": 52195.45405349,
198
+ "end_time": 52407.302247922,
199
+ "total_evaluation_time_seconds": "211.84819443200104"
200
+ }
results_2024-11-30T13-44-48.543919.json ADDED
The diff for this file is too large to render. See raw diff