ggbetz commited on
Commit
42edd49
1 Parent(s): 63be100

Delete data/mistralai/Mistral-7B-Instruct-v0.2/orig/results_24-01-28-20:06:10.json

Browse files
data/mistralai/Mistral-7B-Instruct-v0.2/orig/results_24-01-28-20:06:10.json DELETED
@@ -1,90 +0,0 @@
1
- {
2
- "results": {
3
- "logiqa_base": {
4
- "acc,none": 0.2987220447284345,
5
- "acc_stderr,none": 0.01830790800596066,
6
- "alias": "logiqa_base"
7
- },
8
- "lsat-lr_base": {
9
- "acc,none": 0.2803921568627451,
10
- "acc_stderr,none": 0.01991003317147411,
11
- "alias": "lsat-lr_base"
12
- }
13
- },
14
- "configs": {
15
- "logiqa_base": {
16
- "task": "logiqa_base",
17
- "group": "logikon-bench",
18
- "dataset_path": "logikon/logikon-bench",
19
- "dataset_name": "logiqa",
20
- "test_split": "test",
21
- "doc_to_text": "<function doc_to_text at 0x7f18885f1c60>",
22
- "doc_to_target": "{{answer}}",
23
- "doc_to_choice": "{{options}}",
24
- "description": "",
25
- "target_delimiter": " ",
26
- "fewshot_delimiter": "\n\n",
27
- "num_fewshot": 0,
28
- "metric_list": [
29
- {
30
- "metric": "acc",
31
- "aggregation": "mean",
32
- "higher_is_better": true
33
- }
34
- ],
35
- "output_type": "multiple_choice",
36
- "repeats": 1,
37
- "should_decontaminate": false,
38
- "metadata": {
39
- "version": 0.0
40
- }
41
- },
42
- "lsat-lr_base": {
43
- "task": "lsat-lr_base",
44
- "group": "logikon-bench",
45
- "dataset_path": "logikon/logikon-bench",
46
- "dataset_name": "lsat-lr",
47
- "test_split": "test",
48
- "doc_to_text": "<function doc_to_text at 0x7f188879fd00>",
49
- "doc_to_target": "{{answer}}",
50
- "doc_to_choice": "{{options}}",
51
- "description": "",
52
- "target_delimiter": " ",
53
- "fewshot_delimiter": "\n\n",
54
- "num_fewshot": 0,
55
- "metric_list": [
56
- {
57
- "metric": "acc",
58
- "aggregation": "mean",
59
- "higher_is_better": true
60
- }
61
- ],
62
- "output_type": "multiple_choice",
63
- "repeats": 1,
64
- "should_decontaminate": false,
65
- "metadata": {
66
- "version": 0.0
67
- }
68
- }
69
- },
70
- "versions": {
71
- "logiqa_base": 0.0,
72
- "lsat-lr_base": 0.0
73
- },
74
- "n-shot": {
75
- "logiqa_base": 0,
76
- "lsat-lr_base": 0
77
- },
78
- "config": {
79
- "model": "vllm",
80
- "model_args": "pretrained=mistralai/Mistral-7B-Instruct-v0.2,revision=main,dtype=auto,gpu_memory_utilization=0.9,trust_remote_code=true,max_length=4096",
81
- "batch_size": "auto",
82
- "batch_sizes": [],
83
- "device": null,
84
- "use_cache": null,
85
- "limit": null,
86
- "bootstrap_iters": 100000,
87
- "gen_kwargs": null
88
- },
89
- "git_hash": "92b0637"
90
- }