nvan15 commited on
Commit
ba22426
·
verified ·
1 Parent(s): d4b473b

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d18h10m47,sd43/ft/tokenizer.json +0 -0
  2. reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d18h10m47,sd43/ft/tokenizer_config.json +60 -0
  3. reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d18h10m47,sd43/ft2/README.md +205 -0
  4. reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d18h10m47,sd43/ft2/adapter_config.json +34 -0
  5. reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d21h00m07,sd44/SST-2.tsv +1822 -0
  6. reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d21h00m07,sd44/all_results.json +9 -0
  7. reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d21h00m07,sd44/eval_results.json +9 -0
  8. reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d21h00m07,sd44/ft/added_tokens.json +3 -0
  9. reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d21h00m07,sd44/ft/special_tokens_map.json +15 -0
  10. reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d21h00m07,sd44/ft/tokenizer.json +0 -0
  11. reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d21h00m07,sd44/ft/tokenizer_config.json +60 -0
  12. reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d21h00m07,sd44/ft2/README.md +205 -0
  13. reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d21h00m07,sd44/ft2/adapter_config.json +34 -0
  14. reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d21h00m07,sd44/trainer_state.json +1883 -0
  15. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=22d16h17m48/STS-B.tsv +1380 -0
  16. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=22d16h17m48/all_results.json +11 -0
  17. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=22d16h17m48/eval_results.json +11 -0
  18. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=22d16h17m48/ft/added_tokens.json +3 -0
  19. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=22d16h17m48/ft/special_tokens_map.json +15 -0
  20. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=22d16h17m48/ft/tokenizer.json +0 -0
  21. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=22d16h17m48/ft/tokenizer_config.json +60 -0
  22. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=22d16h17m48/ft2/README.md +205 -0
  23. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=22d16h17m48/ft2/adapter_config.json +34 -0
  24. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=22d16h17m48/trainer_state.json +169 -0
  25. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h35m28,sd43/STS-B.tsv +1380 -0
  26. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h35m28,sd43/all_results.json +11 -0
  27. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h35m28,sd43/eval_results.json +11 -0
  28. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h35m28,sd43/ft/added_tokens.json +3 -0
  29. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h35m28,sd43/ft/special_tokens_map.json +15 -0
  30. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h35m28,sd43/ft/tokenizer.json +0 -0
  31. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h35m28,sd43/ft/tokenizer_config.json +60 -0
  32. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h35m28,sd43/ft2/README.md +205 -0
  33. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h35m28,sd43/ft2/adapter_config.json +34 -0
  34. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h35m28,sd43/trainer_state.json +169 -0
  35. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h42m49,sd44/STS-B.tsv +1380 -0
  36. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h42m49,sd44/all_results.json +11 -0
  37. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h42m49,sd44/eval_results.json +11 -0
  38. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h42m49,sd44/ft/added_tokens.json +3 -0
  39. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h42m49,sd44/ft/special_tokens_map.json +15 -0
  40. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h42m49,sd44/ft/tokenizer.json +0 -0
  41. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h42m49,sd44/ft/tokenizer_config.json +60 -0
  42. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h42m49,sd44/ft2/README.md +205 -0
  43. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h42m49,sd44/ft2/adapter_config.json +34 -0
  44. reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h42m49,sd44/trainer_state.json +169 -0
  45. reproduction/glue_expHRA/cola/dr0.0,mlr9e-03,clr9e-03,ep=34.0t=24d12h04m24,sd43/CoLA.tsv +1064 -0
  46. reproduction/glue_expHRA/cola/dr0.0,mlr9e-03,clr9e-03,ep=34.0t=24d12h04m24,sd43/all_results.json +9 -0
  47. reproduction/glue_expHRA/cola/dr0.0,mlr9e-03,clr9e-03,ep=34.0t=24d12h04m24,sd43/eval_results.json +9 -0
  48. reproduction/glue_expHRA/cola/dr0.0,mlr9e-03,clr9e-03,ep=34.0t=24d12h04m24,sd43/ft/added_tokens.json +3 -0
  49. reproduction/glue_expHRA/cola/dr0.0,mlr9e-03,clr9e-03,ep=34.0t=24d12h04m24,sd43/ft/special_tokens_map.json +15 -0
  50. reproduction/glue_expHRA/cola/dr0.0,mlr9e-03,clr9e-03,ep=34.0t=24d12h04m24,sd43/ft/tokenizer.json +0 -0
reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d18h10m47,sd43/ft/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d18h10m47,sd43/ft/tokenizer_config.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[CLS]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[SEP]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[UNK]",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "128000": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "[CLS]",
45
+ "clean_up_tokenization_spaces": false,
46
+ "cls_token": "[CLS]",
47
+ "do_lower_case": false,
48
+ "eos_token": "[SEP]",
49
+ "extra_special_tokens": {},
50
+ "mask_token": "[MASK]",
51
+ "model_max_length": 512,
52
+ "pad_token": "[PAD]",
53
+ "padding_side": "right",
54
+ "sep_token": "[SEP]",
55
+ "sp_model_kwargs": {},
56
+ "split_by_punct": false,
57
+ "tokenizer_class": "DebertaV2Tokenizer",
58
+ "unk_token": "[UNK]",
59
+ "vocab_type": "spm"
60
+ }
reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d18h10m47,sd43/ft2/README.md ADDED
@@ -0,0 +1,205 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: microsoft/deberta-v3-base
3
+ library_name: peft
4
+ tags:
5
+ - base_model:adapter:microsoft/deberta-v3-base
6
+ - transformers
7
+ ---
8
+
9
+ # Model Card for Model ID
10
+
11
+ <!-- Provide a quick summary of what the model is/does. -->
12
+
13
+
14
+
15
+ ## Model Details
16
+
17
+ ### Model Description
18
+
19
+ <!-- Provide a longer summary of what this model is. -->
20
+
21
+
22
+
23
+ - **Developed by:** [More Information Needed]
24
+ - **Funded by [optional]:** [More Information Needed]
25
+ - **Shared by [optional]:** [More Information Needed]
26
+ - **Model type:** [More Information Needed]
27
+ - **Language(s) (NLP):** [More Information Needed]
28
+ - **License:** [More Information Needed]
29
+ - **Finetuned from model [optional]:** [More Information Needed]
30
+
31
+ ### Model Sources [optional]
32
+
33
+ <!-- Provide the basic links for the model. -->
34
+
35
+ - **Repository:** [More Information Needed]
36
+ - **Paper [optional]:** [More Information Needed]
37
+ - **Demo [optional]:** [More Information Needed]
38
+
39
+ ## Uses
40
+
41
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
42
+
43
+ ### Direct Use
44
+
45
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
46
+
47
+ [More Information Needed]
48
+
49
+ ### Downstream Use [optional]
50
+
51
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
52
+
53
+ [More Information Needed]
54
+
55
+ ### Out-of-Scope Use
56
+
57
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
58
+
59
+ [More Information Needed]
60
+
61
+ ## Bias, Risks, and Limitations
62
+
63
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
64
+
65
+ [More Information Needed]
66
+
67
+ ### Recommendations
68
+
69
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
70
+
71
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
72
+
73
+ ## How to Get Started with the Model
74
+
75
+ Use the code below to get started with the model.
76
+
77
+ [More Information Needed]
78
+
79
+ ## Training Details
80
+
81
+ ### Training Data
82
+
83
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
84
+
85
+ [More Information Needed]
86
+
87
+ ### Training Procedure
88
+
89
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
90
+
91
+ #### Preprocessing [optional]
92
+
93
+ [More Information Needed]
94
+
95
+
96
+ #### Training Hyperparameters
97
+
98
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
99
+
100
+ #### Speeds, Sizes, Times [optional]
101
+
102
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
103
+
104
+ [More Information Needed]
105
+
106
+ ## Evaluation
107
+
108
+ <!-- This section describes the evaluation protocols and provides the results. -->
109
+
110
+ ### Testing Data, Factors & Metrics
111
+
112
+ #### Testing Data
113
+
114
+ <!-- This should link to a Dataset Card if possible. -->
115
+
116
+ [More Information Needed]
117
+
118
+ #### Factors
119
+
120
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
121
+
122
+ [More Information Needed]
123
+
124
+ #### Metrics
125
+
126
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
127
+
128
+ [More Information Needed]
129
+
130
+ ### Results
131
+
132
+ [More Information Needed]
133
+
134
+ #### Summary
135
+
136
+
137
+
138
+ ## Model Examination [optional]
139
+
140
+ <!-- Relevant interpretability work for the model goes here -->
141
+
142
+ [More Information Needed]
143
+
144
+ ## Environmental Impact
145
+
146
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
147
+
148
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
149
+
150
+ - **Hardware Type:** [More Information Needed]
151
+ - **Hours used:** [More Information Needed]
152
+ - **Cloud Provider:** [More Information Needed]
153
+ - **Compute Region:** [More Information Needed]
154
+ - **Carbon Emitted:** [More Information Needed]
155
+
156
+ ## Technical Specifications [optional]
157
+
158
+ ### Model Architecture and Objective
159
+
160
+ [More Information Needed]
161
+
162
+ ### Compute Infrastructure
163
+
164
+ [More Information Needed]
165
+
166
+ #### Hardware
167
+
168
+ [More Information Needed]
169
+
170
+ #### Software
171
+
172
+ [More Information Needed]
173
+
174
+ ## Citation [optional]
175
+
176
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
177
+
178
+ **BibTeX:**
179
+
180
+ [More Information Needed]
181
+
182
+ **APA:**
183
+
184
+ [More Information Needed]
185
+
186
+ ## Glossary [optional]
187
+
188
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
189
+
190
+ [More Information Needed]
191
+
192
+ ## More Information [optional]
193
+
194
+ [More Information Needed]
195
+
196
+ ## Model Card Authors [optional]
197
+
198
+ [More Information Needed]
199
+
200
+ ## Model Card Contact
201
+
202
+ [More Information Needed]
203
+ ### Framework versions
204
+
205
+ - PEFT 0.18.0
reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d18h10m47,sd43/ft2/adapter_config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": {
3
+ "base_model_class": "DebertaV2ForSequenceClassification",
4
+ "parent_library": "transformers.models.deberta_v2.modeling_deberta_v2"
5
+ },
6
+ "base_model_name_or_path": "microsoft/deberta-v3-base",
7
+ "bias": "none",
8
+ "boft_block_num": 0,
9
+ "boft_block_size": 4,
10
+ "boft_dropout": 0.15,
11
+ "boft_n_butterfly_factor": 2,
12
+ "exclude_modules": null,
13
+ "fan_in_fan_out": false,
14
+ "inference_mode": true,
15
+ "init_weights": true,
16
+ "layers_pattern": null,
17
+ "layers_to_transform": null,
18
+ "modules_to_save": [
19
+ "classifier",
20
+ "pooler"
21
+ ],
22
+ "peft_type": "BOFT",
23
+ "peft_version": "0.18.0",
24
+ "revision": null,
25
+ "target_modules": [
26
+ "key_proj",
27
+ "value_proj",
28
+ "output.dense",
29
+ "attention.output.dense",
30
+ "query_proj",
31
+ "intermediate.dense"
32
+ ],
33
+ "task_type": null
34
+ }
reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d21h00m07,sd44/SST-2.tsv ADDED
@@ -0,0 +1,1822 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ index prediction
2
+ 0 0
3
+ 1 0
4
+ 2 1
5
+ 3 1
6
+ 4 1
7
+ 5 1
8
+ 6 0
9
+ 7 1
10
+ 8 1
11
+ 9 0
12
+ 10 0
13
+ 11 0
14
+ 12 0
15
+ 13 1
16
+ 14 1
17
+ 15 1
18
+ 16 1
19
+ 17 1
20
+ 18 1
21
+ 19 0
22
+ 20 0
23
+ 21 1
24
+ 22 0
25
+ 23 1
26
+ 24 0
27
+ 25 0
28
+ 26 0
29
+ 27 0
30
+ 28 1
31
+ 29 1
32
+ 30 1
33
+ 31 0
34
+ 32 0
35
+ 33 1
36
+ 34 0
37
+ 35 1
38
+ 36 0
39
+ 37 0
40
+ 38 0
41
+ 39 0
42
+ 40 1
43
+ 41 0
44
+ 42 1
45
+ 43 1
46
+ 44 0
47
+ 45 1
48
+ 46 1
49
+ 47 0
50
+ 48 0
51
+ 49 0
52
+ 50 0
53
+ 51 1
54
+ 52 0
55
+ 53 0
56
+ 54 1
57
+ 55 0
58
+ 56 0
59
+ 57 0
60
+ 58 0
61
+ 59 1
62
+ 60 1
63
+ 61 0
64
+ 62 1
65
+ 63 1
66
+ 64 1
67
+ 65 1
68
+ 66 1
69
+ 67 0
70
+ 68 0
71
+ 69 0
72
+ 70 0
73
+ 71 1
74
+ 72 1
75
+ 73 1
76
+ 74 0
77
+ 75 0
78
+ 76 0
79
+ 77 1
80
+ 78 0
81
+ 79 1
82
+ 80 0
83
+ 81 1
84
+ 82 1
85
+ 83 1
86
+ 84 1
87
+ 85 1
88
+ 86 0
89
+ 87 0
90
+ 88 0
91
+ 89 0
92
+ 90 1
93
+ 91 1
94
+ 92 0
95
+ 93 1
96
+ 94 1
97
+ 95 1
98
+ 96 1
99
+ 97 1
100
+ 98 0
101
+ 99 0
102
+ 100 0
103
+ 101 0
104
+ 102 0
105
+ 103 1
106
+ 104 1
107
+ 105 0
108
+ 106 1
109
+ 107 1
110
+ 108 0
111
+ 109 0
112
+ 110 0
113
+ 111 1
114
+ 112 0
115
+ 113 1
116
+ 114 1
117
+ 115 1
118
+ 116 1
119
+ 117 0
120
+ 118 0
121
+ 119 1
122
+ 120 1
123
+ 121 0
124
+ 122 1
125
+ 123 0
126
+ 124 1
127
+ 125 1
128
+ 126 1
129
+ 127 0
130
+ 128 0
131
+ 129 1
132
+ 130 1
133
+ 131 0
134
+ 132 1
135
+ 133 1
136
+ 134 0
137
+ 135 0
138
+ 136 1
139
+ 137 1
140
+ 138 1
141
+ 139 1
142
+ 140 1
143
+ 141 1
144
+ 142 0
145
+ 143 1
146
+ 144 1
147
+ 145 1
148
+ 146 1
149
+ 147 0
150
+ 148 0
151
+ 149 0
152
+ 150 1
153
+ 151 0
154
+ 152 1
155
+ 153 0
156
+ 154 0
157
+ 155 1
158
+ 156 1
159
+ 157 1
160
+ 158 1
161
+ 159 1
162
+ 160 1
163
+ 161 1
164
+ 162 0
165
+ 163 0
166
+ 164 0
167
+ 165 0
168
+ 166 0
169
+ 167 1
170
+ 168 1
171
+ 169 1
172
+ 170 1
173
+ 171 1
174
+ 172 0
175
+ 173 0
176
+ 174 0
177
+ 175 0
178
+ 176 1
179
+ 177 0
180
+ 178 1
181
+ 179 1
182
+ 180 1
183
+ 181 0
184
+ 182 1
185
+ 183 1
186
+ 184 0
187
+ 185 1
188
+ 186 1
189
+ 187 1
190
+ 188 0
191
+ 189 1
192
+ 190 1
193
+ 191 0
194
+ 192 1
195
+ 193 0
196
+ 194 0
197
+ 195 0
198
+ 196 1
199
+ 197 1
200
+ 198 1
201
+ 199 0
202
+ 200 0
203
+ 201 0
204
+ 202 0
205
+ 203 0
206
+ 204 1
207
+ 205 1
208
+ 206 1
209
+ 207 1
210
+ 208 0
211
+ 209 0
212
+ 210 1
213
+ 211 1
214
+ 212 1
215
+ 213 0
216
+ 214 1
217
+ 215 1
218
+ 216 1
219
+ 217 0
220
+ 218 1
221
+ 219 1
222
+ 220 0
223
+ 221 1
224
+ 222 1
225
+ 223 0
226
+ 224 0
227
+ 225 1
228
+ 226 0
229
+ 227 1
230
+ 228 1
231
+ 229 1
232
+ 230 0
233
+ 231 0
234
+ 232 0
235
+ 233 0
236
+ 234 0
237
+ 235 0
238
+ 236 0
239
+ 237 0
240
+ 238 0
241
+ 239 0
242
+ 240 0
243
+ 241 0
244
+ 242 1
245
+ 243 1
246
+ 244 1
247
+ 245 0
248
+ 246 0
249
+ 247 0
250
+ 248 0
251
+ 249 1
252
+ 250 0
253
+ 251 0
254
+ 252 1
255
+ 253 0
256
+ 254 1
257
+ 255 0
258
+ 256 0
259
+ 257 0
260
+ 258 0
261
+ 259 0
262
+ 260 0
263
+ 261 0
264
+ 262 0
265
+ 263 0
266
+ 264 1
267
+ 265 1
268
+ 266 1
269
+ 267 1
270
+ 268 1
271
+ 269 0
272
+ 270 0
273
+ 271 1
274
+ 272 1
275
+ 273 1
276
+ 274 0
277
+ 275 1
278
+ 276 0
279
+ 277 0
280
+ 278 0
281
+ 279 1
282
+ 280 1
283
+ 281 1
284
+ 282 0
285
+ 283 0
286
+ 284 1
287
+ 285 1
288
+ 286 1
289
+ 287 1
290
+ 288 1
291
+ 289 1
292
+ 290 0
293
+ 291 1
294
+ 292 0
295
+ 293 1
296
+ 294 0
297
+ 295 1
298
+ 296 0
299
+ 297 0
300
+ 298 1
301
+ 299 1
302
+ 300 0
303
+ 301 1
304
+ 302 1
305
+ 303 1
306
+ 304 0
307
+ 305 1
308
+ 306 0
309
+ 307 1
310
+ 308 1
311
+ 309 1
312
+ 310 0
313
+ 311 1
314
+ 312 1
315
+ 313 0
316
+ 314 0
317
+ 315 0
318
+ 316 1
319
+ 317 1
320
+ 318 1
321
+ 319 0
322
+ 320 0
323
+ 321 0
324
+ 322 0
325
+ 323 0
326
+ 324 1
327
+ 325 1
328
+ 326 0
329
+ 327 1
330
+ 328 1
331
+ 329 1
332
+ 330 1
333
+ 331 1
334
+ 332 0
335
+ 333 0
336
+ 334 0
337
+ 335 1
338
+ 336 0
339
+ 337 1
340
+ 338 1
341
+ 339 0
342
+ 340 1
343
+ 341 0
344
+ 342 0
345
+ 343 1
346
+ 344 0
347
+ 345 0
348
+ 346 1
349
+ 347 1
350
+ 348 1
351
+ 349 0
352
+ 350 1
353
+ 351 0
354
+ 352 1
355
+ 353 0
356
+ 354 1
357
+ 355 1
358
+ 356 1
359
+ 357 1
360
+ 358 0
361
+ 359 0
362
+ 360 1
363
+ 361 1
364
+ 362 1
365
+ 363 0
366
+ 364 1
367
+ 365 0
368
+ 366 1
369
+ 367 1
370
+ 368 1
371
+ 369 0
372
+ 370 0
373
+ 371 0
374
+ 372 0
375
+ 373 1
376
+ 374 0
377
+ 375 0
378
+ 376 1
379
+ 377 0
380
+ 378 0
381
+ 379 1
382
+ 380 0
383
+ 381 0
384
+ 382 0
385
+ 383 1
386
+ 384 1
387
+ 385 0
388
+ 386 1
389
+ 387 0
390
+ 388 1
391
+ 389 0
392
+ 390 1
393
+ 391 1
394
+ 392 1
395
+ 393 1
396
+ 394 1
397
+ 395 0
398
+ 396 0
399
+ 397 0
400
+ 398 0
401
+ 399 0
402
+ 400 1
403
+ 401 1
404
+ 402 1
405
+ 403 1
406
+ 404 0
407
+ 405 1
408
+ 406 1
409
+ 407 1
410
+ 408 0
411
+ 409 0
412
+ 410 0
413
+ 411 0
414
+ 412 0
415
+ 413 1
416
+ 414 0
417
+ 415 0
418
+ 416 1
419
+ 417 1
420
+ 418 1
421
+ 419 1
422
+ 420 0
423
+ 421 1
424
+ 422 1
425
+ 423 0
426
+ 424 0
427
+ 425 0
428
+ 426 1
429
+ 427 1
430
+ 428 1
431
+ 429 1
432
+ 430 0
433
+ 431 1
434
+ 432 0
435
+ 433 0
436
+ 434 1
437
+ 435 0
438
+ 436 0
439
+ 437 1
440
+ 438 0
441
+ 439 0
442
+ 440 1
443
+ 441 1
444
+ 442 1
445
+ 443 0
446
+ 444 1
447
+ 445 0
448
+ 446 0
449
+ 447 0
450
+ 448 1
451
+ 449 1
452
+ 450 1
453
+ 451 1
454
+ 452 1
455
+ 453 0
456
+ 454 1
457
+ 455 0
458
+ 456 1
459
+ 457 0
460
+ 458 1
461
+ 459 0
462
+ 460 0
463
+ 461 1
464
+ 462 0
465
+ 463 0
466
+ 464 1
467
+ 465 0
468
+ 466 0
469
+ 467 1
470
+ 468 0
471
+ 469 1
472
+ 470 1
473
+ 471 0
474
+ 472 0
475
+ 473 1
476
+ 474 0
477
+ 475 0
478
+ 476 0
479
+ 477 1
480
+ 478 0
481
+ 479 1
482
+ 480 1
483
+ 481 1
484
+ 482 0
485
+ 483 1
486
+ 484 0
487
+ 485 0
488
+ 486 0
489
+ 487 0
490
+ 488 0
491
+ 489 1
492
+ 490 1
493
+ 491 1
494
+ 492 0
495
+ 493 1
496
+ 494 1
497
+ 495 1
498
+ 496 0
499
+ 497 1
500
+ 498 1
501
+ 499 0
502
+ 500 0
503
+ 501 0
504
+ 502 0
505
+ 503 0
506
+ 504 1
507
+ 505 0
508
+ 506 1
509
+ 507 1
510
+ 508 1
511
+ 509 0
512
+ 510 0
513
+ 511 0
514
+ 512 0
515
+ 513 0
516
+ 514 0
517
+ 515 1
518
+ 516 1
519
+ 517 1
520
+ 518 1
521
+ 519 0
522
+ 520 0
523
+ 521 1
524
+ 522 0
525
+ 523 0
526
+ 524 0
527
+ 525 1
528
+ 526 0
529
+ 527 1
530
+ 528 0
531
+ 529 1
532
+ 530 0
533
+ 531 0
534
+ 532 0
535
+ 533 1
536
+ 534 0
537
+ 535 1
538
+ 536 1
539
+ 537 1
540
+ 538 0
541
+ 539 0
542
+ 540 1
543
+ 541 0
544
+ 542 1
545
+ 543 1
546
+ 544 0
547
+ 545 1
548
+ 546 1
549
+ 547 0
550
+ 548 1
551
+ 549 0
552
+ 550 0
553
+ 551 0
554
+ 552 0
555
+ 553 1
556
+ 554 1
557
+ 555 1
558
+ 556 0
559
+ 557 1
560
+ 558 0
561
+ 559 0
562
+ 560 1
563
+ 561 0
564
+ 562 1
565
+ 563 0
566
+ 564 0
567
+ 565 0
568
+ 566 0
569
+ 567 1
570
+ 568 1
571
+ 569 1
572
+ 570 0
573
+ 571 0
574
+ 572 1
575
+ 573 1
576
+ 574 1
577
+ 575 0
578
+ 576 1
579
+ 577 1
580
+ 578 0
581
+ 579 0
582
+ 580 0
583
+ 581 1
584
+ 582 1
585
+ 583 0
586
+ 584 1
587
+ 585 1
588
+ 586 0
589
+ 587 0
590
+ 588 1
591
+ 589 1
592
+ 590 0
593
+ 591 1
594
+ 592 0
595
+ 593 0
596
+ 594 0
597
+ 595 0
598
+ 596 1
599
+ 597 0
600
+ 598 1
601
+ 599 0
602
+ 600 0
603
+ 601 0
604
+ 602 1
605
+ 603 1
606
+ 604 1
607
+ 605 1
608
+ 606 0
609
+ 607 0
610
+ 608 1
611
+ 609 1
612
+ 610 1
613
+ 611 0
614
+ 612 1
615
+ 613 1
616
+ 614 1
617
+ 615 0
618
+ 616 0
619
+ 617 1
620
+ 618 0
621
+ 619 1
622
+ 620 0
623
+ 621 1
624
+ 622 1
625
+ 623 0
626
+ 624 1
627
+ 625 0
628
+ 626 1
629
+ 627 1
630
+ 628 1
631
+ 629 1
632
+ 630 0
633
+ 631 0
634
+ 632 0
635
+ 633 1
636
+ 634 1
637
+ 635 0
638
+ 636 0
639
+ 637 1
640
+ 638 1
641
+ 639 1
642
+ 640 0
643
+ 641 1
644
+ 642 1
645
+ 643 1
646
+ 644 0
647
+ 645 0
648
+ 646 1
649
+ 647 0
650
+ 648 1
651
+ 649 1
652
+ 650 0
653
+ 651 0
654
+ 652 1
655
+ 653 1
656
+ 654 1
657
+ 655 1
658
+ 656 1
659
+ 657 1
660
+ 658 0
661
+ 659 1
662
+ 660 1
663
+ 661 0
664
+ 662 0
665
+ 663 0
666
+ 664 0
667
+ 665 1
668
+ 666 0
669
+ 667 1
670
+ 668 1
671
+ 669 0
672
+ 670 0
673
+ 671 1
674
+ 672 1
675
+ 673 0
676
+ 674 1
677
+ 675 1
678
+ 676 0
679
+ 677 1
680
+ 678 0
681
+ 679 1
682
+ 680 0
683
+ 681 1
684
+ 682 1
685
+ 683 0
686
+ 684 0
687
+ 685 1
688
+ 686 0
689
+ 687 1
690
+ 688 0
691
+ 689 1
692
+ 690 1
693
+ 691 0
694
+ 692 0
695
+ 693 1
696
+ 694 0
697
+ 695 1
698
+ 696 1
699
+ 697 0
700
+ 698 0
701
+ 699 1
702
+ 700 0
703
+ 701 0
704
+ 702 1
705
+ 703 0
706
+ 704 1
707
+ 705 1
708
+ 706 1
709
+ 707 0
710
+ 708 1
711
+ 709 0
712
+ 710 1
713
+ 711 1
714
+ 712 0
715
+ 713 1
716
+ 714 0
717
+ 715 0
718
+ 716 1
719
+ 717 0
720
+ 718 1
721
+ 719 1
722
+ 720 0
723
+ 721 1
724
+ 722 1
725
+ 723 1
726
+ 724 1
727
+ 725 1
728
+ 726 1
729
+ 727 0
730
+ 728 0
731
+ 729 0
732
+ 730 0
733
+ 731 1
734
+ 732 0
735
+ 733 1
736
+ 734 0
737
+ 735 0
738
+ 736 0
739
+ 737 1
740
+ 738 0
741
+ 739 0
742
+ 740 1
743
+ 741 1
744
+ 742 0
745
+ 743 1
746
+ 744 1
747
+ 745 1
748
+ 746 0
749
+ 747 1
750
+ 748 1
751
+ 749 1
752
+ 750 1
753
+ 751 0
754
+ 752 0
755
+ 753 1
756
+ 754 0
757
+ 755 1
758
+ 756 0
759
+ 757 0
760
+ 758 1
761
+ 759 1
762
+ 760 0
763
+ 761 0
764
+ 762 0
765
+ 763 1
766
+ 764 1
767
+ 765 0
768
+ 766 0
769
+ 767 0
770
+ 768 1
771
+ 769 1
772
+ 770 1
773
+ 771 0
774
+ 772 1
775
+ 773 1
776
+ 774 1
777
+ 775 1
778
+ 776 1
779
+ 777 1
780
+ 778 0
781
+ 779 1
782
+ 780 1
783
+ 781 0
784
+ 782 0
785
+ 783 0
786
+ 784 1
787
+ 785 0
788
+ 786 0
789
+ 787 0
790
+ 788 1
791
+ 789 0
792
+ 790 1
793
+ 791 1
794
+ 792 0
795
+ 793 0
796
+ 794 0
797
+ 795 0
798
+ 796 1
799
+ 797 1
800
+ 798 0
801
+ 799 1
802
+ 800 0
803
+ 801 1
804
+ 802 0
805
+ 803 1
806
+ 804 1
807
+ 805 0
808
+ 806 1
809
+ 807 0
810
+ 808 1
811
+ 809 1
812
+ 810 0
813
+ 811 0
814
+ 812 1
815
+ 813 0
816
+ 814 0
817
+ 815 1
818
+ 816 1
819
+ 817 1
820
+ 818 1
821
+ 819 1
822
+ 820 0
823
+ 821 0
824
+ 822 0
825
+ 823 0
826
+ 824 1
827
+ 825 0
828
+ 826 1
829
+ 827 0
830
+ 828 0
831
+ 829 0
832
+ 830 0
833
+ 831 0
834
+ 832 0
835
+ 833 0
836
+ 834 0
837
+ 835 0
838
+ 836 0
839
+ 837 1
840
+ 838 0
841
+ 839 0
842
+ 840 1
843
+ 841 1
844
+ 842 1
845
+ 843 0
846
+ 844 0
847
+ 845 1
848
+ 846 1
849
+ 847 0
850
+ 848 0
851
+ 849 1
852
+ 850 1
853
+ 851 1
854
+ 852 0
855
+ 853 1
856
+ 854 1
857
+ 855 1
858
+ 856 0
859
+ 857 1
860
+ 858 1
861
+ 859 0
862
+ 860 1
863
+ 861 0
864
+ 862 0
865
+ 863 0
866
+ 864 1
867
+ 865 0
868
+ 866 1
869
+ 867 1
870
+ 868 0
871
+ 869 0
872
+ 870 1
873
+ 871 1
874
+ 872 0
875
+ 873 0
876
+ 874 1
877
+ 875 0
878
+ 876 1
879
+ 877 1
880
+ 878 0
881
+ 879 0
882
+ 880 0
883
+ 881 0
884
+ 882 0
885
+ 883 1
886
+ 884 1
887
+ 885 1
888
+ 886 0
889
+ 887 0
890
+ 888 0
891
+ 889 1
892
+ 890 0
893
+ 891 1
894
+ 892 1
895
+ 893 0
896
+ 894 1
897
+ 895 1
898
+ 896 0
899
+ 897 0
900
+ 898 0
901
+ 899 1
902
+ 900 0
903
+ 901 0
904
+ 902 1
905
+ 903 0
906
+ 904 1
907
+ 905 0
908
+ 906 0
909
+ 907 0
910
+ 908 1
911
+ 909 0
912
+ 910 0
913
+ 911 1
914
+ 912 1
915
+ 913 1
916
+ 914 0
917
+ 915 0
918
+ 916 1
919
+ 917 1
920
+ 918 0
921
+ 919 1
922
+ 920 1
923
+ 921 1
924
+ 922 1
925
+ 923 0
926
+ 924 0
927
+ 925 1
928
+ 926 1
929
+ 927 1
930
+ 928 1
931
+ 929 0
932
+ 930 0
933
+ 931 0
934
+ 932 1
935
+ 933 0
936
+ 934 0
937
+ 935 1
938
+ 936 0
939
+ 937 0
940
+ 938 0
941
+ 939 1
942
+ 940 1
943
+ 941 1
944
+ 942 0
945
+ 943 1
946
+ 944 1
947
+ 945 0
948
+ 946 1
949
+ 947 0
950
+ 948 0
951
+ 949 1
952
+ 950 1
953
+ 951 1
954
+ 952 0
955
+ 953 1
956
+ 954 0
957
+ 955 0
958
+ 956 0
959
+ 957 0
960
+ 958 1
961
+ 959 0
962
+ 960 0
963
+ 961 1
964
+ 962 1
965
+ 963 0
966
+ 964 0
967
+ 965 1
968
+ 966 1
969
+ 967 0
970
+ 968 1
971
+ 969 0
972
+ 970 0
973
+ 971 0
974
+ 972 0
975
+ 973 1
976
+ 974 1
977
+ 975 0
978
+ 976 1
979
+ 977 1
980
+ 978 1
981
+ 979 0
982
+ 980 1
983
+ 981 1
984
+ 982 1
985
+ 983 1
986
+ 984 0
987
+ 985 1
988
+ 986 1
989
+ 987 1
990
+ 988 0
991
+ 989 1
992
+ 990 0
993
+ 991 0
994
+ 992 1
995
+ 993 1
996
+ 994 1
997
+ 995 1
998
+ 996 1
999
+ 997 1
1000
+ 998 0
1001
+ 999 0
1002
+ 1000 1
1003
+ 1001 0
1004
+ 1002 1
1005
+ 1003 0
1006
+ 1004 1
1007
+ 1005 0
1008
+ 1006 1
1009
+ 1007 1
1010
+ 1008 1
1011
+ 1009 1
1012
+ 1010 0
1013
+ 1011 0
1014
+ 1012 1
1015
+ 1013 1
1016
+ 1014 0
1017
+ 1015 1
1018
+ 1016 0
1019
+ 1017 1
1020
+ 1018 0
1021
+ 1019 1
1022
+ 1020 0
1023
+ 1021 0
1024
+ 1022 1
1025
+ 1023 1
1026
+ 1024 1
1027
+ 1025 1
1028
+ 1026 1
1029
+ 1027 0
1030
+ 1028 1
1031
+ 1029 0
1032
+ 1030 0
1033
+ 1031 1
1034
+ 1032 0
1035
+ 1033 1
1036
+ 1034 0
1037
+ 1035 0
1038
+ 1036 0
1039
+ 1037 0
1040
+ 1038 1
1041
+ 1039 1
1042
+ 1040 1
1043
+ 1041 1
1044
+ 1042 1
1045
+ 1043 0
1046
+ 1044 1
1047
+ 1045 1
1048
+ 1046 1
1049
+ 1047 0
1050
+ 1048 0
1051
+ 1049 1
1052
+ 1050 0
1053
+ 1051 1
1054
+ 1052 0
1055
+ 1053 1
1056
+ 1054 1
1057
+ 1055 0
1058
+ 1056 1
1059
+ 1057 0
1060
+ 1058 0
1061
+ 1059 0
1062
+ 1060 0
1063
+ 1061 1
1064
+ 1062 1
1065
+ 1063 0
1066
+ 1064 1
1067
+ 1065 0
1068
+ 1066 0
1069
+ 1067 1
1070
+ 1068 1
1071
+ 1069 0
1072
+ 1070 1
1073
+ 1071 0
1074
+ 1072 1
1075
+ 1073 0
1076
+ 1074 1
1077
+ 1075 1
1078
+ 1076 1
1079
+ 1077 0
1080
+ 1078 0
1081
+ 1079 1
1082
+ 1080 0
1083
+ 1081 0
1084
+ 1082 1
1085
+ 1083 0
1086
+ 1084 1
1087
+ 1085 1
1088
+ 1086 0
1089
+ 1087 0
1090
+ 1088 0
1091
+ 1089 1
1092
+ 1090 0
1093
+ 1091 0
1094
+ 1092 0
1095
+ 1093 0
1096
+ 1094 0
1097
+ 1095 1
1098
+ 1096 0
1099
+ 1097 1
1100
+ 1098 1
1101
+ 1099 1
1102
+ 1100 1
1103
+ 1101 1
1104
+ 1102 1
1105
+ 1103 1
1106
+ 1104 0
1107
+ 1105 0
1108
+ 1106 0
1109
+ 1107 0
1110
+ 1108 0
1111
+ 1109 1
1112
+ 1110 0
1113
+ 1111 1
1114
+ 1112 1
1115
+ 1113 1
1116
+ 1114 0
1117
+ 1115 0
1118
+ 1116 0
1119
+ 1117 1
1120
+ 1118 0
1121
+ 1119 1
1122
+ 1120 0
1123
+ 1121 1
1124
+ 1122 1
1125
+ 1123 0
1126
+ 1124 0
1127
+ 1125 1
1128
+ 1126 1
1129
+ 1127 1
1130
+ 1128 1
1131
+ 1129 1
1132
+ 1130 0
1133
+ 1131 1
1134
+ 1132 1
1135
+ 1133 1
1136
+ 1134 1
1137
+ 1135 1
1138
+ 1136 0
1139
+ 1137 0
1140
+ 1138 0
1141
+ 1139 1
1142
+ 1140 0
1143
+ 1141 0
1144
+ 1142 1
1145
+ 1143 1
1146
+ 1144 1
1147
+ 1145 1
1148
+ 1146 1
1149
+ 1147 0
1150
+ 1148 1
1151
+ 1149 1
1152
+ 1150 1
1153
+ 1151 1
1154
+ 1152 1
1155
+ 1153 1
1156
+ 1154 1
1157
+ 1155 1
1158
+ 1156 1
1159
+ 1157 1
1160
+ 1158 1
1161
+ 1159 0
1162
+ 1160 1
1163
+ 1161 1
1164
+ 1162 1
1165
+ 1163 0
1166
+ 1164 0
1167
+ 1165 0
1168
+ 1166 1
1169
+ 1167 1
1170
+ 1168 1
1171
+ 1169 1
1172
+ 1170 0
1173
+ 1171 0
1174
+ 1172 1
1175
+ 1173 1
1176
+ 1174 1
1177
+ 1175 1
1178
+ 1176 0
1179
+ 1177 0
1180
+ 1178 0
1181
+ 1179 1
1182
+ 1180 0
1183
+ 1181 1
1184
+ 1182 0
1185
+ 1183 0
1186
+ 1184 0
1187
+ 1185 0
1188
+ 1186 0
1189
+ 1187 0
1190
+ 1188 1
1191
+ 1189 0
1192
+ 1190 0
1193
+ 1191 0
1194
+ 1192 0
1195
+ 1193 0
1196
+ 1194 1
1197
+ 1195 0
1198
+ 1196 0
1199
+ 1197 1
1200
+ 1198 1
1201
+ 1199 1
1202
+ 1200 1
1203
+ 1201 1
1204
+ 1202 0
1205
+ 1203 0
1206
+ 1204 0
1207
+ 1205 0
1208
+ 1206 1
1209
+ 1207 0
1210
+ 1208 0
1211
+ 1209 1
1212
+ 1210 0
1213
+ 1211 0
1214
+ 1212 1
1215
+ 1213 1
1216
+ 1214 0
1217
+ 1215 0
1218
+ 1216 1
1219
+ 1217 1
1220
+ 1218 1
1221
+ 1219 0
1222
+ 1220 1
1223
+ 1221 0
1224
+ 1222 1
1225
+ 1223 0
1226
+ 1224 0
1227
+ 1225 0
1228
+ 1226 0
1229
+ 1227 0
1230
+ 1228 0
1231
+ 1229 1
1232
+ 1230 1
1233
+ 1231 0
1234
+ 1232 1
1235
+ 1233 1
1236
+ 1234 1
1237
+ 1235 0
1238
+ 1236 1
1239
+ 1237 1
1240
+ 1238 0
1241
+ 1239 0
1242
+ 1240 1
1243
+ 1241 0
1244
+ 1242 1
1245
+ 1243 1
1246
+ 1244 1
1247
+ 1245 0
1248
+ 1246 0
1249
+ 1247 1
1250
+ 1248 0
1251
+ 1249 0
1252
+ 1250 0
1253
+ 1251 1
1254
+ 1252 1
1255
+ 1253 0
1256
+ 1254 0
1257
+ 1255 0
1258
+ 1256 0
1259
+ 1257 1
1260
+ 1258 0
1261
+ 1259 1
1262
+ 1260 1
1263
+ 1261 0
1264
+ 1262 0
1265
+ 1263 0
1266
+ 1264 0
1267
+ 1265 1
1268
+ 1266 0
1269
+ 1267 1
1270
+ 1268 1
1271
+ 1269 1
1272
+ 1270 1
1273
+ 1271 0
1274
+ 1272 0
1275
+ 1273 0
1276
+ 1274 1
1277
+ 1275 0
1278
+ 1276 0
1279
+ 1277 0
1280
+ 1278 0
1281
+ 1279 0
1282
+ 1280 1
1283
+ 1281 0
1284
+ 1282 0
1285
+ 1283 1
1286
+ 1284 1
1287
+ 1285 0
1288
+ 1286 0
1289
+ 1287 0
1290
+ 1288 1
1291
+ 1289 1
1292
+ 1290 0
1293
+ 1291 0
1294
+ 1292 1
1295
+ 1293 1
1296
+ 1294 1
1297
+ 1295 1
1298
+ 1296 0
1299
+ 1297 1
1300
+ 1298 0
1301
+ 1299 1
1302
+ 1300 0
1303
+ 1301 1
1304
+ 1302 1
1305
+ 1303 0
1306
+ 1304 0
1307
+ 1305 1
1308
+ 1306 1
1309
+ 1307 0
1310
+ 1308 1
1311
+ 1309 1
1312
+ 1310 0
1313
+ 1311 0
1314
+ 1312 0
1315
+ 1313 0
1316
+ 1314 1
1317
+ 1315 0
1318
+ 1316 0
1319
+ 1317 0
1320
+ 1318 0
1321
+ 1319 1
1322
+ 1320 0
1323
+ 1321 0
1324
+ 1322 0
1325
+ 1323 1
1326
+ 1324 1
1327
+ 1325 0
1328
+ 1326 1
1329
+ 1327 0
1330
+ 1328 1
1331
+ 1329 0
1332
+ 1330 1
1333
+ 1331 1
1334
+ 1332 0
1335
+ 1333 0
1336
+ 1334 0
1337
+ 1335 0
1338
+ 1336 1
1339
+ 1337 0
1340
+ 1338 0
1341
+ 1339 0
1342
+ 1340 0
1343
+ 1341 1
1344
+ 1342 0
1345
+ 1343 1
1346
+ 1344 1
1347
+ 1345 0
1348
+ 1346 0
1349
+ 1347 1
1350
+ 1348 0
1351
+ 1349 0
1352
+ 1350 0
1353
+ 1351 0
1354
+ 1352 0
1355
+ 1353 0
1356
+ 1354 1
1357
+ 1355 0
1358
+ 1356 0
1359
+ 1357 0
1360
+ 1358 1
1361
+ 1359 1
1362
+ 1360 1
1363
+ 1361 1
1364
+ 1362 1
1365
+ 1363 0
1366
+ 1364 0
1367
+ 1365 0
1368
+ 1366 1
1369
+ 1367 0
1370
+ 1368 1
1371
+ 1369 0
1372
+ 1370 0
1373
+ 1371 0
1374
+ 1372 0
1375
+ 1373 0
1376
+ 1374 0
1377
+ 1375 1
1378
+ 1376 0
1379
+ 1377 0
1380
+ 1378 0
1381
+ 1379 0
1382
+ 1380 0
1383
+ 1381 1
1384
+ 1382 0
1385
+ 1383 1
1386
+ 1384 1
1387
+ 1385 1
1388
+ 1386 0
1389
+ 1387 1
1390
+ 1388 1
1391
+ 1389 0
1392
+ 1390 0
1393
+ 1391 1
1394
+ 1392 1
1395
+ 1393 1
1396
+ 1394 0
1397
+ 1395 0
1398
+ 1396 1
1399
+ 1397 1
1400
+ 1398 0
1401
+ 1399 1
1402
+ 1400 1
1403
+ 1401 1
1404
+ 1402 1
1405
+ 1403 1
1406
+ 1404 0
1407
+ 1405 0
1408
+ 1406 0
1409
+ 1407 0
1410
+ 1408 0
1411
+ 1409 0
1412
+ 1410 1
1413
+ 1411 1
1414
+ 1412 1
1415
+ 1413 1
1416
+ 1414 0
1417
+ 1415 1
1418
+ 1416 0
1419
+ 1417 1
1420
+ 1418 1
1421
+ 1419 1
1422
+ 1420 1
1423
+ 1421 0
1424
+ 1422 0
1425
+ 1423 1
1426
+ 1424 1
1427
+ 1425 1
1428
+ 1426 1
1429
+ 1427 0
1430
+ 1428 0
1431
+ 1429 0
1432
+ 1430 1
1433
+ 1431 1
1434
+ 1432 1
1435
+ 1433 0
1436
+ 1434 0
1437
+ 1435 0
1438
+ 1436 1
1439
+ 1437 1
1440
+ 1438 1
1441
+ 1439 0
1442
+ 1440 0
1443
+ 1441 0
1444
+ 1442 0
1445
+ 1443 1
1446
+ 1444 0
1447
+ 1445 1
1448
+ 1446 1
1449
+ 1447 0
1450
+ 1448 0
1451
+ 1449 0
1452
+ 1450 1
1453
+ 1451 0
1454
+ 1452 1
1455
+ 1453 0
1456
+ 1454 1
1457
+ 1455 1
1458
+ 1456 0
1459
+ 1457 0
1460
+ 1458 1
1461
+ 1459 0
1462
+ 1460 1
1463
+ 1461 0
1464
+ 1462 1
1465
+ 1463 1
1466
+ 1464 1
1467
+ 1465 0
1468
+ 1466 0
1469
+ 1467 0
1470
+ 1468 0
1471
+ 1469 0
1472
+ 1470 0
1473
+ 1471 1
1474
+ 1472 1
1475
+ 1473 1
1476
+ 1474 0
1477
+ 1475 1
1478
+ 1476 0
1479
+ 1477 1
1480
+ 1478 0
1481
+ 1479 1
1482
+ 1480 0
1483
+ 1481 1
1484
+ 1482 1
1485
+ 1483 1
1486
+ 1484 1
1487
+ 1485 0
1488
+ 1486 1
1489
+ 1487 0
1490
+ 1488 1
1491
+ 1489 0
1492
+ 1490 1
1493
+ 1491 1
1494
+ 1492 1
1495
+ 1493 0
1496
+ 1494 1
1497
+ 1495 1
1498
+ 1496 0
1499
+ 1497 0
1500
+ 1498 0
1501
+ 1499 0
1502
+ 1500 1
1503
+ 1501 1
1504
+ 1502 1
1505
+ 1503 1
1506
+ 1504 1
1507
+ 1505 1
1508
+ 1506 0
1509
+ 1507 1
1510
+ 1508 0
1511
+ 1509 1
1512
+ 1510 0
1513
+ 1511 1
1514
+ 1512 0
1515
+ 1513 0
1516
+ 1514 1
1517
+ 1515 1
1518
+ 1516 1
1519
+ 1517 0
1520
+ 1518 1
1521
+ 1519 0
1522
+ 1520 0
1523
+ 1521 0
1524
+ 1522 1
1525
+ 1523 1
1526
+ 1524 0
1527
+ 1525 0
1528
+ 1526 1
1529
+ 1527 0
1530
+ 1528 1
1531
+ 1529 0
1532
+ 1530 0
1533
+ 1531 1
1534
+ 1532 1
1535
+ 1533 0
1536
+ 1534 1
1537
+ 1535 1
1538
+ 1536 1
1539
+ 1537 1
1540
+ 1538 0
1541
+ 1539 1
1542
+ 1540 0
1543
+ 1541 0
1544
+ 1542 0
1545
+ 1543 0
1546
+ 1544 0
1547
+ 1545 1
1548
+ 1546 1
1549
+ 1547 1
1550
+ 1548 0
1551
+ 1549 0
1552
+ 1550 1
1553
+ 1551 0
1554
+ 1552 0
1555
+ 1553 0
1556
+ 1554 1
1557
+ 1555 1
1558
+ 1556 0
1559
+ 1557 0
1560
+ 1558 0
1561
+ 1559 0
1562
+ 1560 0
1563
+ 1561 0
1564
+ 1562 0
1565
+ 1563 1
1566
+ 1564 1
1567
+ 1565 1
1568
+ 1566 0
1569
+ 1567 1
1570
+ 1568 1
1571
+ 1569 1
1572
+ 1570 1
1573
+ 1571 0
1574
+ 1572 1
1575
+ 1573 1
1576
+ 1574 0
1577
+ 1575 0
1578
+ 1576 1
1579
+ 1577 0
1580
+ 1578 1
1581
+ 1579 0
1582
+ 1580 1
1583
+ 1581 1
1584
+ 1582 0
1585
+ 1583 1
1586
+ 1584 0
1587
+ 1585 0
1588
+ 1586 1
1589
+ 1587 0
1590
+ 1588 1
1591
+ 1589 1
1592
+ 1590 0
1593
+ 1591 1
1594
+ 1592 0
1595
+ 1593 0
1596
+ 1594 1
1597
+ 1595 0
1598
+ 1596 0
1599
+ 1597 1
1600
+ 1598 0
1601
+ 1599 1
1602
+ 1600 0
1603
+ 1601 1
1604
+ 1602 0
1605
+ 1603 1
1606
+ 1604 0
1607
+ 1605 0
1608
+ 1606 1
1609
+ 1607 1
1610
+ 1608 1
1611
+ 1609 1
1612
+ 1610 0
1613
+ 1611 0
1614
+ 1612 1
1615
+ 1613 0
1616
+ 1614 1
1617
+ 1615 0
1618
+ 1616 0
1619
+ 1617 0
1620
+ 1618 0
1621
+ 1619 1
1622
+ 1620 0
1623
+ 1621 1
1624
+ 1622 0
1625
+ 1623 1
1626
+ 1624 0
1627
+ 1625 1
1628
+ 1626 0
1629
+ 1627 1
1630
+ 1628 0
1631
+ 1629 0
1632
+ 1630 0
1633
+ 1631 1
1634
+ 1632 1
1635
+ 1633 0
1636
+ 1634 1
1637
+ 1635 0
1638
+ 1636 0
1639
+ 1637 0
1640
+ 1638 1
1641
+ 1639 1
1642
+ 1640 0
1643
+ 1641 1
1644
+ 1642 1
1645
+ 1643 0
1646
+ 1644 0
1647
+ 1645 0
1648
+ 1646 1
1649
+ 1647 1
1650
+ 1648 1
1651
+ 1649 0
1652
+ 1650 1
1653
+ 1651 1
1654
+ 1652 1
1655
+ 1653 0
1656
+ 1654 1
1657
+ 1655 1
1658
+ 1656 0
1659
+ 1657 1
1660
+ 1658 0
1661
+ 1659 1
1662
+ 1660 1
1663
+ 1661 1
1664
+ 1662 1
1665
+ 1663 1
1666
+ 1664 0
1667
+ 1665 0
1668
+ 1666 0
1669
+ 1667 0
1670
+ 1668 0
1671
+ 1669 1
1672
+ 1670 1
1673
+ 1671 0
1674
+ 1672 1
1675
+ 1673 0
1676
+ 1674 1
1677
+ 1675 1
1678
+ 1676 0
1679
+ 1677 1
1680
+ 1678 1
1681
+ 1679 0
1682
+ 1680 0
1683
+ 1681 0
1684
+ 1682 1
1685
+ 1683 0
1686
+ 1684 1
1687
+ 1685 0
1688
+ 1686 1
1689
+ 1687 1
1690
+ 1688 0
1691
+ 1689 0
1692
+ 1690 0
1693
+ 1691 1
1694
+ 1692 1
1695
+ 1693 1
1696
+ 1694 0
1697
+ 1695 0
1698
+ 1696 1
1699
+ 1697 0
1700
+ 1698 1
1701
+ 1699 0
1702
+ 1700 0
1703
+ 1701 0
1704
+ 1702 1
1705
+ 1703 1
1706
+ 1704 1
1707
+ 1705 0
1708
+ 1706 0
1709
+ 1707 1
1710
+ 1708 0
1711
+ 1709 1
1712
+ 1710 1
1713
+ 1711 1
1714
+ 1712 1
1715
+ 1713 0
1716
+ 1714 1
1717
+ 1715 1
1718
+ 1716 0
1719
+ 1717 0
1720
+ 1718 0
1721
+ 1719 0
1722
+ 1720 0
1723
+ 1721 0
1724
+ 1722 0
1725
+ 1723 1
1726
+ 1724 1
1727
+ 1725 1
1728
+ 1726 1
1729
+ 1727 1
1730
+ 1728 1
1731
+ 1729 0
1732
+ 1730 1
1733
+ 1731 1
1734
+ 1732 0
1735
+ 1733 1
1736
+ 1734 0
1737
+ 1735 0
1738
+ 1736 1
1739
+ 1737 1
1740
+ 1738 0
1741
+ 1739 0
1742
+ 1740 1
1743
+ 1741 0
1744
+ 1742 0
1745
+ 1743 1
1746
+ 1744 1
1747
+ 1745 1
1748
+ 1746 1
1749
+ 1747 1
1750
+ 1748 1
1751
+ 1749 1
1752
+ 1750 1
1753
+ 1751 0
1754
+ 1752 0
1755
+ 1753 0
1756
+ 1754 1
1757
+ 1755 1
1758
+ 1756 1
1759
+ 1757 0
1760
+ 1758 1
1761
+ 1759 1
1762
+ 1760 0
1763
+ 1761 0
1764
+ 1762 0
1765
+ 1763 0
1766
+ 1764 1
1767
+ 1765 1
1768
+ 1766 1
1769
+ 1767 1
1770
+ 1768 0
1771
+ 1769 1
1772
+ 1770 0
1773
+ 1771 1
1774
+ 1772 0
1775
+ 1773 0
1776
+ 1774 1
1777
+ 1775 1
1778
+ 1776 0
1779
+ 1777 0
1780
+ 1778 0
1781
+ 1779 1
1782
+ 1780 0
1783
+ 1781 1
1784
+ 1782 1
1785
+ 1783 0
1786
+ 1784 1
1787
+ 1785 1
1788
+ 1786 0
1789
+ 1787 0
1790
+ 1788 1
1791
+ 1789 0
1792
+ 1790 0
1793
+ 1791 0
1794
+ 1792 0
1795
+ 1793 0
1796
+ 1794 0
1797
+ 1795 0
1798
+ 1796 0
1799
+ 1797 1
1800
+ 1798 1
1801
+ 1799 0
1802
+ 1800 1
1803
+ 1801 1
1804
+ 1802 1
1805
+ 1803 1
1806
+ 1804 0
1807
+ 1805 0
1808
+ 1806 0
1809
+ 1807 0
1810
+ 1808 0
1811
+ 1809 0
1812
+ 1810 1
1813
+ 1811 0
1814
+ 1812 0
1815
+ 1813 1
1816
+ 1814 1
1817
+ 1815 0
1818
+ 1816 0
1819
+ 1817 1
1820
+ 1818 1
1821
+ 1819 1
1822
+ 1820 1
reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d21h00m07,sd44/all_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 11.0,
3
+ "eval_accuracy": 0.9564220183486238,
4
+ "eval_loss": 0.15713490545749664,
5
+ "eval_runtime": 0.6469,
6
+ "eval_samples": 872,
7
+ "eval_samples_per_second": 1348.038,
8
+ "eval_steps_per_second": 3.092
9
+ }
reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d21h00m07,sd44/eval_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 11.0,
3
+ "eval_accuracy": 0.9564220183486238,
4
+ "eval_loss": 0.15713490545749664,
5
+ "eval_runtime": 0.6469,
6
+ "eval_samples": 872,
7
+ "eval_samples_per_second": 1348.038,
8
+ "eval_steps_per_second": 3.092
9
+ }
reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d21h00m07,sd44/ft/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[MASK]": 128000
3
+ }
reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d21h00m07,sd44/ft/special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": {
9
+ "content": "[UNK]",
10
+ "lstrip": false,
11
+ "normalized": true,
12
+ "rstrip": false,
13
+ "single_word": false
14
+ }
15
+ }
reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d21h00m07,sd44/ft/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d21h00m07,sd44/ft/tokenizer_config.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[CLS]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[SEP]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[UNK]",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "128000": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "[CLS]",
45
+ "clean_up_tokenization_spaces": false,
46
+ "cls_token": "[CLS]",
47
+ "do_lower_case": false,
48
+ "eos_token": "[SEP]",
49
+ "extra_special_tokens": {},
50
+ "mask_token": "[MASK]",
51
+ "model_max_length": 512,
52
+ "pad_token": "[PAD]",
53
+ "padding_side": "right",
54
+ "sep_token": "[SEP]",
55
+ "sp_model_kwargs": {},
56
+ "split_by_punct": false,
57
+ "tokenizer_class": "DebertaV2Tokenizer",
58
+ "unk_token": "[UNK]",
59
+ "vocab_type": "spm"
60
+ }
reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d21h00m07,sd44/ft2/README.md ADDED
@@ -0,0 +1,205 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: microsoft/deberta-v3-base
3
+ library_name: peft
4
+ tags:
5
+ - base_model:adapter:microsoft/deberta-v3-base
6
+ - transformers
7
+ ---
8
+
9
+ # Model Card for Model ID
10
+
11
+ <!-- Provide a quick summary of what the model is/does. -->
12
+
13
+
14
+
15
+ ## Model Details
16
+
17
+ ### Model Description
18
+
19
+ <!-- Provide a longer summary of what this model is. -->
20
+
21
+
22
+
23
+ - **Developed by:** [More Information Needed]
24
+ - **Funded by [optional]:** [More Information Needed]
25
+ - **Shared by [optional]:** [More Information Needed]
26
+ - **Model type:** [More Information Needed]
27
+ - **Language(s) (NLP):** [More Information Needed]
28
+ - **License:** [More Information Needed]
29
+ - **Finetuned from model [optional]:** [More Information Needed]
30
+
31
+ ### Model Sources [optional]
32
+
33
+ <!-- Provide the basic links for the model. -->
34
+
35
+ - **Repository:** [More Information Needed]
36
+ - **Paper [optional]:** [More Information Needed]
37
+ - **Demo [optional]:** [More Information Needed]
38
+
39
+ ## Uses
40
+
41
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
42
+
43
+ ### Direct Use
44
+
45
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
46
+
47
+ [More Information Needed]
48
+
49
+ ### Downstream Use [optional]
50
+
51
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
52
+
53
+ [More Information Needed]
54
+
55
+ ### Out-of-Scope Use
56
+
57
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
58
+
59
+ [More Information Needed]
60
+
61
+ ## Bias, Risks, and Limitations
62
+
63
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
64
+
65
+ [More Information Needed]
66
+
67
+ ### Recommendations
68
+
69
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
70
+
71
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
72
+
73
+ ## How to Get Started with the Model
74
+
75
+ Use the code below to get started with the model.
76
+
77
+ [More Information Needed]
78
+
79
+ ## Training Details
80
+
81
+ ### Training Data
82
+
83
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
84
+
85
+ [More Information Needed]
86
+
87
+ ### Training Procedure
88
+
89
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
90
+
91
+ #### Preprocessing [optional]
92
+
93
+ [More Information Needed]
94
+
95
+
96
+ #### Training Hyperparameters
97
+
98
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
99
+
100
+ #### Speeds, Sizes, Times [optional]
101
+
102
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
103
+
104
+ [More Information Needed]
105
+
106
+ ## Evaluation
107
+
108
+ <!-- This section describes the evaluation protocols and provides the results. -->
109
+
110
+ ### Testing Data, Factors & Metrics
111
+
112
+ #### Testing Data
113
+
114
+ <!-- This should link to a Dataset Card if possible. -->
115
+
116
+ [More Information Needed]
117
+
118
+ #### Factors
119
+
120
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
121
+
122
+ [More Information Needed]
123
+
124
+ #### Metrics
125
+
126
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
127
+
128
+ [More Information Needed]
129
+
130
+ ### Results
131
+
132
+ [More Information Needed]
133
+
134
+ #### Summary
135
+
136
+
137
+
138
+ ## Model Examination [optional]
139
+
140
+ <!-- Relevant interpretability work for the model goes here -->
141
+
142
+ [More Information Needed]
143
+
144
+ ## Environmental Impact
145
+
146
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
147
+
148
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
149
+
150
+ - **Hardware Type:** [More Information Needed]
151
+ - **Hours used:** [More Information Needed]
152
+ - **Cloud Provider:** [More Information Needed]
153
+ - **Compute Region:** [More Information Needed]
154
+ - **Carbon Emitted:** [More Information Needed]
155
+
156
+ ## Technical Specifications [optional]
157
+
158
+ ### Model Architecture and Objective
159
+
160
+ [More Information Needed]
161
+
162
+ ### Compute Infrastructure
163
+
164
+ [More Information Needed]
165
+
166
+ #### Hardware
167
+
168
+ [More Information Needed]
169
+
170
+ #### Software
171
+
172
+ [More Information Needed]
173
+
174
+ ## Citation [optional]
175
+
176
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
177
+
178
+ **BibTeX:**
179
+
180
+ [More Information Needed]
181
+
182
+ **APA:**
183
+
184
+ [More Information Needed]
185
+
186
+ ## Glossary [optional]
187
+
188
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
189
+
190
+ [More Information Needed]
191
+
192
+ ## More Information [optional]
193
+
194
+ [More Information Needed]
195
+
196
+ ## Model Card Authors [optional]
197
+
198
+ [More Information Needed]
199
+
200
+ ## Model Card Contact
201
+
202
+ [More Information Needed]
203
+ ### Framework versions
204
+
205
+ - PEFT 0.18.0
reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d21h00m07,sd44/ft2/adapter_config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": {
3
+ "base_model_class": "DebertaV2ForSequenceClassification",
4
+ "parent_library": "transformers.models.deberta_v2.modeling_deberta_v2"
5
+ },
6
+ "base_model_name_or_path": "microsoft/deberta-v3-base",
7
+ "bias": "none",
8
+ "boft_block_num": 0,
9
+ "boft_block_size": 4,
10
+ "boft_dropout": 0.15,
11
+ "boft_n_butterfly_factor": 2,
12
+ "exclude_modules": null,
13
+ "fan_in_fan_out": false,
14
+ "inference_mode": true,
15
+ "init_weights": true,
16
+ "layers_pattern": null,
17
+ "layers_to_transform": null,
18
+ "modules_to_save": [
19
+ "classifier",
20
+ "pooler"
21
+ ],
22
+ "peft_type": "BOFT",
23
+ "peft_version": "0.18.0",
24
+ "revision": null,
25
+ "target_modules": [
26
+ "attention.output.dense",
27
+ "output.dense",
28
+ "key_proj",
29
+ "intermediate.dense",
30
+ "value_proj",
31
+ "query_proj"
32
+ ],
33
+ "task_type": null
34
+ }
reproduction/glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d21h00m07,sd44/trainer_state.json ADDED
@@ -0,0 +1,1883 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 9600,
3
+ "best_metric": 0.9564220183486238,
4
+ "best_model_checkpoint": "./glue_expBOFT/sst2/dr0.15,mlr5e-05,clr5e-05,ep=11.0t=25d21h00m07,sd44/checkpoint-9600",
5
+ "epoch": 11.0,
6
+ "eval_steps": 200,
7
+ "global_step": 23155,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.09501187648456057,
14
+ "grad_norm": 1.7514725923538208,
15
+ "learning_rate": 1.9900000000000003e-05,
16
+ "loss": 0.6904,
17
+ "step": 200
18
+ },
19
+ {
20
+ "epoch": 0.09501187648456057,
21
+ "eval_accuracy": 0.5091743119266054,
22
+ "eval_loss": 0.6906887888908386,
23
+ "eval_runtime": 0.6759,
24
+ "eval_samples_per_second": 1290.205,
25
+ "eval_steps_per_second": 2.959,
26
+ "step": 200
27
+ },
28
+ {
29
+ "epoch": 0.19002375296912113,
30
+ "grad_norm": 1.6420567035675049,
31
+ "learning_rate": 3.99e-05,
32
+ "loss": 0.5409,
33
+ "step": 400
34
+ },
35
+ {
36
+ "epoch": 0.19002375296912113,
37
+ "eval_accuracy": 0.8922018348623854,
38
+ "eval_loss": 0.2870236337184906,
39
+ "eval_runtime": 1.0327,
40
+ "eval_samples_per_second": 844.371,
41
+ "eval_steps_per_second": 1.937,
42
+ "step": 400
43
+ },
44
+ {
45
+ "epoch": 0.2850356294536817,
46
+ "grad_norm": 4.483543395996094,
47
+ "learning_rate": 4.999764187388674e-05,
48
+ "loss": 0.2904,
49
+ "step": 600
50
+ },
51
+ {
52
+ "epoch": 0.2850356294536817,
53
+ "eval_accuracy": 0.9162844036697247,
54
+ "eval_loss": 0.236904114484787,
55
+ "eval_runtime": 1.0035,
56
+ "eval_samples_per_second": 868.964,
57
+ "eval_steps_per_second": 1.993,
58
+ "step": 600
59
+ },
60
+ {
61
+ "epoch": 0.38004750593824227,
62
+ "grad_norm": 2.000626564025879,
63
+ "learning_rate": 4.997849281538165e-05,
64
+ "loss": 0.2539,
65
+ "step": 800
66
+ },
67
+ {
68
+ "epoch": 0.38004750593824227,
69
+ "eval_accuracy": 0.9220183486238532,
70
+ "eval_loss": 0.22290953993797302,
71
+ "eval_runtime": 1.0184,
72
+ "eval_samples_per_second": 856.239,
73
+ "eval_steps_per_second": 1.964,
74
+ "step": 800
75
+ },
76
+ {
77
+ "epoch": 0.4750593824228028,
78
+ "grad_norm": 2.609875440597534,
79
+ "learning_rate": 4.9940113202284656e-05,
80
+ "loss": 0.2328,
81
+ "step": 1000
82
+ },
83
+ {
84
+ "epoch": 0.4750593824228028,
85
+ "eval_accuracy": 0.9139908256880734,
86
+ "eval_loss": 0.23303523659706116,
87
+ "eval_runtime": 1.0229,
88
+ "eval_samples_per_second": 852.488,
89
+ "eval_steps_per_second": 1.955,
90
+ "step": 1000
91
+ },
92
+ {
93
+ "epoch": 0.5700712589073634,
94
+ "grad_norm": 1.1132041215896606,
95
+ "learning_rate": 4.988253258246521e-05,
96
+ "loss": 0.237,
97
+ "step": 1200
98
+ },
99
+ {
100
+ "epoch": 0.5700712589073634,
101
+ "eval_accuracy": 0.9208715596330275,
102
+ "eval_loss": 0.2123883217573166,
103
+ "eval_runtime": 0.9702,
104
+ "eval_samples_per_second": 898.746,
105
+ "eval_steps_per_second": 2.061,
106
+ "step": 1200
107
+ },
108
+ {
109
+ "epoch": 0.665083135391924,
110
+ "grad_norm": 2.244969606399536,
111
+ "learning_rate": 4.980579528635033e-05,
112
+ "loss": 0.2248,
113
+ "step": 1400
114
+ },
115
+ {
116
+ "epoch": 0.665083135391924,
117
+ "eval_accuracy": 0.9288990825688074,
118
+ "eval_loss": 0.1928652971982956,
119
+ "eval_runtime": 1.0529,
120
+ "eval_samples_per_second": 828.182,
121
+ "eval_steps_per_second": 1.9,
122
+ "step": 1400
123
+ },
124
+ {
125
+ "epoch": 0.7600950118764845,
126
+ "grad_norm": 1.3691550493240356,
127
+ "learning_rate": 4.9709960392795314e-05,
128
+ "loss": 0.229,
129
+ "step": 1600
130
+ },
131
+ {
132
+ "epoch": 0.7600950118764845,
133
+ "eval_accuracy": 0.930045871559633,
134
+ "eval_loss": 0.1818411648273468,
135
+ "eval_runtime": 1.6058,
136
+ "eval_samples_per_second": 543.019,
137
+ "eval_steps_per_second": 1.245,
138
+ "step": 1600
139
+ },
140
+ {
141
+ "epoch": 0.8551068883610451,
142
+ "grad_norm": 2.4221930503845215,
143
+ "learning_rate": 4.959510168359979e-05,
144
+ "loss": 0.2062,
145
+ "step": 1800
146
+ },
147
+ {
148
+ "epoch": 0.8551068883610451,
149
+ "eval_accuracy": 0.930045871559633,
150
+ "eval_loss": 0.188993439078331,
151
+ "eval_runtime": 0.9925,
152
+ "eval_samples_per_second": 878.622,
153
+ "eval_steps_per_second": 2.015,
154
+ "step": 1800
155
+ },
156
+ {
157
+ "epoch": 0.9501187648456056,
158
+ "grad_norm": 2.23431658744812,
159
+ "learning_rate": 4.946130758670431e-05,
160
+ "loss": 0.201,
161
+ "step": 2000
162
+ },
163
+ {
164
+ "epoch": 0.9501187648456056,
165
+ "eval_accuracy": 0.9357798165137615,
166
+ "eval_loss": 0.1779988557100296,
167
+ "eval_runtime": 0.9388,
168
+ "eval_samples_per_second": 928.888,
169
+ "eval_steps_per_second": 2.13,
170
+ "step": 2000
171
+ },
172
+ {
173
+ "epoch": 1.0451306413301662,
174
+ "grad_norm": 2.040349006652832,
175
+ "learning_rate": 4.9308681108111e-05,
176
+ "loss": 0.2121,
177
+ "step": 2200
178
+ },
179
+ {
180
+ "epoch": 1.0451306413301662,
181
+ "eval_accuracy": 0.9346330275229358,
182
+ "eval_loss": 0.1772075593471527,
183
+ "eval_runtime": 1.0101,
184
+ "eval_samples_per_second": 863.297,
185
+ "eval_steps_per_second": 1.98,
186
+ "step": 2200
187
+ },
188
+ {
189
+ "epoch": 1.1401425178147269,
190
+ "grad_norm": 2.165803909301758,
191
+ "learning_rate": 4.9137339752580894e-05,
192
+ "loss": 0.1837,
193
+ "step": 2400
194
+ },
195
+ {
196
+ "epoch": 1.1401425178147269,
197
+ "eval_accuracy": 0.9380733944954128,
198
+ "eval_loss": 0.1726435273885727,
199
+ "eval_runtime": 0.9742,
200
+ "eval_samples_per_second": 895.11,
201
+ "eval_steps_per_second": 2.053,
202
+ "step": 2400
203
+ },
204
+ {
205
+ "epoch": 1.2351543942992875,
206
+ "grad_norm": 1.3453562259674072,
207
+ "learning_rate": 4.894741543316877e-05,
208
+ "loss": 0.1894,
209
+ "step": 2600
210
+ },
211
+ {
212
+ "epoch": 1.2351543942992875,
213
+ "eval_accuracy": 0.9334862385321101,
214
+ "eval_loss": 0.18274469673633575,
215
+ "eval_runtime": 0.9684,
216
+ "eval_samples_per_second": 900.483,
217
+ "eval_steps_per_second": 2.065,
218
+ "step": 2600
219
+ },
220
+ {
221
+ "epoch": 1.330166270783848,
222
+ "grad_norm": 1.084747314453125,
223
+ "learning_rate": 4.8739054369665395e-05,
224
+ "loss": 0.1923,
225
+ "step": 2800
226
+ },
227
+ {
228
+ "epoch": 1.330166270783848,
229
+ "eval_accuracy": 0.9346330275229358,
230
+ "eval_loss": 0.16488640010356903,
231
+ "eval_runtime": 1.0324,
232
+ "eval_samples_per_second": 844.596,
233
+ "eval_steps_per_second": 1.937,
234
+ "step": 2800
235
+ },
236
+ {
237
+ "epoch": 1.4251781472684084,
238
+ "grad_norm": 3.704699754714966,
239
+ "learning_rate": 4.851241697602512e-05,
240
+ "loss": 0.1708,
241
+ "step": 3000
242
+ },
243
+ {
244
+ "epoch": 1.4251781472684084,
245
+ "eval_accuracy": 0.9392201834862385,
246
+ "eval_loss": 0.168609619140625,
247
+ "eval_runtime": 1.0541,
248
+ "eval_samples_per_second": 827.255,
249
+ "eval_steps_per_second": 1.897,
250
+ "step": 3000
251
+ },
252
+ {
253
+ "epoch": 1.520190023752969,
254
+ "grad_norm": 3.629873275756836,
255
+ "learning_rate": 4.8267677736865697e-05,
256
+ "loss": 0.1779,
257
+ "step": 3200
258
+ },
259
+ {
260
+ "epoch": 1.520190023752969,
261
+ "eval_accuracy": 0.9438073394495413,
262
+ "eval_loss": 0.1655423790216446,
263
+ "eval_runtime": 0.9981,
264
+ "eval_samples_per_second": 873.643,
265
+ "eval_steps_per_second": 2.004,
266
+ "step": 3200
267
+ },
268
+ {
269
+ "epoch": 1.6152019002375297,
270
+ "grad_norm": 1.4464539289474487,
271
+ "learning_rate": 4.800502507313527e-05,
272
+ "loss": 0.1849,
273
+ "step": 3400
274
+ },
275
+ {
276
+ "epoch": 1.6152019002375297,
277
+ "eval_accuracy": 0.9369266055045872,
278
+ "eval_loss": 0.16120491921901703,
279
+ "eval_runtime": 1.0069,
280
+ "eval_samples_per_second": 866.023,
281
+ "eval_steps_per_second": 1.986,
282
+ "step": 3400
283
+ },
284
+ {
285
+ "epoch": 1.7102137767220902,
286
+ "grad_norm": 0.7518911957740784,
287
+ "learning_rate": 4.772466119705001e-05,
288
+ "loss": 0.167,
289
+ "step": 3600
290
+ },
291
+ {
292
+ "epoch": 1.7102137767220902,
293
+ "eval_accuracy": 0.9380733944954128,
294
+ "eval_loss": 0.16436979174613953,
295
+ "eval_runtime": 1.5794,
296
+ "eval_samples_per_second": 552.124,
297
+ "eval_steps_per_second": 1.266,
298
+ "step": 3600
299
+ },
300
+ {
301
+ "epoch": 1.8052256532066508,
302
+ "grad_norm": 1.3362587690353394,
303
+ "learning_rate": 4.742680195641403e-05,
304
+ "loss": 0.1717,
305
+ "step": 3800
306
+ },
307
+ {
308
+ "epoch": 1.8052256532066508,
309
+ "eval_accuracy": 0.944954128440367,
310
+ "eval_loss": 0.16524536907672882,
311
+ "eval_runtime": 0.993,
312
+ "eval_samples_per_second": 878.19,
313
+ "eval_steps_per_second": 2.014,
314
+ "step": 3800
315
+ },
316
+ {
317
+ "epoch": 1.9002375296912115,
318
+ "grad_norm": 2.359588384628296,
319
+ "learning_rate": 4.7111676668441565e-05,
320
+ "loss": 0.1747,
321
+ "step": 4000
322
+ },
323
+ {
324
+ "epoch": 1.9002375296912115,
325
+ "eval_accuracy": 0.9415137614678899,
326
+ "eval_loss": 0.15950897336006165,
327
+ "eval_runtime": 0.9576,
328
+ "eval_samples_per_second": 910.587,
329
+ "eval_steps_per_second": 2.089,
330
+ "step": 4000
331
+ },
332
+ {
333
+ "epoch": 1.995249406175772,
334
+ "grad_norm": 0.9058042168617249,
335
+ "learning_rate": 4.677952794320914e-05,
336
+ "loss": 0.1728,
337
+ "step": 4200
338
+ },
339
+ {
340
+ "epoch": 1.995249406175772,
341
+ "eval_accuracy": 0.9426605504587156,
342
+ "eval_loss": 0.159332737326622,
343
+ "eval_runtime": 0.6202,
344
+ "eval_samples_per_second": 1406.017,
345
+ "eval_steps_per_second": 3.225,
346
+ "step": 4200
347
+ },
348
+ {
349
+ "epoch": 2.0902612826603324,
350
+ "grad_norm": 1.3049640655517578,
351
+ "learning_rate": 4.643061149687386e-05,
352
+ "loss": 0.1636,
353
+ "step": 4400
354
+ },
355
+ {
356
+ "epoch": 2.0902612826603324,
357
+ "eval_accuracy": 0.9472477064220184,
358
+ "eval_loss": 0.15445983409881592,
359
+ "eval_runtime": 0.8053,
360
+ "eval_samples_per_second": 1082.783,
361
+ "eval_steps_per_second": 2.483,
362
+ "step": 4400
363
+ },
364
+ {
365
+ "epoch": 2.1852731591448933,
366
+ "grad_norm": 1.5209531784057617,
367
+ "learning_rate": 4.606519595480153e-05,
368
+ "loss": 0.1588,
369
+ "step": 4600
370
+ },
371
+ {
372
+ "epoch": 2.1852731591448933,
373
+ "eval_accuracy": 0.9461009174311926,
374
+ "eval_loss": 0.15329764783382416,
375
+ "eval_runtime": 0.6484,
376
+ "eval_samples_per_second": 1344.902,
377
+ "eval_steps_per_second": 3.085,
378
+ "step": 4600
379
+ },
380
+ {
381
+ "epoch": 2.2802850356294537,
382
+ "grad_norm": 1.5452911853790283,
383
+ "learning_rate": 4.568356264475607e-05,
384
+ "loss": 0.1612,
385
+ "step": 4800
386
+ },
387
+ {
388
+ "epoch": 2.2802850356294537,
389
+ "eval_accuracy": 0.9495412844036697,
390
+ "eval_loss": 0.16638603806495667,
391
+ "eval_runtime": 0.6156,
392
+ "eval_samples_per_second": 1416.546,
393
+ "eval_steps_per_second": 3.249,
394
+ "step": 4800
395
+ },
396
+ {
397
+ "epoch": 2.375296912114014,
398
+ "grad_norm": 0.9725423455238342,
399
+ "learning_rate": 4.5286005380309736e-05,
400
+ "loss": 0.162,
401
+ "step": 5000
402
+ },
403
+ {
404
+ "epoch": 2.375296912114014,
405
+ "eval_accuracy": 0.9518348623853211,
406
+ "eval_loss": 0.15117758512496948,
407
+ "eval_runtime": 0.7062,
408
+ "eval_samples_per_second": 1234.707,
409
+ "eval_steps_per_second": 2.832,
410
+ "step": 5000
411
+ },
412
+ {
413
+ "epoch": 2.470308788598575,
414
+ "grad_norm": 2.4135422706604004,
415
+ "learning_rate": 4.48728302346405e-05,
416
+ "loss": 0.1636,
417
+ "step": 5200
418
+ },
419
+ {
420
+ "epoch": 2.470308788598575,
421
+ "eval_accuracy": 0.9472477064220184,
422
+ "eval_loss": 0.163556307554245,
423
+ "eval_runtime": 0.6149,
424
+ "eval_samples_per_second": 1418.203,
425
+ "eval_steps_per_second": 3.253,
426
+ "step": 5200
427
+ },
428
+ {
429
+ "epoch": 2.5653206650831355,
430
+ "grad_norm": 0.9368196129798889,
431
+ "learning_rate": 4.44443553048911e-05,
432
+ "loss": 0.1585,
433
+ "step": 5400
434
+ },
435
+ {
436
+ "epoch": 2.5653206650831355,
437
+ "eval_accuracy": 0.944954128440367,
438
+ "eval_loss": 0.15284310281276703,
439
+ "eval_runtime": 0.7463,
440
+ "eval_samples_per_second": 1168.426,
441
+ "eval_steps_per_second": 2.68,
442
+ "step": 5400
443
+ },
444
+ {
445
+ "epoch": 2.660332541567696,
446
+ "grad_norm": 1.0074721574783325,
447
+ "learning_rate": 4.400091046727101e-05,
448
+ "loss": 0.1686,
449
+ "step": 5600
450
+ },
451
+ {
452
+ "epoch": 2.660332541567696,
453
+ "eval_accuracy": 0.948394495412844,
454
+ "eval_loss": 0.14752088487148285,
455
+ "eval_runtime": 0.9594,
456
+ "eval_samples_per_second": 908.947,
457
+ "eval_steps_per_second": 2.085,
458
+ "step": 5600
459
+ },
460
+ {
461
+ "epoch": 2.7553444180522564,
462
+ "grad_norm": 0.5866758823394775,
463
+ "learning_rate": 4.3542837123089805e-05,
464
+ "loss": 0.1449,
465
+ "step": 5800
466
+ },
467
+ {
468
+ "epoch": 2.7553444180522564,
469
+ "eval_accuracy": 0.9495412844036697,
470
+ "eval_loss": 0.14833877980709076,
471
+ "eval_runtime": 0.6098,
472
+ "eval_samples_per_second": 1429.886,
473
+ "eval_steps_per_second": 3.28,
474
+ "step": 5800
475
+ },
476
+ {
477
+ "epoch": 2.850356294536817,
478
+ "grad_norm": 0.47707095742225647,
479
+ "learning_rate": 4.307048793591759e-05,
480
+ "loss": 0.1584,
481
+ "step": 6000
482
+ },
483
+ {
484
+ "epoch": 2.850356294536817,
485
+ "eval_accuracy": 0.9506880733944955,
486
+ "eval_loss": 0.1465180665254593,
487
+ "eval_runtime": 0.6085,
488
+ "eval_samples_per_second": 1432.969,
489
+ "eval_steps_per_second": 3.287,
490
+ "step": 6000
491
+ },
492
+ {
493
+ "epoch": 2.9453681710213777,
494
+ "grad_norm": 1.0461047887802124,
495
+ "learning_rate": 4.25842265600748e-05,
496
+ "loss": 0.1529,
497
+ "step": 6200
498
+ },
499
+ {
500
+ "epoch": 2.9453681710213777,
501
+ "eval_accuracy": 0.9518348623853211,
502
+ "eval_loss": 0.15420927107334137,
503
+ "eval_runtime": 0.6322,
504
+ "eval_samples_per_second": 1379.208,
505
+ "eval_steps_per_second": 3.163,
506
+ "step": 6200
507
+ },
508
+ {
509
+ "epoch": 3.040380047505938,
510
+ "grad_norm": 0.9840183258056641,
511
+ "learning_rate": 4.208442736066038e-05,
512
+ "loss": 0.1574,
513
+ "step": 6400
514
+ },
515
+ {
516
+ "epoch": 3.040380047505938,
517
+ "eval_accuracy": 0.9495412844036697,
518
+ "eval_loss": 0.14834153652191162,
519
+ "eval_runtime": 0.7199,
520
+ "eval_samples_per_second": 1211.27,
521
+ "eval_steps_per_second": 2.778,
522
+ "step": 6400
523
+ },
524
+ {
525
+ "epoch": 3.1353919239904986,
526
+ "grad_norm": 2.284499168395996,
527
+ "learning_rate": 4.157147512533388e-05,
528
+ "loss": 0.1259,
529
+ "step": 6600
530
+ },
531
+ {
532
+ "epoch": 3.1353919239904986,
533
+ "eval_accuracy": 0.9506880733944955,
534
+ "eval_loss": 0.1611321121454239,
535
+ "eval_runtime": 0.6196,
536
+ "eval_samples_per_second": 1407.364,
537
+ "eval_steps_per_second": 3.228,
538
+ "step": 6600
539
+ },
540
+ {
541
+ "epoch": 3.2304038004750595,
542
+ "grad_norm": 1.1971311569213867,
543
+ "learning_rate": 4.104576476807338e-05,
544
+ "loss": 0.1444,
545
+ "step": 6800
546
+ },
547
+ {
548
+ "epoch": 3.2304038004750595,
549
+ "eval_accuracy": 0.948394495412844,
550
+ "eval_loss": 0.15331239998340607,
551
+ "eval_runtime": 0.6719,
552
+ "eval_samples_per_second": 1297.768,
553
+ "eval_steps_per_second": 2.977,
554
+ "step": 6800
555
+ },
556
+ {
557
+ "epoch": 3.32541567695962,
558
+ "grad_norm": 1.705183982849121,
559
+ "learning_rate": 4.050770102513736e-05,
560
+ "loss": 0.1517,
561
+ "step": 7000
562
+ },
563
+ {
564
+ "epoch": 3.32541567695962,
565
+ "eval_accuracy": 0.948394495412844,
566
+ "eval_loss": 0.1468544751405716,
567
+ "eval_runtime": 0.6123,
568
+ "eval_samples_per_second": 1424.167,
569
+ "eval_steps_per_second": 3.266,
570
+ "step": 7000
571
+ },
572
+ {
573
+ "epoch": 3.4204275534441804,
574
+ "grad_norm": 1.1557941436767578,
575
+ "learning_rate": 3.995769814346456e-05,
576
+ "loss": 0.1445,
577
+ "step": 7200
578
+ },
579
+ {
580
+ "epoch": 3.4204275534441804,
581
+ "eval_accuracy": 0.9541284403669725,
582
+ "eval_loss": 0.1495801955461502,
583
+ "eval_runtime": 0.7309,
584
+ "eval_samples_per_second": 1193.06,
585
+ "eval_steps_per_second": 2.736,
586
+ "step": 7200
587
+ },
588
+ {
589
+ "epoch": 3.5154394299287413,
590
+ "grad_norm": 0.9594547748565674,
591
+ "learning_rate": 3.939617956175159e-05,
592
+ "loss": 0.1453,
593
+ "step": 7400
594
+ },
595
+ {
596
+ "epoch": 3.5154394299287413,
597
+ "eval_accuracy": 0.9495412844036697,
598
+ "eval_loss": 0.15841937065124512,
599
+ "eval_runtime": 0.7203,
600
+ "eval_samples_per_second": 1210.537,
601
+ "eval_steps_per_second": 2.776,
602
+ "step": 7400
603
+ },
604
+ {
605
+ "epoch": 3.6104513064133017,
606
+ "grad_norm": 1.175558090209961,
607
+ "learning_rate": 3.8823577584454146e-05,
608
+ "loss": 0.1482,
609
+ "step": 7600
610
+ },
611
+ {
612
+ "epoch": 3.6104513064133017,
613
+ "eval_accuracy": 0.9529816513761468,
614
+ "eval_loss": 0.1476680189371109,
615
+ "eval_runtime": 1.1101,
616
+ "eval_samples_per_second": 785.547,
617
+ "eval_steps_per_second": 1.802,
618
+ "step": 7600
619
+ },
620
+ {
621
+ "epoch": 3.705463182897862,
622
+ "grad_norm": 1.2012368440628052,
623
+ "learning_rate": 3.8240333048962405e-05,
624
+ "loss": 0.1525,
625
+ "step": 7800
626
+ },
627
+ {
628
+ "epoch": 3.705463182897862,
629
+ "eval_accuracy": 0.9506880733944955,
630
+ "eval_loss": 0.14647004008293152,
631
+ "eval_runtime": 1.0142,
632
+ "eval_samples_per_second": 859.798,
633
+ "eval_steps_per_second": 1.972,
634
+ "step": 7800
635
+ },
636
+ {
637
+ "epoch": 3.800475059382423,
638
+ "grad_norm": 1.2092653512954712,
639
+ "learning_rate": 3.7646894986207234e-05,
640
+ "loss": 0.1399,
641
+ "step": 8000
642
+ },
643
+ {
644
+ "epoch": 3.800475059382423,
645
+ "eval_accuracy": 0.9518348623853211,
646
+ "eval_loss": 0.15854564309120178,
647
+ "eval_runtime": 0.6193,
648
+ "eval_samples_per_second": 1407.962,
649
+ "eval_steps_per_second": 3.229,
650
+ "step": 8000
651
+ },
652
+ {
653
+ "epoch": 3.8954869358669835,
654
+ "grad_norm": 1.5754486322402954,
655
+ "learning_rate": 3.704372027495819e-05,
656
+ "loss": 0.1479,
657
+ "step": 8200
658
+ },
659
+ {
660
+ "epoch": 3.8954869358669835,
661
+ "eval_accuracy": 0.9552752293577982,
662
+ "eval_loss": 0.15056897699832916,
663
+ "eval_runtime": 0.5866,
664
+ "eval_samples_per_second": 1486.601,
665
+ "eval_steps_per_second": 3.41,
666
+ "step": 8200
667
+ },
668
+ {
669
+ "epoch": 3.990498812351544,
670
+ "grad_norm": 1.033606767654419,
671
+ "learning_rate": 3.6431273290079624e-05,
672
+ "loss": 0.1442,
673
+ "step": 8400
674
+ },
675
+ {
676
+ "epoch": 3.990498812351544,
677
+ "eval_accuracy": 0.9552752293577982,
678
+ "eval_loss": 0.14493830502033234,
679
+ "eval_runtime": 0.5148,
680
+ "eval_samples_per_second": 1693.849,
681
+ "eval_steps_per_second": 3.885,
682
+ "step": 8400
683
+ },
684
+ {
685
+ "epoch": 4.085510688836105,
686
+ "grad_norm": 1.3973227739334106,
687
+ "learning_rate": 3.581002554501569e-05,
688
+ "loss": 0.1378,
689
+ "step": 8600
690
+ },
691
+ {
692
+ "epoch": 4.085510688836105,
693
+ "eval_accuracy": 0.9529816513761468,
694
+ "eval_loss": 0.154870867729187,
695
+ "eval_runtime": 0.8675,
696
+ "eval_samples_per_second": 1005.17,
697
+ "eval_steps_per_second": 2.305,
698
+ "step": 8600
699
+ },
700
+ {
701
+ "epoch": 4.180522565320665,
702
+ "grad_norm": 0.45185843110084534,
703
+ "learning_rate": 3.5180455328779474e-05,
704
+ "loss": 0.1363,
705
+ "step": 8800
706
+ },
707
+ {
708
+ "epoch": 4.180522565320665,
709
+ "eval_accuracy": 0.9552752293577982,
710
+ "eval_loss": 0.16305947303771973,
711
+ "eval_runtime": 0.8381,
712
+ "eval_samples_per_second": 1040.493,
713
+ "eval_steps_per_second": 2.386,
714
+ "step": 8800
715
+ },
716
+ {
717
+ "epoch": 4.275534441805226,
718
+ "grad_norm": 0.37426725029945374,
719
+ "learning_rate": 3.454304733772563e-05,
720
+ "loss": 0.1416,
721
+ "step": 9000
722
+ },
723
+ {
724
+ "epoch": 4.275534441805226,
725
+ "eval_accuracy": 0.9541284403669725,
726
+ "eval_loss": 0.1468331217765808,
727
+ "eval_runtime": 0.8685,
728
+ "eval_samples_per_second": 1004.038,
729
+ "eval_steps_per_second": 2.303,
730
+ "step": 9000
731
+ },
732
+ {
733
+ "epoch": 4.370546318289787,
734
+ "grad_norm": 1.494674563407898,
735
+ "learning_rate": 3.3898292302390225e-05,
736
+ "loss": 0.1358,
737
+ "step": 9200
738
+ },
739
+ {
740
+ "epoch": 4.370546318289787,
741
+ "eval_accuracy": 0.9529816513761468,
742
+ "eval_loss": 0.16494396328926086,
743
+ "eval_runtime": 0.8188,
744
+ "eval_samples_per_second": 1064.942,
745
+ "eval_steps_per_second": 2.443,
746
+ "step": 9200
747
+ },
748
+ {
749
+ "epoch": 4.465558194774347,
750
+ "grad_norm": 1.4736316204071045,
751
+ "learning_rate": 3.324668660968487e-05,
752
+ "loss": 0.1335,
753
+ "step": 9400
754
+ },
755
+ {
756
+ "epoch": 4.465558194774347,
757
+ "eval_accuracy": 0.9518348623853211,
758
+ "eval_loss": 0.15829212963581085,
759
+ "eval_runtime": 0.8486,
760
+ "eval_samples_per_second": 1027.574,
761
+ "eval_steps_per_second": 2.357,
762
+ "step": 9400
763
+ },
764
+ {
765
+ "epoch": 4.5605700712589075,
766
+ "grad_norm": 1.7827177047729492,
767
+ "learning_rate": 3.258873192073616e-05,
768
+ "loss": 0.1275,
769
+ "step": 9600
770
+ },
771
+ {
772
+ "epoch": 4.5605700712589075,
773
+ "eval_accuracy": 0.9564220183486238,
774
+ "eval_loss": 0.15713490545749664,
775
+ "eval_runtime": 0.8365,
776
+ "eval_samples_per_second": 1042.439,
777
+ "eval_steps_per_second": 2.391,
778
+ "step": 9600
779
+ },
780
+ {
781
+ "epoch": 4.6555819477434675,
782
+ "grad_norm": 1.2970054149627686,
783
+ "learning_rate": 3.1924934784664485e-05,
784
+ "loss": 0.142,
785
+ "step": 9800
786
+ },
787
+ {
788
+ "epoch": 4.6555819477434675,
789
+ "eval_accuracy": 0.9541284403669725,
790
+ "eval_loss": 0.1443370133638382,
791
+ "eval_runtime": 0.8915,
792
+ "eval_samples_per_second": 978.155,
793
+ "eval_steps_per_second": 2.243,
794
+ "step": 9800
795
+ },
796
+ {
797
+ "epoch": 4.750593824228028,
798
+ "grad_norm": 1.1736654043197632,
799
+ "learning_rate": 3.1255806248599775e-05,
800
+ "loss": 0.1361,
801
+ "step": 10000
802
+ },
803
+ {
804
+ "epoch": 4.750593824228028,
805
+ "eval_accuracy": 0.9518348623853211,
806
+ "eval_loss": 0.14679767191410065,
807
+ "eval_runtime": 0.8603,
808
+ "eval_samples_per_second": 1013.553,
809
+ "eval_steps_per_second": 2.325,
810
+ "step": 10000
811
+ },
812
+ {
813
+ "epoch": 4.845605700712589,
814
+ "grad_norm": 1.8192143440246582,
815
+ "learning_rate": 3.0581861464234214e-05,
816
+ "loss": 0.1373,
817
+ "step": 10200
818
+ },
819
+ {
820
+ "epoch": 4.845605700712589,
821
+ "eval_accuracy": 0.9506880733944955,
822
+ "eval_loss": 0.15044112503528595,
823
+ "eval_runtime": 0.8533,
824
+ "eval_samples_per_second": 1021.896,
825
+ "eval_steps_per_second": 2.344,
826
+ "step": 10200
827
+ },
828
+ {
829
+ "epoch": 4.94061757719715,
830
+ "grad_norm": 0.5954006910324097,
831
+ "learning_rate": 2.9903619291214902e-05,
832
+ "loss": 0.1447,
833
+ "step": 10400
834
+ },
835
+ {
836
+ "epoch": 4.94061757719715,
837
+ "eval_accuracy": 0.9541284403669725,
838
+ "eval_loss": 0.1495528221130371,
839
+ "eval_runtime": 0.8588,
840
+ "eval_samples_per_second": 1015.328,
841
+ "eval_steps_per_second": 2.329,
842
+ "step": 10400
843
+ },
844
+ {
845
+ "epoch": 5.03562945368171,
846
+ "grad_norm": 2.016547203063965,
847
+ "learning_rate": 2.9221601897681907e-05,
848
+ "loss": 0.1322,
849
+ "step": 10600
850
+ },
851
+ {
852
+ "epoch": 5.03562945368171,
853
+ "eval_accuracy": 0.9541284403669725,
854
+ "eval_loss": 0.15634067356586456,
855
+ "eval_runtime": 0.8391,
856
+ "eval_samples_per_second": 1039.173,
857
+ "eval_steps_per_second": 2.383,
858
+ "step": 10600
859
+ },
860
+ {
861
+ "epoch": 5.130641330166271,
862
+ "grad_norm": 1.5706390142440796,
863
+ "learning_rate": 2.853633435825906e-05,
864
+ "loss": 0.1329,
865
+ "step": 10800
866
+ },
867
+ {
868
+ "epoch": 5.130641330166271,
869
+ "eval_accuracy": 0.9495412844036697,
870
+ "eval_loss": 0.15066424012184143,
871
+ "eval_runtime": 0.8846,
872
+ "eval_samples_per_second": 985.705,
873
+ "eval_steps_per_second": 2.261,
874
+ "step": 10800
875
+ },
876
+ {
877
+ "epoch": 5.225653206650831,
878
+ "grad_norm": 1.007319688796997,
879
+ "learning_rate": 2.7848344249807153e-05,
880
+ "loss": 0.1331,
881
+ "step": 11000
882
+ },
883
+ {
884
+ "epoch": 5.225653206650831,
885
+ "eval_accuracy": 0.9495412844036697,
886
+ "eval_loss": 0.1453312486410141,
887
+ "eval_runtime": 0.8528,
888
+ "eval_samples_per_second": 1022.539,
889
+ "eval_steps_per_second": 2.345,
890
+ "step": 11000
891
+ },
892
+ {
893
+ "epoch": 5.320665083135392,
894
+ "grad_norm": 1.0420247316360474,
895
+ "learning_rate": 2.7158161245250686e-05,
896
+ "loss": 0.126,
897
+ "step": 11200
898
+ },
899
+ {
900
+ "epoch": 5.320665083135392,
901
+ "eval_accuracy": 0.9495412844036697,
902
+ "eval_loss": 0.15748272836208344,
903
+ "eval_runtime": 0.8885,
904
+ "eval_samples_per_second": 981.479,
905
+ "eval_steps_per_second": 2.251,
906
+ "step": 11200
907
+ },
908
+ {
909
+ "epoch": 5.415676959619953,
910
+ "grad_norm": 1.3743844032287598,
911
+ "learning_rate": 2.646631670579091e-05,
912
+ "loss": 0.1342,
913
+ "step": 11400
914
+ },
915
+ {
916
+ "epoch": 5.415676959619953,
917
+ "eval_accuracy": 0.9495412844036697,
918
+ "eval_loss": 0.14626795053482056,
919
+ "eval_runtime": 0.8487,
920
+ "eval_samples_per_second": 1027.513,
921
+ "eval_steps_per_second": 2.357,
922
+ "step": 11400
923
+ },
924
+ {
925
+ "epoch": 5.510688836104513,
926
+ "grad_norm": 1.2073663473129272,
927
+ "learning_rate": 2.5773343271819046e-05,
928
+ "loss": 0.1299,
929
+ "step": 11600
930
+ },
931
+ {
932
+ "epoch": 5.510688836104513,
933
+ "eval_accuracy": 0.9518348623853211,
934
+ "eval_loss": 0.14876167476177216,
935
+ "eval_runtime": 0.8834,
936
+ "eval_samples_per_second": 987.119,
937
+ "eval_steps_per_second": 2.264,
938
+ "step": 11600
939
+ },
940
+ {
941
+ "epoch": 5.605700712589074,
942
+ "grad_norm": 2.149050235748291,
943
+ "learning_rate": 2.5079774452844683e-05,
944
+ "loss": 0.1307,
945
+ "step": 11800
946
+ },
947
+ {
948
+ "epoch": 5.605700712589074,
949
+ "eval_accuracy": 0.9529816513761468,
950
+ "eval_loss": 0.15025578439235687,
951
+ "eval_runtime": 0.9048,
952
+ "eval_samples_per_second": 963.741,
953
+ "eval_steps_per_second": 2.21,
954
+ "step": 11800
955
+ },
956
+ {
957
+ "epoch": 5.7007125890736345,
958
+ "grad_norm": 1.3404911756515503,
959
+ "learning_rate": 2.4386144216755135e-05,
960
+ "loss": 0.1318,
961
+ "step": 12000
962
+ },
963
+ {
964
+ "epoch": 5.7007125890736345,
965
+ "eval_accuracy": 0.9529816513761468,
966
+ "eval_loss": 0.14999549090862274,
967
+ "eval_runtime": 0.832,
968
+ "eval_samples_per_second": 1048.065,
969
+ "eval_steps_per_second": 2.404,
970
+ "step": 12000
971
+ },
972
+ {
973
+ "epoch": 5.7957244655581945,
974
+ "grad_norm": 0.4969980716705322,
975
+ "learning_rate": 2.3692986578721787e-05,
976
+ "loss": 0.1257,
977
+ "step": 12200
978
+ },
979
+ {
980
+ "epoch": 5.7957244655581945,
981
+ "eval_accuracy": 0.9518348623853211,
982
+ "eval_loss": 0.15145188570022583,
983
+ "eval_runtime": 0.851,
984
+ "eval_samples_per_second": 1024.662,
985
+ "eval_steps_per_second": 2.35,
986
+ "step": 12200
987
+ },
988
+ {
989
+ "epoch": 5.890736342042755,
990
+ "grad_norm": 1.4386001825332642,
991
+ "learning_rate": 2.300083519007009e-05,
992
+ "loss": 0.1321,
993
+ "step": 12400
994
+ },
995
+ {
996
+ "epoch": 5.890736342042755,
997
+ "eval_accuracy": 0.9529816513761468,
998
+ "eval_loss": 0.15149681270122528,
999
+ "eval_runtime": 0.8313,
1000
+ "eval_samples_per_second": 1049.009,
1001
+ "eval_steps_per_second": 2.406,
1002
+ "step": 12400
1003
+ },
1004
+ {
1005
+ "epoch": 5.985748218527316,
1006
+ "grad_norm": 0.8061667680740356,
1007
+ "learning_rate": 2.231022292742972e-05,
1008
+ "loss": 0.1342,
1009
+ "step": 12600
1010
+ },
1011
+ {
1012
+ "epoch": 5.985748218527316,
1013
+ "eval_accuracy": 0.9529816513761468,
1014
+ "eval_loss": 0.15307840704917908,
1015
+ "eval_runtime": 0.6313,
1016
+ "eval_samples_per_second": 1381.312,
1017
+ "eval_steps_per_second": 3.168,
1018
+ "step": 12600
1019
+ },
1020
+ {
1021
+ "epoch": 6.080760095011876,
1022
+ "grad_norm": 1.0963289737701416,
1023
+ "learning_rate": 2.1621681482481037e-05,
1024
+ "loss": 0.1301,
1025
+ "step": 12800
1026
+ },
1027
+ {
1028
+ "epoch": 6.080760095011876,
1029
+ "eval_accuracy": 0.9564220183486238,
1030
+ "eval_loss": 0.15626297891139984,
1031
+ "eval_runtime": 0.6422,
1032
+ "eval_samples_per_second": 1357.787,
1033
+ "eval_steps_per_second": 3.114,
1034
+ "step": 12800
1035
+ },
1036
+ {
1037
+ "epoch": 6.175771971496437,
1038
+ "grad_norm": 1.1481024026870728,
1039
+ "learning_rate": 2.0935740952613994e-05,
1040
+ "loss": 0.1274,
1041
+ "step": 13000
1042
+ },
1043
+ {
1044
+ "epoch": 6.175771971496437,
1045
+ "eval_accuracy": 0.9529816513761468,
1046
+ "eval_loss": 0.15459315478801727,
1047
+ "eval_runtime": 0.6608,
1048
+ "eval_samples_per_second": 1319.593,
1049
+ "eval_steps_per_second": 3.027,
1050
+ "step": 13000
1051
+ },
1052
+ {
1053
+ "epoch": 6.270783847980997,
1054
+ "grad_norm": 0.7984941005706787,
1055
+ "learning_rate": 2.025292943281429e-05,
1056
+ "loss": 0.1184,
1057
+ "step": 13200
1058
+ },
1059
+ {
1060
+ "epoch": 6.270783847980997,
1061
+ "eval_accuracy": 0.9529816513761468,
1062
+ "eval_loss": 0.15706071257591248,
1063
+ "eval_runtime": 0.5212,
1064
+ "eval_samples_per_second": 1672.928,
1065
+ "eval_steps_per_second": 3.837,
1066
+ "step": 13200
1067
+ },
1068
+ {
1069
+ "epoch": 6.365795724465558,
1070
+ "grad_norm": 0.7598547339439392,
1071
+ "learning_rate": 1.957377260909128e-05,
1072
+ "loss": 0.1241,
1073
+ "step": 13400
1074
+ },
1075
+ {
1076
+ "epoch": 6.365795724465558,
1077
+ "eval_accuracy": 0.9552752293577982,
1078
+ "eval_loss": 0.15746158361434937,
1079
+ "eval_runtime": 0.5284,
1080
+ "eval_samples_per_second": 1650.363,
1081
+ "eval_steps_per_second": 3.785,
1082
+ "step": 13400
1083
+ },
1084
+ {
1085
+ "epoch": 6.460807600950119,
1086
+ "grad_norm": 1.695999264717102,
1087
+ "learning_rate": 1.8898793353760407e-05,
1088
+ "loss": 0.1239,
1089
+ "step": 13600
1090
+ },
1091
+ {
1092
+ "epoch": 6.460807600950119,
1093
+ "eval_accuracy": 0.9541284403669725,
1094
+ "eval_loss": 0.1590254157781601,
1095
+ "eval_runtime": 0.5235,
1096
+ "eval_samples_per_second": 1665.588,
1097
+ "eval_steps_per_second": 3.82,
1098
+ "step": 13600
1099
+ },
1100
+ {
1101
+ "epoch": 6.555819477434679,
1102
+ "grad_norm": 0.2552671432495117,
1103
+ "learning_rate": 1.8228511322891968e-05,
1104
+ "loss": 0.1175,
1105
+ "step": 13800
1106
+ },
1107
+ {
1108
+ "epoch": 6.555819477434679,
1109
+ "eval_accuracy": 0.9529816513761468,
1110
+ "eval_loss": 0.16055269539356232,
1111
+ "eval_runtime": 0.5547,
1112
+ "eval_samples_per_second": 1571.983,
1113
+ "eval_steps_per_second": 3.605,
1114
+ "step": 13800
1115
+ },
1116
+ {
1117
+ "epoch": 6.65083135391924,
1118
+ "grad_norm": 0.6147297024726868,
1119
+ "learning_rate": 1.7563442556235898e-05,
1120
+ "loss": 0.1244,
1121
+ "step": 14000
1122
+ },
1123
+ {
1124
+ "epoch": 6.65083135391924,
1125
+ "eval_accuracy": 0.9541284403669725,
1126
+ "eval_loss": 0.15496574342250824,
1127
+ "eval_runtime": 0.5259,
1128
+ "eval_samples_per_second": 1658.232,
1129
+ "eval_steps_per_second": 3.803,
1130
+ "step": 14000
1131
+ },
1132
+ {
1133
+ "epoch": 6.745843230403801,
1134
+ "grad_norm": 0.3021308481693268,
1135
+ "learning_rate": 1.690409907993083e-05,
1136
+ "loss": 0.1199,
1137
+ "step": 14200
1138
+ },
1139
+ {
1140
+ "epoch": 6.745843230403801,
1141
+ "eval_accuracy": 0.9552752293577982,
1142
+ "eval_loss": 0.15175311267375946,
1143
+ "eval_runtime": 0.6235,
1144
+ "eval_samples_per_second": 1398.668,
1145
+ "eval_steps_per_second": 3.208,
1146
+ "step": 14200
1147
+ },
1148
+ {
1149
+ "epoch": 6.840855106888361,
1150
+ "grad_norm": 1.5534125566482544,
1151
+ "learning_rate": 1.625098851230302e-05,
1152
+ "loss": 0.1281,
1153
+ "step": 14400
1154
+ },
1155
+ {
1156
+ "epoch": 6.840855106888361,
1157
+ "eval_accuracy": 0.9541284403669725,
1158
+ "eval_loss": 0.15342016518115997,
1159
+ "eval_runtime": 0.5996,
1160
+ "eval_samples_per_second": 1454.266,
1161
+ "eval_steps_per_second": 3.335,
1162
+ "step": 14400
1163
+ },
1164
+ {
1165
+ "epoch": 6.935866983372922,
1166
+ "grad_norm": 1.0003083944320679,
1167
+ "learning_rate": 1.5604613673058945e-05,
1168
+ "loss": 0.126,
1169
+ "step": 14600
1170
+ },
1171
+ {
1172
+ "epoch": 6.935866983372922,
1173
+ "eval_accuracy": 0.9552752293577982,
1174
+ "eval_loss": 0.15357157588005066,
1175
+ "eval_runtime": 0.5221,
1176
+ "eval_samples_per_second": 1670.207,
1177
+ "eval_steps_per_second": 3.831,
1178
+ "step": 14600
1179
+ },
1180
+ {
1181
+ "epoch": 7.0308788598574825,
1182
+ "grad_norm": 0.6248033046722412,
1183
+ "learning_rate": 1.4965472196172137e-05,
1184
+ "loss": 0.1238,
1185
+ "step": 14800
1186
+ },
1187
+ {
1188
+ "epoch": 7.0308788598574825,
1189
+ "eval_accuracy": 0.9529816513761468,
1190
+ "eval_loss": 0.1497410237789154,
1191
+ "eval_runtime": 0.5195,
1192
+ "eval_samples_per_second": 1678.517,
1193
+ "eval_steps_per_second": 3.85,
1194
+ "step": 14800
1195
+ },
1196
+ {
1197
+ "epoch": 7.1258907363420425,
1198
+ "grad_norm": 1.2488328218460083,
1199
+ "learning_rate": 1.4334056146762565e-05,
1200
+ "loss": 0.1245,
1201
+ "step": 15000
1202
+ },
1203
+ {
1204
+ "epoch": 7.1258907363420425,
1205
+ "eval_accuracy": 0.9552752293577982,
1206
+ "eval_loss": 0.1517316848039627,
1207
+ "eval_runtime": 0.6476,
1208
+ "eval_samples_per_second": 1346.533,
1209
+ "eval_steps_per_second": 3.088,
1210
+ "step": 15000
1211
+ },
1212
+ {
1213
+ "epoch": 7.220902612826603,
1214
+ "grad_norm": 0.7753092646598816,
1215
+ "learning_rate": 1.3710851642263234e-05,
1216
+ "loss": 0.1239,
1217
+ "step": 15200
1218
+ },
1219
+ {
1220
+ "epoch": 7.220902612826603,
1221
+ "eval_accuracy": 0.9529816513761468,
1222
+ "eval_loss": 0.15445499122142792,
1223
+ "eval_runtime": 0.6332,
1224
+ "eval_samples_per_second": 1377.182,
1225
+ "eval_steps_per_second": 3.159,
1226
+ "step": 15200
1227
+ },
1228
+ {
1229
+ "epoch": 7.315914489311164,
1230
+ "grad_norm": 1.4783904552459717,
1231
+ "learning_rate": 1.309633847816599e-05,
1232
+ "loss": 0.1205,
1233
+ "step": 15400
1234
+ },
1235
+ {
1236
+ "epoch": 7.315914489311164,
1237
+ "eval_accuracy": 0.9529816513761468,
1238
+ "eval_loss": 0.15255992114543915,
1239
+ "eval_runtime": 0.5177,
1240
+ "eval_samples_per_second": 1684.429,
1241
+ "eval_steps_per_second": 3.863,
1242
+ "step": 15400
1243
+ },
1244
+ {
1245
+ "epoch": 7.410926365795724,
1246
+ "grad_norm": 1.1244440078735352,
1247
+ "learning_rate": 1.2490989758634352e-05,
1248
+ "loss": 0.116,
1249
+ "step": 15600
1250
+ },
1251
+ {
1252
+ "epoch": 7.410926365795724,
1253
+ "eval_accuracy": 0.9529816513761468,
1254
+ "eval_loss": 0.15762777626514435,
1255
+ "eval_runtime": 0.5458,
1256
+ "eval_samples_per_second": 1597.769,
1257
+ "eval_steps_per_second": 3.665,
1258
+ "step": 15600
1259
+ },
1260
+ {
1261
+ "epoch": 7.505938242280285,
1262
+ "grad_norm": 0.2549656331539154,
1263
+ "learning_rate": 1.1895271532268005e-05,
1264
+ "loss": 0.1134,
1265
+ "step": 15800
1266
+ },
1267
+ {
1268
+ "epoch": 7.505938242280285,
1269
+ "eval_accuracy": 0.9564220183486238,
1270
+ "eval_loss": 0.15428341925144196,
1271
+ "eval_runtime": 0.5404,
1272
+ "eval_samples_per_second": 1613.532,
1273
+ "eval_steps_per_second": 3.701,
1274
+ "step": 15800
1275
+ },
1276
+ {
1277
+ "epoch": 7.600950118764846,
1278
+ "grad_norm": 1.8452253341674805,
1279
+ "learning_rate": 1.1309642433299092e-05,
1280
+ "loss": 0.1268,
1281
+ "step": 16000
1282
+ },
1283
+ {
1284
+ "epoch": 7.600950118764846,
1285
+ "eval_accuracy": 0.9518348623853211,
1286
+ "eval_loss": 0.15276707708835602,
1287
+ "eval_runtime": 0.5594,
1288
+ "eval_samples_per_second": 1558.735,
1289
+ "eval_steps_per_second": 3.575,
1290
+ "step": 16000
1291
+ },
1292
+ {
1293
+ "epoch": 7.695961995249406,
1294
+ "grad_norm": 0.6518962979316711,
1295
+ "learning_rate": 1.0734553328496972e-05,
1296
+ "loss": 0.1238,
1297
+ "step": 16200
1298
+ },
1299
+ {
1300
+ "epoch": 7.695961995249406,
1301
+ "eval_accuracy": 0.9564220183486238,
1302
+ "eval_loss": 0.14917632937431335,
1303
+ "eval_runtime": 0.5095,
1304
+ "eval_samples_per_second": 1711.495,
1305
+ "eval_steps_per_second": 3.925,
1306
+ "step": 16200
1307
+ },
1308
+ {
1309
+ "epoch": 7.790973871733967,
1310
+ "grad_norm": 1.0892926454544067,
1311
+ "learning_rate": 1.0170446970052763e-05,
1312
+ "loss": 0.1252,
1313
+ "step": 16400
1314
+ },
1315
+ {
1316
+ "epoch": 7.790973871733967,
1317
+ "eval_accuracy": 0.9495412844036697,
1318
+ "eval_loss": 0.1537962406873703,
1319
+ "eval_runtime": 0.6627,
1320
+ "eval_samples_per_second": 1315.823,
1321
+ "eval_steps_per_second": 3.018,
1322
+ "step": 16400
1323
+ },
1324
+ {
1325
+ "epoch": 7.885985748218527,
1326
+ "grad_norm": 1.2721507549285889,
1327
+ "learning_rate": 9.617757654711344e-06,
1328
+ "loss": 0.1185,
1329
+ "step": 16600
1330
+ },
1331
+ {
1332
+ "epoch": 7.885985748218527,
1333
+ "eval_accuracy": 0.9552752293577982,
1334
+ "eval_loss": 0.1545173078775406,
1335
+ "eval_runtime": 0.7112,
1336
+ "eval_samples_per_second": 1226.064,
1337
+ "eval_steps_per_second": 2.812,
1338
+ "step": 16600
1339
+ },
1340
+ {
1341
+ "epoch": 7.980997624703088,
1342
+ "grad_norm": 1.68338143825531,
1343
+ "learning_rate": 9.07691088941301e-06,
1344
+ "loss": 0.1165,
1345
+ "step": 16800
1346
+ },
1347
+ {
1348
+ "epoch": 7.980997624703088,
1349
+ "eval_accuracy": 0.9552752293577982,
1350
+ "eval_loss": 0.1595546454191208,
1351
+ "eval_runtime": 0.5416,
1352
+ "eval_samples_per_second": 1610.0,
1353
+ "eval_steps_per_second": 3.693,
1354
+ "step": 16800
1355
+ },
1356
+ {
1357
+ "epoch": 8.076009501187649,
1358
+ "grad_norm": 0.8316041231155396,
1359
+ "learning_rate": 8.548323063702269e-06,
1360
+ "loss": 0.132,
1361
+ "step": 17000
1362
+ },
1363
+ {
1364
+ "epoch": 8.076009501187649,
1365
+ "eval_accuracy": 0.9529816513761468,
1366
+ "eval_loss": 0.1475267857313156,
1367
+ "eval_runtime": 0.6916,
1368
+ "eval_samples_per_second": 1260.788,
1369
+ "eval_steps_per_second": 2.892,
1370
+ "step": 17000
1371
+ },
1372
+ {
1373
+ "epoch": 8.17102137767221,
1374
+ "grad_norm": 1.104248285293579,
1375
+ "learning_rate": 8.032401129155977e-06,
1376
+ "loss": 0.1155,
1377
+ "step": 17200
1378
+ },
1379
+ {
1380
+ "epoch": 8.17102137767221,
1381
+ "eval_accuracy": 0.9552752293577982,
1382
+ "eval_loss": 0.1523580700159073,
1383
+ "eval_runtime": 0.5785,
1384
+ "eval_samples_per_second": 1507.22,
1385
+ "eval_steps_per_second": 3.457,
1386
+ "step": 17200
1387
+ },
1388
+ {
1389
+ "epoch": 8.26603325415677,
1390
+ "grad_norm": 1.3649719953536987,
1391
+ "learning_rate": 7.52954228607767e-06,
1392
+ "loss": 0.118,
1393
+ "step": 17400
1394
+ },
1395
+ {
1396
+ "epoch": 8.26603325415677,
1397
+ "eval_accuracy": 0.9541284403669725,
1398
+ "eval_loss": 0.154517263174057,
1399
+ "eval_runtime": 0.5581,
1400
+ "eval_samples_per_second": 1562.523,
1401
+ "eval_steps_per_second": 3.584,
1402
+ "step": 17400
1403
+ },
1404
+ {
1405
+ "epoch": 8.36104513064133,
1406
+ "grad_norm": 0.7670822143554688,
1407
+ "learning_rate": 7.0401336776991095e-06,
1408
+ "loss": 0.1222,
1409
+ "step": 17600
1410
+ },
1411
+ {
1412
+ "epoch": 8.36104513064133,
1413
+ "eval_accuracy": 0.9541284403669725,
1414
+ "eval_loss": 0.15460576117038727,
1415
+ "eval_runtime": 0.9693,
1416
+ "eval_samples_per_second": 899.641,
1417
+ "eval_steps_per_second": 2.063,
1418
+ "step": 17600
1419
+ },
1420
+ {
1421
+ "epoch": 8.45605700712589,
1422
+ "grad_norm": 2.21932053565979,
1423
+ "learning_rate": 6.564552092124815e-06,
1424
+ "loss": 0.1162,
1425
+ "step": 17800
1426
+ },
1427
+ {
1428
+ "epoch": 8.45605700712589,
1429
+ "eval_accuracy": 0.9541284403669725,
1430
+ "eval_loss": 0.15456785261631012,
1431
+ "eval_runtime": 0.9057,
1432
+ "eval_samples_per_second": 962.79,
1433
+ "eval_steps_per_second": 2.208,
1434
+ "step": 17800
1435
+ },
1436
+ {
1437
+ "epoch": 8.551068883610451,
1438
+ "grad_norm": 0.9780743718147278,
1439
+ "learning_rate": 6.103163672248663e-06,
1440
+ "loss": 0.1049,
1441
+ "step": 18000
1442
+ },
1443
+ {
1444
+ "epoch": 8.551068883610451,
1445
+ "eval_accuracy": 0.9541284403669725,
1446
+ "eval_loss": 0.15727345645427704,
1447
+ "eval_runtime": 0.9078,
1448
+ "eval_samples_per_second": 960.53,
1449
+ "eval_steps_per_second": 2.203,
1450
+ "step": 18000
1451
+ },
1452
+ {
1453
+ "epoch": 8.646080760095012,
1454
+ "grad_norm": 1.746826171875,
1455
+ "learning_rate": 5.656323633866184e-06,
1456
+ "loss": 0.132,
1457
+ "step": 18200
1458
+ },
1459
+ {
1460
+ "epoch": 8.646080760095012,
1461
+ "eval_accuracy": 0.9552752293577982,
1462
+ "eval_loss": 0.15054485201835632,
1463
+ "eval_runtime": 0.959,
1464
+ "eval_samples_per_second": 909.28,
1465
+ "eval_steps_per_second": 2.086,
1466
+ "step": 18200
1467
+ },
1468
+ {
1469
+ "epoch": 8.741092636579573,
1470
+ "grad_norm": 0.7549273371696472,
1471
+ "learning_rate": 5.22437599219941e-06,
1472
+ "loss": 0.1184,
1473
+ "step": 18400
1474
+ },
1475
+ {
1476
+ "epoch": 8.741092636579573,
1477
+ "eval_accuracy": 0.9529816513761468,
1478
+ "eval_loss": 0.1521771401166916,
1479
+ "eval_runtime": 0.8683,
1480
+ "eval_samples_per_second": 1004.308,
1481
+ "eval_steps_per_second": 2.303,
1482
+ "step": 18400
1483
+ },
1484
+ {
1485
+ "epoch": 8.836104513064132,
1486
+ "grad_norm": 0.8423401117324829,
1487
+ "learning_rate": 4.807653297044895e-06,
1488
+ "loss": 0.1134,
1489
+ "step": 18600
1490
+ },
1491
+ {
1492
+ "epoch": 8.836104513064132,
1493
+ "eval_accuracy": 0.9552752293577982,
1494
+ "eval_loss": 0.15427321195602417,
1495
+ "eval_runtime": 0.9316,
1496
+ "eval_samples_per_second": 936.009,
1497
+ "eval_steps_per_second": 2.147,
1498
+ "step": 18600
1499
+ },
1500
+ {
1501
+ "epoch": 8.931116389548693,
1502
+ "grad_norm": 0.3886643350124359,
1503
+ "learning_rate": 4.406476376748761e-06,
1504
+ "loss": 0.123,
1505
+ "step": 18800
1506
+ },
1507
+ {
1508
+ "epoch": 8.931116389548693,
1509
+ "eval_accuracy": 0.9552752293577982,
1510
+ "eval_loss": 0.15342247486114502,
1511
+ "eval_runtime": 0.8887,
1512
+ "eval_samples_per_second": 981.18,
1513
+ "eval_steps_per_second": 2.25,
1514
+ "step": 18800
1515
+ },
1516
+ {
1517
+ "epoch": 9.026128266033254,
1518
+ "grad_norm": 1.1008235216140747,
1519
+ "learning_rate": 4.021154091205976e-06,
1520
+ "loss": 0.1221,
1521
+ "step": 19000
1522
+ },
1523
+ {
1524
+ "epoch": 9.026128266033254,
1525
+ "eval_accuracy": 0.9552752293577982,
1526
+ "eval_loss": 0.15262193977832794,
1527
+ "eval_runtime": 0.8891,
1528
+ "eval_samples_per_second": 980.719,
1529
+ "eval_steps_per_second": 2.249,
1530
+ "step": 19000
1531
+ },
1532
+ {
1533
+ "epoch": 9.121140142517815,
1534
+ "grad_norm": 0.7351152300834656,
1535
+ "learning_rate": 3.6519830940738103e-06,
1536
+ "loss": 0.1171,
1537
+ "step": 19200
1538
+ },
1539
+ {
1540
+ "epoch": 9.121140142517815,
1541
+ "eval_accuracy": 0.9552752293577982,
1542
+ "eval_loss": 0.15378212928771973,
1543
+ "eval_runtime": 0.961,
1544
+ "eval_samples_per_second": 907.398,
1545
+ "eval_steps_per_second": 2.081,
1546
+ "step": 19200
1547
+ },
1548
+ {
1549
+ "epoch": 9.216152019002376,
1550
+ "grad_norm": 0.9485660195350647,
1551
+ "learning_rate": 3.299247604382899e-06,
1552
+ "loss": 0.1158,
1553
+ "step": 19400
1554
+ },
1555
+ {
1556
+ "epoch": 9.216152019002376,
1557
+ "eval_accuracy": 0.9552752293577982,
1558
+ "eval_loss": 0.15435412526130676,
1559
+ "eval_runtime": 0.8558,
1560
+ "eval_samples_per_second": 1018.971,
1561
+ "eval_steps_per_second": 2.337,
1562
+ "step": 19400
1563
+ },
1564
+ {
1565
+ "epoch": 9.311163895486937,
1566
+ "grad_norm": 1.1230568885803223,
1567
+ "learning_rate": 2.9632191877213223e-06,
1568
+ "loss": 0.1208,
1569
+ "step": 19600
1570
+ },
1571
+ {
1572
+ "epoch": 9.311163895486937,
1573
+ "eval_accuracy": 0.9552752293577982,
1574
+ "eval_loss": 0.15259253978729248,
1575
+ "eval_runtime": 0.8748,
1576
+ "eval_samples_per_second": 996.848,
1577
+ "eval_steps_per_second": 2.286,
1578
+ "step": 19600
1579
+ },
1580
+ {
1581
+ "epoch": 9.406175771971496,
1582
+ "grad_norm": 1.1961973905563354,
1583
+ "learning_rate": 2.644156547160512e-06,
1584
+ "loss": 0.1191,
1585
+ "step": 19800
1586
+ },
1587
+ {
1588
+ "epoch": 9.406175771971496,
1589
+ "eval_accuracy": 0.9541284403669725,
1590
+ "eval_loss": 0.1536150872707367,
1591
+ "eval_runtime": 0.8576,
1592
+ "eval_samples_per_second": 1016.825,
1593
+ "eval_steps_per_second": 2.332,
1594
+ "step": 19800
1595
+ },
1596
+ {
1597
+ "epoch": 9.501187648456057,
1598
+ "grad_norm": 1.8314346075057983,
1599
+ "learning_rate": 2.3423053240837515e-06,
1600
+ "loss": 0.1147,
1601
+ "step": 20000
1602
+ },
1603
+ {
1604
+ "epoch": 9.501187648456057,
1605
+ "eval_accuracy": 0.9541284403669725,
1606
+ "eval_loss": 0.1538364589214325,
1607
+ "eval_runtime": 0.847,
1608
+ "eval_samples_per_second": 1029.564,
1609
+ "eval_steps_per_second": 2.361,
1610
+ "step": 20000
1611
+ },
1612
+ {
1613
+ "epoch": 9.596199524940618,
1614
+ "grad_norm": 1.333292841911316,
1615
+ "learning_rate": 2.0578979090706885e-06,
1616
+ "loss": 0.113,
1617
+ "step": 20200
1618
+ },
1619
+ {
1620
+ "epoch": 9.596199524940618,
1621
+ "eval_accuracy": 0.9541284403669725,
1622
+ "eval_loss": 0.1541624367237091,
1623
+ "eval_runtime": 0.8962,
1624
+ "eval_samples_per_second": 972.955,
1625
+ "eval_steps_per_second": 2.232,
1626
+ "step": 20200
1627
+ },
1628
+ {
1629
+ "epoch": 9.691211401425178,
1630
+ "grad_norm": 1.4652808904647827,
1631
+ "learning_rate": 1.7911532629834167e-06,
1632
+ "loss": 0.1332,
1633
+ "step": 20400
1634
+ },
1635
+ {
1636
+ "epoch": 9.691211401425178,
1637
+ "eval_accuracy": 0.9564220183486238,
1638
+ "eval_loss": 0.15292838215827942,
1639
+ "eval_runtime": 0.9022,
1640
+ "eval_samples_per_second": 966.574,
1641
+ "eval_steps_per_second": 2.217,
1642
+ "step": 20400
1643
+ },
1644
+ {
1645
+ "epoch": 9.78622327790974,
1646
+ "grad_norm": 0.8483909964561462,
1647
+ "learning_rate": 1.5422767483919132e-06,
1648
+ "loss": 0.1291,
1649
+ "step": 20600
1650
+ },
1651
+ {
1652
+ "epoch": 9.78622327790974,
1653
+ "eval_accuracy": 0.9564220183486238,
1654
+ "eval_loss": 0.15221929550170898,
1655
+ "eval_runtime": 1.4558,
1656
+ "eval_samples_per_second": 598.986,
1657
+ "eval_steps_per_second": 1.374,
1658
+ "step": 20600
1659
+ },
1660
+ {
1661
+ "epoch": 9.8812351543943,
1662
+ "grad_norm": 0.8579058051109314,
1663
+ "learning_rate": 1.3114599714685676e-06,
1664
+ "loss": 0.1055,
1665
+ "step": 20800
1666
+ },
1667
+ {
1668
+ "epoch": 9.8812351543943,
1669
+ "eval_accuracy": 0.9552752293577982,
1670
+ "eval_loss": 0.15338386595249176,
1671
+ "eval_runtime": 0.8535,
1672
+ "eval_samples_per_second": 1021.63,
1673
+ "eval_steps_per_second": 2.343,
1674
+ "step": 20800
1675
+ },
1676
+ {
1677
+ "epoch": 9.97624703087886,
1678
+ "grad_norm": 1.1100027561187744,
1679
+ "learning_rate": 1.0988806344735787e-06,
1680
+ "loss": 0.1205,
1681
+ "step": 21000
1682
+ },
1683
+ {
1684
+ "epoch": 9.97624703087886,
1685
+ "eval_accuracy": 0.9564220183486238,
1686
+ "eval_loss": 0.1530323177576065,
1687
+ "eval_runtime": 0.9569,
1688
+ "eval_samples_per_second": 911.291,
1689
+ "eval_steps_per_second": 2.09,
1690
+ "step": 21000
1691
+ },
1692
+ {
1693
+ "epoch": 10.07125890736342,
1694
+ "grad_norm": 0.19341756403446198,
1695
+ "learning_rate": 9.047023989446929e-07,
1696
+ "loss": 0.1231,
1697
+ "step": 21200
1698
+ },
1699
+ {
1700
+ "epoch": 10.07125890736342,
1701
+ "eval_accuracy": 0.9564220183486238,
1702
+ "eval_loss": 0.1525803804397583,
1703
+ "eval_runtime": 0.8635,
1704
+ "eval_samples_per_second": 1009.879,
1705
+ "eval_steps_per_second": 2.316,
1706
+ "step": 21200
1707
+ },
1708
+ {
1709
+ "epoch": 10.166270783847981,
1710
+ "grad_norm": 1.0303609371185303,
1711
+ "learning_rate": 7.290747596967823e-07,
1712
+ "loss": 0.1171,
1713
+ "step": 21400
1714
+ },
1715
+ {
1716
+ "epoch": 10.166270783847981,
1717
+ "eval_accuracy": 0.9564220183486238,
1718
+ "eval_loss": 0.1528913825750351,
1719
+ "eval_runtime": 0.8866,
1720
+ "eval_samples_per_second": 983.505,
1721
+ "eval_steps_per_second": 2.256,
1722
+ "step": 21400
1723
+ },
1724
+ {
1725
+ "epoch": 10.261282660332542,
1726
+ "grad_norm": 1.1353927850723267,
1727
+ "learning_rate": 5.721329297280559e-07,
1728
+ "loss": 0.1331,
1729
+ "step": 21600
1730
+ },
1731
+ {
1732
+ "epoch": 10.261282660332542,
1733
+ "eval_accuracy": 0.9564220183486238,
1734
+ "eval_loss": 0.152116060256958,
1735
+ "eval_runtime": 0.9013,
1736
+ "eval_samples_per_second": 967.48,
1737
+ "eval_steps_per_second": 2.219,
1738
+ "step": 21600
1739
+ },
1740
+ {
1741
+ "epoch": 10.356294536817103,
1742
+ "grad_norm": 0.6838251352310181,
1743
+ "learning_rate": 4.339977361216857e-07,
1744
+ "loss": 0.12,
1745
+ "step": 21800
1746
+ },
1747
+ {
1748
+ "epoch": 10.356294536817103,
1749
+ "eval_accuracy": 0.9541284403669725,
1750
+ "eval_loss": 0.15240301191806793,
1751
+ "eval_runtime": 0.7036,
1752
+ "eval_samples_per_second": 1239.387,
1753
+ "eval_steps_per_second": 2.843,
1754
+ "step": 21800
1755
+ },
1756
+ {
1757
+ "epoch": 10.451306413301662,
1758
+ "grad_norm": 1.2753151655197144,
1759
+ "learning_rate": 3.147755270229119e-07,
1760
+ "loss": 0.116,
1761
+ "step": 22000
1762
+ },
1763
+ {
1764
+ "epoch": 10.451306413301662,
1765
+ "eval_accuracy": 0.9541284403669725,
1766
+ "eval_loss": 0.15241561830043793,
1767
+ "eval_runtime": 0.5893,
1768
+ "eval_samples_per_second": 1479.653,
1769
+ "eval_steps_per_second": 3.394,
1770
+ "step": 22000
1771
+ },
1772
+ {
1773
+ "epoch": 10.546318289786223,
1774
+ "grad_norm": 0.616693913936615,
1775
+ "learning_rate": 2.1455808976323933e-07,
1776
+ "loss": 0.1115,
1777
+ "step": 22200
1778
+ },
1779
+ {
1780
+ "epoch": 10.546318289786223,
1781
+ "eval_accuracy": 0.9541284403669725,
1782
+ "eval_loss": 0.15251894295215607,
1783
+ "eval_runtime": 0.6592,
1784
+ "eval_samples_per_second": 1322.773,
1785
+ "eval_steps_per_second": 3.034,
1786
+ "step": 22200
1787
+ },
1788
+ {
1789
+ "epoch": 10.641330166270784,
1790
+ "grad_norm": 1.795017957687378,
1791
+ "learning_rate": 1.3342258019480858e-07,
1792
+ "loss": 0.1095,
1793
+ "step": 22400
1794
+ },
1795
+ {
1796
+ "epoch": 10.641330166270784,
1797
+ "eval_accuracy": 0.9541284403669725,
1798
+ "eval_loss": 0.15266826748847961,
1799
+ "eval_runtime": 0.648,
1800
+ "eval_samples_per_second": 1345.601,
1801
+ "eval_steps_per_second": 3.086,
1802
+ "step": 22400
1803
+ },
1804
+ {
1805
+ "epoch": 10.736342042755345,
1806
+ "grad_norm": 1.1101276874542236,
1807
+ "learning_rate": 7.143146328928985e-08,
1808
+ "loss": 0.1061,
1809
+ "step": 22600
1810
+ },
1811
+ {
1812
+ "epoch": 10.736342042755345,
1813
+ "eval_accuracy": 0.9541284403669725,
1814
+ "eval_loss": 0.15278586745262146,
1815
+ "eval_runtime": 0.5568,
1816
+ "eval_samples_per_second": 1566.143,
1817
+ "eval_steps_per_second": 3.592,
1818
+ "step": 22600
1819
+ },
1820
+ {
1821
+ "epoch": 10.831353919239906,
1822
+ "grad_norm": 1.0985181331634521,
1823
+ "learning_rate": 2.8632465047070955e-08,
1824
+ "loss": 0.1116,
1825
+ "step": 22800
1826
+ },
1827
+ {
1828
+ "epoch": 10.831353919239906,
1829
+ "eval_accuracy": 0.9541284403669725,
1830
+ "eval_loss": 0.152838334441185,
1831
+ "eval_runtime": 0.5705,
1832
+ "eval_samples_per_second": 1528.353,
1833
+ "eval_steps_per_second": 3.505,
1834
+ "step": 22800
1835
+ },
1836
+ {
1837
+ "epoch": 10.926365795724465,
1838
+ "grad_norm": 1.4312000274658203,
1839
+ "learning_rate": 5.0585357537602515e-09,
1840
+ "loss": 0.1249,
1841
+ "step": 23000
1842
+ },
1843
+ {
1844
+ "epoch": 10.926365795724465,
1845
+ "eval_accuracy": 0.9541284403669725,
1846
+ "eval_loss": 0.15280665457248688,
1847
+ "eval_runtime": 0.5525,
1848
+ "eval_samples_per_second": 1578.28,
1849
+ "eval_steps_per_second": 3.62,
1850
+ "step": 23000
1851
+ },
1852
+ {
1853
+ "epoch": 11.0,
1854
+ "step": 23155,
1855
+ "total_flos": 4.949323718936678e+16,
1856
+ "train_loss": 0.1508518957056924,
1857
+ "train_runtime": 9289.2724,
1858
+ "train_samples_per_second": 79.752,
1859
+ "train_steps_per_second": 2.493
1860
+ }
1861
+ ],
1862
+ "logging_steps": 200,
1863
+ "max_steps": 23155,
1864
+ "num_input_tokens_seen": 0,
1865
+ "num_train_epochs": 11,
1866
+ "save_steps": 200,
1867
+ "stateful_callbacks": {
1868
+ "TrainerControl": {
1869
+ "args": {
1870
+ "should_epoch_stop": false,
1871
+ "should_evaluate": false,
1872
+ "should_log": false,
1873
+ "should_save": true,
1874
+ "should_training_stop": true
1875
+ },
1876
+ "attributes": {}
1877
+ }
1878
+ },
1879
+ "total_flos": 4.949323718936678e+16,
1880
+ "train_batch_size": 32,
1881
+ "trial_name": null,
1882
+ "trial_params": null
1883
+ }
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=22d16h17m48/STS-B.tsv ADDED
@@ -0,0 +1,1380 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ index prediction
2
+ 0 3.249
3
+ 1 4.039
4
+ 2 5.000
5
+ 3 5.000
6
+ 4 2.089
7
+ 5 1.825
8
+ 6 3.781
9
+ 7 1.662
10
+ 8 2.353
11
+ 9 1.821
12
+ 10 1.821
13
+ 11 5.000
14
+ 12 0.773
15
+ 13 4.907
16
+ 14 1.928
17
+ 15 1.785
18
+ 16 4.740
19
+ 17 3.153
20
+ 18 3.686
21
+ 19 1.255
22
+ 20 1.707
23
+ 21 0.416
24
+ 22 5.000
25
+ 23 5.000
26
+ 24 1.447
27
+ 25 2.900
28
+ 26 0.969
29
+ 27 2.137
30
+ 28 0.788
31
+ 29 2.880
32
+ 30 2.720
33
+ 31 4.750
34
+ 32 0.980
35
+ 33 4.778
36
+ 34 3.539
37
+ 35 0.437
38
+ 36 0.715
39
+ 37 2.108
40
+ 38 4.991
41
+ 39 0.494
42
+ 40 3.992
43
+ 41 2.849
44
+ 42 3.246
45
+ 43 0.874
46
+ 44 1.623
47
+ 45 0.392
48
+ 46 0.148
49
+ 47 0.572
50
+ 48 5.000
51
+ 49 1.884
52
+ 50 0.950
53
+ 51 1.613
54
+ 52 0.359
55
+ 53 0.887
56
+ 54 1.700
57
+ 55 0.157
58
+ 56 3.998
59
+ 57 0.714
60
+ 58 4.461
61
+ 59 0.704
62
+ 60 5.000
63
+ 61 0.000
64
+ 62 4.291
65
+ 63 1.158
66
+ 64 3.689
67
+ 65 0.036
68
+ 66 0.224
69
+ 67 4.227
70
+ 68 0.111
71
+ 69 4.406
72
+ 70 2.589
73
+ 71 4.418
74
+ 72 0.105
75
+ 73 4.094
76
+ 74 4.266
77
+ 75 0.110
78
+ 76 0.550
79
+ 77 0.000
80
+ 78 2.720
81
+ 79 3.651
82
+ 80 0.113
83
+ 81 0.594
84
+ 82 3.037
85
+ 83 0.443
86
+ 84 0.196
87
+ 85 1.486
88
+ 86 3.998
89
+ 87 4.797
90
+ 88 3.232
91
+ 89 1.603
92
+ 90 1.638
93
+ 91 3.703
94
+ 92 1.023
95
+ 93 3.935
96
+ 94 2.949
97
+ 95 4.733
98
+ 96 4.904
99
+ 97 3.681
100
+ 98 2.562
101
+ 99 3.421
102
+ 100 0.000
103
+ 101 0.553
104
+ 102 3.026
105
+ 103 2.928
106
+ 104 3.983
107
+ 105 2.879
108
+ 106 0.072
109
+ 107 2.761
110
+ 108 1.973
111
+ 109 3.587
112
+ 110 3.362
113
+ 111 4.064
114
+ 112 3.427
115
+ 113 4.237
116
+ 114 0.000
117
+ 115 0.000
118
+ 116 4.162
119
+ 117 3.025
120
+ 118 0.040
121
+ 119 0.182
122
+ 120 0.545
123
+ 121 0.253
124
+ 122 0.032
125
+ 123 0.253
126
+ 124 0.000
127
+ 125 5.000
128
+ 126 0.479
129
+ 127 2.255
130
+ 128 5.000
131
+ 129 3.722
132
+ 130 4.966
133
+ 131 2.262
134
+ 132 3.108
135
+ 133 5.000
136
+ 134 3.520
137
+ 135 5.000
138
+ 136 5.000
139
+ 137 1.689
140
+ 138 1.806
141
+ 139 4.969
142
+ 140 5.000
143
+ 141 3.719
144
+ 142 4.459
145
+ 143 4.113
146
+ 144 2.148
147
+ 145 1.416
148
+ 146 4.338
149
+ 147 3.059
150
+ 148 1.722
151
+ 149 0.859
152
+ 150 1.826
153
+ 151 1.831
154
+ 152 3.789
155
+ 153 3.615
156
+ 154 3.643
157
+ 155 3.739
158
+ 156 0.268
159
+ 157 0.732
160
+ 158 4.170
161
+ 159 0.724
162
+ 160 5.000
163
+ 161 2.313
164
+ 162 0.847
165
+ 163 3.389
166
+ 164 4.896
167
+ 165 4.382
168
+ 166 4.190
169
+ 167 3.522
170
+ 168 4.684
171
+ 169 4.703
172
+ 170 1.603
173
+ 171 0.337
174
+ 172 4.668
175
+ 173 4.935
176
+ 174 4.210
177
+ 175 3.601
178
+ 176 3.005
179
+ 177 4.216
180
+ 178 0.185
181
+ 179 0.226
182
+ 180 0.288
183
+ 181 3.002
184
+ 182 3.467
185
+ 183 1.141
186
+ 184 1.589
187
+ 185 0.194
188
+ 186 0.000
189
+ 187 0.234
190
+ 188 0.297
191
+ 189 0.118
192
+ 190 4.064
193
+ 191 4.539
194
+ 192 1.593
195
+ 193 3.107
196
+ 194 3.935
197
+ 195 3.275
198
+ 196 3.159
199
+ 197 0.233
200
+ 198 0.000
201
+ 199 3.293
202
+ 200 0.442
203
+ 201 0.617
204
+ 202 0.215
205
+ 203 0.469
206
+ 204 0.212
207
+ 205 0.535
208
+ 206 3.759
209
+ 207 0.081
210
+ 208 2.309
211
+ 209 0.447
212
+ 210 0.697
213
+ 211 0.246
214
+ 212 0.111
215
+ 213 0.777
216
+ 214 3.340
217
+ 215 0.546
218
+ 216 1.380
219
+ 217 3.481
220
+ 218 3.852
221
+ 219 2.975
222
+ 220 0.733
223
+ 221 0.042
224
+ 222 0.543
225
+ 223 1.231
226
+ 224 0.096
227
+ 225 0.069
228
+ 226 2.784
229
+ 227 2.117
230
+ 228 0.000
231
+ 229 1.697
232
+ 230 0.000
233
+ 231 0.023
234
+ 232 4.750
235
+ 233 2.584
236
+ 234 4.261
237
+ 235 2.457
238
+ 236 0.053
239
+ 237 0.000
240
+ 238 0.000
241
+ 239 0.000
242
+ 240 2.857
243
+ 241 0.000
244
+ 242 0.000
245
+ 243 0.014
246
+ 244 3.378
247
+ 245 0.040
248
+ 246 3.060
249
+ 247 3.544
250
+ 248 0.000
251
+ 249 0.299
252
+ 250 4.081
253
+ 251 4.545
254
+ 252 3.969
255
+ 253 2.563
256
+ 254 0.588
257
+ 255 3.804
258
+ 256 0.273
259
+ 257 4.124
260
+ 258 3.134
261
+ 259 0.059
262
+ 260 5.000
263
+ 261 3.577
264
+ 262 4.201
265
+ 263 4.257
266
+ 264 3.016
267
+ 265 1.614
268
+ 266 2.638
269
+ 267 3.794
270
+ 268 1.356
271
+ 269 5.000
272
+ 270 4.354
273
+ 271 0.339
274
+ 272 3.681
275
+ 273 5.000
276
+ 274 4.179
277
+ 275 5.000
278
+ 276 0.267
279
+ 277 0.000
280
+ 278 4.120
281
+ 279 3.026
282
+ 280 0.367
283
+ 281 2.329
284
+ 282 3.916
285
+ 283 3.940
286
+ 284 3.850
287
+ 285 4.165
288
+ 286 1.438
289
+ 287 3.848
290
+ 288 3.463
291
+ 289 1.050
292
+ 290 4.997
293
+ 291 0.847
294
+ 292 2.495
295
+ 293 4.851
296
+ 294 0.000
297
+ 295 0.489
298
+ 296 1.995
299
+ 297 3.786
300
+ 298 3.043
301
+ 299 2.971
302
+ 300 5.000
303
+ 301 2.536
304
+ 302 2.085
305
+ 303 4.931
306
+ 304 3.367
307
+ 305 3.974
308
+ 306 4.544
309
+ 307 4.047
310
+ 308 3.687
311
+ 309 4.478
312
+ 310 1.809
313
+ 311 0.029
314
+ 312 0.598
315
+ 313 2.936
316
+ 314 3.542
317
+ 315 4.537
318
+ 316 3.612
319
+ 317 3.857
320
+ 318 0.000
321
+ 319 0.095
322
+ 320 3.913
323
+ 321 3.001
324
+ 322 4.379
325
+ 323 2.784
326
+ 324 2.748
327
+ 325 0.121
328
+ 326 2.644
329
+ 327 3.807
330
+ 328 4.724
331
+ 329 3.827
332
+ 330 5.000
333
+ 331 4.925
334
+ 332 3.815
335
+ 333 2.204
336
+ 334 3.602
337
+ 335 1.489
338
+ 336 3.942
339
+ 337 4.030
340
+ 338 0.000
341
+ 339 0.310
342
+ 340 4.063
343
+ 341 4.785
344
+ 342 4.907
345
+ 343 2.149
346
+ 344 3.856
347
+ 345 0.464
348
+ 346 4.385
349
+ 347 0.061
350
+ 348 4.025
351
+ 349 0.094
352
+ 350 4.562
353
+ 351 3.443
354
+ 352 0.833
355
+ 353 4.276
356
+ 354 3.085
357
+ 355 0.106
358
+ 356 2.577
359
+ 357 4.943
360
+ 358 0.099
361
+ 359 4.086
362
+ 360 4.969
363
+ 361 4.700
364
+ 362 2.655
365
+ 363 3.876
366
+ 364 4.724
367
+ 365 0.000
368
+ 366 1.205
369
+ 367 3.972
370
+ 368 4.210
371
+ 369 4.916
372
+ 370 3.547
373
+ 371 4.721
374
+ 372 0.000
375
+ 373 0.686
376
+ 374 1.103
377
+ 375 4.347
378
+ 376 2.145
379
+ 377 0.000
380
+ 378 2.978
381
+ 379 1.592
382
+ 380 3.845
383
+ 381 4.339
384
+ 382 2.675
385
+ 383 0.276
386
+ 384 3.631
387
+ 385 2.274
388
+ 386 5.000
389
+ 387 1.197
390
+ 388 3.171
391
+ 389 2.021
392
+ 390 4.174
393
+ 391 2.048
394
+ 392 0.000
395
+ 393 2.061
396
+ 394 2.706
397
+ 395 0.100
398
+ 396 0.279
399
+ 397 5.000
400
+ 398 3.044
401
+ 399 2.759
402
+ 400 2.578
403
+ 401 0.134
404
+ 402 2.357
405
+ 403 3.991
406
+ 404 3.470
407
+ 405 4.112
408
+ 406 0.181
409
+ 407 2.874
410
+ 408 2.397
411
+ 409 3.289
412
+ 410 4.660
413
+ 411 5.000
414
+ 412 1.771
415
+ 413 4.903
416
+ 414 0.397
417
+ 415 1.078
418
+ 416 0.000
419
+ 417 4.947
420
+ 418 4.454
421
+ 419 3.993
422
+ 420 5.000
423
+ 421 1.779
424
+ 422 4.652
425
+ 423 1.094
426
+ 424 2.244
427
+ 425 3.574
428
+ 426 0.000
429
+ 427 5.000
430
+ 428 2.956
431
+ 429 5.000
432
+ 430 0.620
433
+ 431 5.000
434
+ 432 0.234
435
+ 433 3.698
436
+ 434 0.093
437
+ 435 2.650
438
+ 436 2.529
439
+ 437 5.000
440
+ 438 3.527
441
+ 439 4.611
442
+ 440 3.183
443
+ 441 2.926
444
+ 442 0.000
445
+ 443 2.840
446
+ 444 3.927
447
+ 445 0.000
448
+ 446 1.311
449
+ 447 3.886
450
+ 448 0.364
451
+ 449 1.638
452
+ 450 3.789
453
+ 451 3.104
454
+ 452 0.222
455
+ 453 0.129
456
+ 454 4.252
457
+ 455 2.092
458
+ 456 2.051
459
+ 457 5.000
460
+ 458 1.114
461
+ 459 2.809
462
+ 460 0.128
463
+ 461 0.777
464
+ 462 0.000
465
+ 463 1.994
466
+ 464 3.410
467
+ 465 4.161
468
+ 466 2.034
469
+ 467 4.961
470
+ 468 4.760
471
+ 469 0.205
472
+ 470 2.311
473
+ 471 4.676
474
+ 472 2.098
475
+ 473 4.890
476
+ 474 3.196
477
+ 475 2.660
478
+ 476 0.810
479
+ 477 3.212
480
+ 478 4.373
481
+ 479 0.000
482
+ 480 4.988
483
+ 481 2.466
484
+ 482 4.075
485
+ 483 0.000
486
+ 484 4.669
487
+ 485 0.172
488
+ 486 3.649
489
+ 487 4.506
490
+ 488 3.671
491
+ 489 5.000
492
+ 490 4.021
493
+ 491 4.321
494
+ 492 3.501
495
+ 493 3.620
496
+ 494 3.852
497
+ 495 2.429
498
+ 496 1.240
499
+ 497 5.000
500
+ 498 0.127
501
+ 499 5.000
502
+ 500 1.736
503
+ 501 2.719
504
+ 502 1.596
505
+ 503 2.345
506
+ 504 1.323
507
+ 505 3.376
508
+ 506 3.426
509
+ 507 0.829
510
+ 508 3.067
511
+ 509 0.000
512
+ 510 1.394
513
+ 511 1.260
514
+ 512 2.451
515
+ 513 0.278
516
+ 514 1.088
517
+ 515 0.186
518
+ 516 1.410
519
+ 517 2.263
520
+ 518 1.329
521
+ 519 3.807
522
+ 520 1.517
523
+ 521 2.463
524
+ 522 4.698
525
+ 523 1.104
526
+ 524 2.886
527
+ 525 3.702
528
+ 526 1.690
529
+ 527 0.716
530
+ 528 2.476
531
+ 529 0.712
532
+ 530 0.792
533
+ 531 0.000
534
+ 532 1.028
535
+ 533 0.499
536
+ 534 0.115
537
+ 535 0.279
538
+ 536 2.158
539
+ 537 1.105
540
+ 538 2.717
541
+ 539 3.007
542
+ 540 2.577
543
+ 541 1.135
544
+ 542 1.897
545
+ 543 3.301
546
+ 544 4.377
547
+ 545 5.000
548
+ 546 1.210
549
+ 547 2.319
550
+ 548 3.093
551
+ 549 4.325
552
+ 550 3.810
553
+ 551 1.182
554
+ 552 3.071
555
+ 553 2.433
556
+ 554 1.658
557
+ 555 2.453
558
+ 556 3.145
559
+ 557 0.850
560
+ 558 5.000
561
+ 559 4.108
562
+ 560 3.916
563
+ 561 0.000
564
+ 562 1.538
565
+ 563 0.899
566
+ 564 0.691
567
+ 565 2.448
568
+ 566 5.000
569
+ 567 2.864
570
+ 568 0.366
571
+ 569 2.075
572
+ 570 1.816
573
+ 571 0.927
574
+ 572 0.147
575
+ 573 3.325
576
+ 574 4.829
577
+ 575 2.438
578
+ 576 4.836
579
+ 577 3.407
580
+ 578 2.215
581
+ 579 1.245
582
+ 580 1.153
583
+ 581 2.954
584
+ 582 2.891
585
+ 583 1.259
586
+ 584 2.765
587
+ 585 4.118
588
+ 586 0.602
589
+ 587 5.000
590
+ 588 1.367
591
+ 589 1.618
592
+ 590 1.723
593
+ 591 2.058
594
+ 592 4.883
595
+ 593 1.510
596
+ 594 0.674
597
+ 595 3.677
598
+ 596 2.036
599
+ 597 0.316
600
+ 598 3.937
601
+ 599 4.820
602
+ 600 2.132
603
+ 601 0.328
604
+ 602 3.646
605
+ 603 1.738
606
+ 604 0.994
607
+ 605 4.043
608
+ 606 1.766
609
+ 607 1.508
610
+ 608 3.161
611
+ 609 2.613
612
+ 610 5.000
613
+ 611 2.895
614
+ 612 1.087
615
+ 613 0.597
616
+ 614 3.258
617
+ 615 1.960
618
+ 616 1.747
619
+ 617 0.345
620
+ 618 2.593
621
+ 619 4.626
622
+ 620 0.592
623
+ 621 2.739
624
+ 622 0.847
625
+ 623 5.000
626
+ 624 2.298
627
+ 625 2.977
628
+ 626 2.734
629
+ 627 0.149
630
+ 628 0.468
631
+ 629 2.092
632
+ 630 1.584
633
+ 631 5.000
634
+ 632 2.797
635
+ 633 2.926
636
+ 634 1.315
637
+ 635 2.688
638
+ 636 4.210
639
+ 637 2.642
640
+ 638 2.741
641
+ 639 3.455
642
+ 640 1.222
643
+ 641 1.497
644
+ 642 2.637
645
+ 643 2.768
646
+ 644 1.661
647
+ 645 4.783
648
+ 646 3.102
649
+ 647 4.109
650
+ 648 4.768
651
+ 649 3.552
652
+ 650 1.621
653
+ 651 4.509
654
+ 652 0.891
655
+ 653 1.348
656
+ 654 1.398
657
+ 655 3.466
658
+ 656 4.564
659
+ 657 1.326
660
+ 658 1.965
661
+ 659 0.806
662
+ 660 1.639
663
+ 661 1.452
664
+ 662 0.346
665
+ 663 2.215
666
+ 664 5.000
667
+ 665 0.959
668
+ 666 3.482
669
+ 667 3.056
670
+ 668 2.762
671
+ 669 2.436
672
+ 670 2.275
673
+ 671 3.832
674
+ 672 1.844
675
+ 673 4.867
676
+ 674 4.899
677
+ 675 0.678
678
+ 676 1.831
679
+ 677 3.029
680
+ 678 1.543
681
+ 679 2.853
682
+ 680 3.309
683
+ 681 1.258
684
+ 682 2.989
685
+ 683 2.162
686
+ 684 4.009
687
+ 685 1.950
688
+ 686 4.053
689
+ 687 1.889
690
+ 688 4.108
691
+ 689 4.606
692
+ 690 2.099
693
+ 691 2.480
694
+ 692 5.000
695
+ 693 1.317
696
+ 694 3.424
697
+ 695 1.408
698
+ 696 2.421
699
+ 697 0.566
700
+ 698 1.545
701
+ 699 4.386
702
+ 700 3.910
703
+ 701 2.358
704
+ 702 4.877
705
+ 703 4.752
706
+ 704 2.765
707
+ 705 2.021
708
+ 706 2.188
709
+ 707 0.905
710
+ 708 3.483
711
+ 709 2.028
712
+ 710 4.167
713
+ 711 1.366
714
+ 712 2.291
715
+ 713 1.508
716
+ 714 1.808
717
+ 715 1.847
718
+ 716 2.179
719
+ 717 4.278
720
+ 718 2.117
721
+ 719 5.000
722
+ 720 5.000
723
+ 721 2.510
724
+ 722 1.842
725
+ 723 1.417
726
+ 724 4.141
727
+ 725 1.093
728
+ 726 3.336
729
+ 727 4.674
730
+ 728 2.672
731
+ 729 3.462
732
+ 730 3.111
733
+ 731 1.770
734
+ 732 2.118
735
+ 733 0.219
736
+ 734 2.934
737
+ 735 1.434
738
+ 736 1.538
739
+ 737 4.610
740
+ 738 2.774
741
+ 739 2.762
742
+ 740 4.712
743
+ 741 1.668
744
+ 742 1.430
745
+ 743 3.591
746
+ 744 1.752
747
+ 745 3.211
748
+ 746 0.821
749
+ 747 1.487
750
+ 748 4.042
751
+ 749 1.498
752
+ 750 2.546
753
+ 751 1.814
754
+ 752 3.063
755
+ 753 1.049
756
+ 754 1.423
757
+ 755 4.705
758
+ 756 4.602
759
+ 757 3.785
760
+ 758 1.701
761
+ 759 3.438
762
+ 760 3.676
763
+ 761 0.264
764
+ 762 0.478
765
+ 763 1.676
766
+ 764 3.458
767
+ 765 4.272
768
+ 766 3.361
769
+ 767 1.957
770
+ 768 4.066
771
+ 769 1.752
772
+ 770 2.985
773
+ 771 1.361
774
+ 772 1.575
775
+ 773 4.978
776
+ 774 4.639
777
+ 775 4.315
778
+ 776 1.650
779
+ 777 0.865
780
+ 778 2.184
781
+ 779 0.719
782
+ 780 2.548
783
+ 781 2.284
784
+ 782 5.000
785
+ 783 4.725
786
+ 784 3.115
787
+ 785 0.706
788
+ 786 0.742
789
+ 787 2.737
790
+ 788 5.000
791
+ 789 2.480
792
+ 790 2.416
793
+ 791 2.489
794
+ 792 1.668
795
+ 793 1.743
796
+ 794 2.370
797
+ 795 3.579
798
+ 796 1.634
799
+ 797 4.193
800
+ 798 0.699
801
+ 799 1.336
802
+ 800 1.612
803
+ 801 2.276
804
+ 802 1.174
805
+ 803 0.399
806
+ 804 1.694
807
+ 805 5.000
808
+ 806 2.449
809
+ 807 2.804
810
+ 808 4.201
811
+ 809 2.308
812
+ 810 3.783
813
+ 811 1.998
814
+ 812 1.904
815
+ 813 4.316
816
+ 814 4.449
817
+ 815 1.023
818
+ 816 3.197
819
+ 817 1.596
820
+ 818 3.691
821
+ 819 1.967
822
+ 820 2.412
823
+ 821 2.265
824
+ 822 1.160
825
+ 823 4.025
826
+ 824 2.577
827
+ 825 1.702
828
+ 826 2.960
829
+ 827 0.979
830
+ 828 3.533
831
+ 829 5.000
832
+ 830 2.405
833
+ 831 1.931
834
+ 832 1.447
835
+ 833 4.770
836
+ 834 1.892
837
+ 835 2.804
838
+ 836 4.554
839
+ 837 1.598
840
+ 838 1.589
841
+ 839 3.637
842
+ 840 3.245
843
+ 841 1.677
844
+ 842 4.455
845
+ 843 3.777
846
+ 844 4.632
847
+ 845 4.058
848
+ 846 2.959
849
+ 847 2.260
850
+ 848 1.965
851
+ 849 3.267
852
+ 850 4.626
853
+ 851 1.569
854
+ 852 4.078
855
+ 853 4.718
856
+ 854 5.000
857
+ 855 1.781
858
+ 856 0.728
859
+ 857 2.658
860
+ 858 2.276
861
+ 859 3.188
862
+ 860 4.061
863
+ 861 4.624
864
+ 862 3.778
865
+ 863 0.170
866
+ 864 0.841
867
+ 865 0.893
868
+ 866 0.834
869
+ 867 0.675
870
+ 868 0.808
871
+ 869 4.785
872
+ 870 4.145
873
+ 871 2.515
874
+ 872 2.668
875
+ 873 4.453
876
+ 874 1.954
877
+ 875 4.167
878
+ 876 2.117
879
+ 877 4.562
880
+ 878 4.214
881
+ 879 1.975
882
+ 880 4.189
883
+ 881 3.445
884
+ 882 3.328
885
+ 883 4.525
886
+ 884 1.531
887
+ 885 3.981
888
+ 886 3.263
889
+ 887 4.358
890
+ 888 3.357
891
+ 889 3.265
892
+ 890 5.000
893
+ 891 3.391
894
+ 892 1.847
895
+ 893 3.705
896
+ 894 4.298
897
+ 895 4.439
898
+ 896 3.481
899
+ 897 2.904
900
+ 898 1.280
901
+ 899 3.560
902
+ 900 3.922
903
+ 901 4.216
904
+ 902 3.488
905
+ 903 4.595
906
+ 904 4.745
907
+ 905 3.269
908
+ 906 3.244
909
+ 907 4.248
910
+ 908 3.841
911
+ 909 3.063
912
+ 910 3.863
913
+ 911 1.944
914
+ 912 2.801
915
+ 913 4.339
916
+ 914 3.448
917
+ 915 4.804
918
+ 916 3.812
919
+ 917 2.747
920
+ 918 4.303
921
+ 919 3.913
922
+ 920 3.559
923
+ 921 3.850
924
+ 922 4.977
925
+ 923 4.452
926
+ 924 4.133
927
+ 925 2.286
928
+ 926 3.427
929
+ 927 2.510
930
+ 928 3.854
931
+ 929 3.370
932
+ 930 3.721
933
+ 931 3.690
934
+ 932 4.332
935
+ 933 2.988
936
+ 934 2.048
937
+ 935 2.224
938
+ 936 3.599
939
+ 937 3.603
940
+ 938 3.258
941
+ 939 2.550
942
+ 940 4.365
943
+ 941 3.655
944
+ 942 3.482
945
+ 943 3.140
946
+ 944 3.912
947
+ 945 4.553
948
+ 946 4.058
949
+ 947 3.984
950
+ 948 3.676
951
+ 949 2.119
952
+ 950 3.254
953
+ 951 4.662
954
+ 952 4.370
955
+ 953 2.979
956
+ 954 3.040
957
+ 955 1.928
958
+ 956 3.484
959
+ 957 1.234
960
+ 958 4.410
961
+ 959 3.004
962
+ 960 3.335
963
+ 961 3.738
964
+ 962 4.375
965
+ 963 2.873
966
+ 964 3.621
967
+ 965 3.550
968
+ 966 4.177
969
+ 967 4.098
970
+ 968 3.379
971
+ 969 3.435
972
+ 970 2.818
973
+ 971 3.158
974
+ 972 3.867
975
+ 973 3.290
976
+ 974 3.918
977
+ 975 2.584
978
+ 976 3.277
979
+ 977 4.816
980
+ 978 4.364
981
+ 979 3.882
982
+ 980 2.314
983
+ 981 4.557
984
+ 982 2.826
985
+ 983 3.020
986
+ 984 2.537
987
+ 985 3.323
988
+ 986 4.503
989
+ 987 3.239
990
+ 988 4.095
991
+ 989 4.153
992
+ 990 4.509
993
+ 991 1.822
994
+ 992 3.590
995
+ 993 3.507
996
+ 994 3.188
997
+ 995 2.956
998
+ 996 1.851
999
+ 997 4.437
1000
+ 998 4.174
1001
+ 999 3.138
1002
+ 1000 3.315
1003
+ 1001 4.002
1004
+ 1002 4.140
1005
+ 1003 3.499
1006
+ 1004 1.682
1007
+ 1005 2.575
1008
+ 1006 4.416
1009
+ 1007 3.284
1010
+ 1008 1.950
1011
+ 1009 3.570
1012
+ 1010 4.202
1013
+ 1011 4.181
1014
+ 1012 4.895
1015
+ 1013 2.351
1016
+ 1014 4.305
1017
+ 1015 3.885
1018
+ 1016 4.330
1019
+ 1017 2.996
1020
+ 1018 4.858
1021
+ 1019 4.113
1022
+ 1020 3.308
1023
+ 1021 4.303
1024
+ 1022 3.842
1025
+ 1023 1.854
1026
+ 1024 3.538
1027
+ 1025 3.522
1028
+ 1026 3.291
1029
+ 1027 2.009
1030
+ 1028 3.879
1031
+ 1029 1.858
1032
+ 1030 4.687
1033
+ 1031 4.207
1034
+ 1032 4.396
1035
+ 1033 2.965
1036
+ 1034 5.000
1037
+ 1035 3.937
1038
+ 1036 3.557
1039
+ 1037 4.146
1040
+ 1038 4.593
1041
+ 1039 4.917
1042
+ 1040 3.256
1043
+ 1041 4.230
1044
+ 1042 3.301
1045
+ 1043 3.353
1046
+ 1044 4.064
1047
+ 1045 2.640
1048
+ 1046 1.519
1049
+ 1047 2.907
1050
+ 1048 4.304
1051
+ 1049 4.805
1052
+ 1050 3.944
1053
+ 1051 3.788
1054
+ 1052 4.051
1055
+ 1053 3.001
1056
+ 1054 2.879
1057
+ 1055 4.526
1058
+ 1056 4.333
1059
+ 1057 3.211
1060
+ 1058 4.644
1061
+ 1059 3.500
1062
+ 1060 3.105
1063
+ 1061 3.515
1064
+ 1062 2.790
1065
+ 1063 4.508
1066
+ 1064 3.185
1067
+ 1065 3.140
1068
+ 1066 4.125
1069
+ 1067 2.912
1070
+ 1068 3.153
1071
+ 1069 3.937
1072
+ 1070 5.000
1073
+ 1071 2.761
1074
+ 1072 3.135
1075
+ 1073 0.868
1076
+ 1074 3.284
1077
+ 1075 2.114
1078
+ 1076 3.896
1079
+ 1077 3.790
1080
+ 1078 3.853
1081
+ 1079 2.036
1082
+ 1080 4.179
1083
+ 1081 3.606
1084
+ 1082 3.691
1085
+ 1083 0.997
1086
+ 1084 3.687
1087
+ 1085 4.041
1088
+ 1086 2.481
1089
+ 1087 4.156
1090
+ 1088 1.622
1091
+ 1089 4.260
1092
+ 1090 3.419
1093
+ 1091 4.818
1094
+ 1092 3.860
1095
+ 1093 3.862
1096
+ 1094 3.310
1097
+ 1095 3.721
1098
+ 1096 3.309
1099
+ 1097 4.095
1100
+ 1098 3.100
1101
+ 1099 2.119
1102
+ 1100 2.790
1103
+ 1101 3.796
1104
+ 1102 3.629
1105
+ 1103 3.676
1106
+ 1104 3.853
1107
+ 1105 3.623
1108
+ 1106 3.401
1109
+ 1107 2.041
1110
+ 1108 3.744
1111
+ 1109 2.466
1112
+ 1110 1.185
1113
+ 1111 4.775
1114
+ 1112 3.755
1115
+ 1113 4.843
1116
+ 1114 3.999
1117
+ 1115 3.662
1118
+ 1116 2.915
1119
+ 1117 3.843
1120
+ 1118 4.533
1121
+ 1119 3.225
1122
+ 1120 2.646
1123
+ 1121 0.915
1124
+ 1122 4.287
1125
+ 1123 3.370
1126
+ 1124 3.191
1127
+ 1125 4.617
1128
+ 1126 3.602
1129
+ 1127 4.459
1130
+ 1128 4.272
1131
+ 1129 1.685
1132
+ 1130 1.480
1133
+ 1131 4.457
1134
+ 1132 5.000
1135
+ 1133 0.698
1136
+ 1134 2.631
1137
+ 1135 2.475
1138
+ 1136 2.416
1139
+ 1137 3.374
1140
+ 1138 0.608
1141
+ 1139 4.980
1142
+ 1140 1.429
1143
+ 1141 0.955
1144
+ 1142 2.288
1145
+ 1143 4.598
1146
+ 1144 3.706
1147
+ 1145 1.968
1148
+ 1146 1.039
1149
+ 1147 4.549
1150
+ 1148 4.166
1151
+ 1149 4.892
1152
+ 1150 0.000
1153
+ 1151 0.182
1154
+ 1152 5.000
1155
+ 1153 4.997
1156
+ 1154 5.000
1157
+ 1155 3.953
1158
+ 1156 4.155
1159
+ 1157 0.000
1160
+ 1158 4.589
1161
+ 1159 1.276
1162
+ 1160 3.932
1163
+ 1161 3.698
1164
+ 1162 0.814
1165
+ 1163 4.971
1166
+ 1164 4.760
1167
+ 1165 0.878
1168
+ 1166 4.203
1169
+ 1167 1.511
1170
+ 1168 3.605
1171
+ 1169 0.172
1172
+ 1170 2.407
1173
+ 1171 4.404
1174
+ 1172 2.728
1175
+ 1173 3.770
1176
+ 1174 0.133
1177
+ 1175 0.019
1178
+ 1176 5.000
1179
+ 1177 1.438
1180
+ 1178 3.340
1181
+ 1179 0.408
1182
+ 1180 0.766
1183
+ 1181 4.026
1184
+ 1182 1.856
1185
+ 1183 2.802
1186
+ 1184 0.606
1187
+ 1185 4.362
1188
+ 1186 5.000
1189
+ 1187 3.945
1190
+ 1188 3.202
1191
+ 1189 4.818
1192
+ 1190 3.579
1193
+ 1191 4.472
1194
+ 1192 1.567
1195
+ 1193 1.966
1196
+ 1194 0.910
1197
+ 1195 2.217
1198
+ 1196 4.982
1199
+ 1197 4.557
1200
+ 1198 3.641
1201
+ 1199 3.935
1202
+ 1200 2.755
1203
+ 1201 0.985
1204
+ 1202 3.625
1205
+ 1203 4.787
1206
+ 1204 4.040
1207
+ 1205 3.255
1208
+ 1206 5.000
1209
+ 1207 4.404
1210
+ 1208 2.172
1211
+ 1209 2.200
1212
+ 1210 2.528
1213
+ 1211 4.269
1214
+ 1212 4.317
1215
+ 1213 2.295
1216
+ 1214 5.000
1217
+ 1215 4.601
1218
+ 1216 1.236
1219
+ 1217 2.433
1220
+ 1218 1.636
1221
+ 1219 1.870
1222
+ 1220 1.466
1223
+ 1221 3.228
1224
+ 1222 4.411
1225
+ 1223 3.282
1226
+ 1224 1.518
1227
+ 1225 3.458
1228
+ 1226 3.152
1229
+ 1227 4.821
1230
+ 1228 4.760
1231
+ 1229 3.130
1232
+ 1230 2.867
1233
+ 1231 3.517
1234
+ 1232 3.414
1235
+ 1233 4.903
1236
+ 1234 4.427
1237
+ 1235 2.205
1238
+ 1236 2.586
1239
+ 1237 0.993
1240
+ 1238 1.708
1241
+ 1239 0.964
1242
+ 1240 1.073
1243
+ 1241 1.784
1244
+ 1242 2.730
1245
+ 1243 2.936
1246
+ 1244 0.000
1247
+ 1245 5.000
1248
+ 1246 4.100
1249
+ 1247 1.290
1250
+ 1248 0.213
1251
+ 1249 5.000
1252
+ 1250 4.302
1253
+ 1251 3.737
1254
+ 1252 4.431
1255
+ 1253 2.457
1256
+ 1254 0.000
1257
+ 1255 4.751
1258
+ 1256 2.445
1259
+ 1257 4.344
1260
+ 1258 3.932
1261
+ 1259 4.482
1262
+ 1260 3.503
1263
+ 1261 1.163
1264
+ 1262 4.277
1265
+ 1263 1.186
1266
+ 1264 3.593
1267
+ 1265 4.110
1268
+ 1266 4.113
1269
+ 1267 4.775
1270
+ 1268 1.211
1271
+ 1269 1.035
1272
+ 1270 3.077
1273
+ 1271 3.663
1274
+ 1272 0.977
1275
+ 1273 4.231
1276
+ 1274 4.182
1277
+ 1275 2.211
1278
+ 1276 0.517
1279
+ 1277 3.478
1280
+ 1278 0.778
1281
+ 1279 4.763
1282
+ 1280 0.244
1283
+ 1281 0.124
1284
+ 1282 3.175
1285
+ 1283 4.775
1286
+ 1284 2.221
1287
+ 1285 4.380
1288
+ 1286 4.251
1289
+ 1287 1.600
1290
+ 1288 0.000
1291
+ 1289 2.137
1292
+ 1290 5.000
1293
+ 1291 0.494
1294
+ 1292 4.124
1295
+ 1293 3.458
1296
+ 1294 1.646
1297
+ 1295 0.229
1298
+ 1296 4.989
1299
+ 1297 1.233
1300
+ 1298 0.818
1301
+ 1299 4.843
1302
+ 1300 2.777
1303
+ 1301 2.338
1304
+ 1302 2.890
1305
+ 1303 2.175
1306
+ 1304 2.479
1307
+ 1305 0.280
1308
+ 1306 2.107
1309
+ 1307 0.956
1310
+ 1308 2.608
1311
+ 1309 3.204
1312
+ 1310 3.630
1313
+ 1311 4.839
1314
+ 1312 0.942
1315
+ 1313 3.304
1316
+ 1314 2.648
1317
+ 1315 4.755
1318
+ 1316 5.000
1319
+ 1317 5.000
1320
+ 1318 2.049
1321
+ 1319 0.015
1322
+ 1320 5.000
1323
+ 1321 1.906
1324
+ 1322 5.000
1325
+ 1323 2.714
1326
+ 1324 5.000
1327
+ 1325 4.754
1328
+ 1326 5.000
1329
+ 1327 3.500
1330
+ 1328 0.559
1331
+ 1329 4.848
1332
+ 1330 4.606
1333
+ 1331 2.601
1334
+ 1332 1.893
1335
+ 1333 1.716
1336
+ 1334 2.539
1337
+ 1335 4.658
1338
+ 1336 4.612
1339
+ 1337 4.765
1340
+ 1338 0.565
1341
+ 1339 2.111
1342
+ 1340 3.704
1343
+ 1341 4.491
1344
+ 1342 2.885
1345
+ 1343 0.389
1346
+ 1344 0.827
1347
+ 1345 2.349
1348
+ 1346 4.714
1349
+ 1347 5.000
1350
+ 1348 5.000
1351
+ 1349 2.385
1352
+ 1350 5.000
1353
+ 1351 5.000
1354
+ 1352 4.982
1355
+ 1353 5.000
1356
+ 1354 4.481
1357
+ 1355 2.698
1358
+ 1356 2.382
1359
+ 1357 5.000
1360
+ 1358 4.090
1361
+ 1359 3.104
1362
+ 1360 0.780
1363
+ 1361 4.055
1364
+ 1362 1.122
1365
+ 1363 2.078
1366
+ 1364 5.000
1367
+ 1365 1.982
1368
+ 1366 1.761
1369
+ 1367 2.317
1370
+ 1368 0.202
1371
+ 1369 0.333
1372
+ 1370 0.916
1373
+ 1371 1.224
1374
+ 1372 1.393
1375
+ 1373 1.922
1376
+ 1374 0.570
1377
+ 1375 1.103
1378
+ 1376 1.858
1379
+ 1377 0.088
1380
+ 1378 1.089
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=22d16h17m48/all_results.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 8.0,
3
+ "eval_combined_score": 0.9131120094564884,
4
+ "eval_loss": 0.42496126890182495,
5
+ "eval_pearson": 0.9132190942764282,
6
+ "eval_runtime": 0.8154,
7
+ "eval_samples": 1500,
8
+ "eval_samples_per_second": 1839.695,
9
+ "eval_spearmanr": 0.9130049246365486,
10
+ "eval_steps_per_second": 3.679
11
+ }
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=22d16h17m48/eval_results.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 8.0,
3
+ "eval_combined_score": 0.9131120094564884,
4
+ "eval_loss": 0.42496126890182495,
5
+ "eval_pearson": 0.9132190942764282,
6
+ "eval_runtime": 0.8154,
7
+ "eval_samples": 1500,
8
+ "eval_samples_per_second": 1839.695,
9
+ "eval_spearmanr": 0.9130049246365486,
10
+ "eval_steps_per_second": 3.679
11
+ }
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=22d16h17m48/ft/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[MASK]": 128000
3
+ }
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=22d16h17m48/ft/special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": {
9
+ "content": "[UNK]",
10
+ "lstrip": false,
11
+ "normalized": true,
12
+ "rstrip": false,
13
+ "single_word": false
14
+ }
15
+ }
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=22d16h17m48/ft/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=22d16h17m48/ft/tokenizer_config.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[CLS]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[SEP]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[UNK]",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "128000": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "[CLS]",
45
+ "clean_up_tokenization_spaces": false,
46
+ "cls_token": "[CLS]",
47
+ "do_lower_case": false,
48
+ "eos_token": "[SEP]",
49
+ "extra_special_tokens": {},
50
+ "mask_token": "[MASK]",
51
+ "model_max_length": 512,
52
+ "pad_token": "[PAD]",
53
+ "padding_side": "right",
54
+ "sep_token": "[SEP]",
55
+ "sp_model_kwargs": {},
56
+ "split_by_punct": false,
57
+ "tokenizer_class": "DebertaV2Tokenizer",
58
+ "unk_token": "[UNK]",
59
+ "vocab_type": "spm"
60
+ }
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=22d16h17m48/ft2/README.md ADDED
@@ -0,0 +1,205 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: microsoft/deberta-v3-base
3
+ library_name: peft
4
+ tags:
5
+ - base_model:adapter:microsoft/deberta-v3-base
6
+ - transformers
7
+ ---
8
+
9
+ # Model Card for Model ID
10
+
11
+ <!-- Provide a quick summary of what the model is/does. -->
12
+
13
+
14
+
15
+ ## Model Details
16
+
17
+ ### Model Description
18
+
19
+ <!-- Provide a longer summary of what this model is. -->
20
+
21
+
22
+
23
+ - **Developed by:** [More Information Needed]
24
+ - **Funded by [optional]:** [More Information Needed]
25
+ - **Shared by [optional]:** [More Information Needed]
26
+ - **Model type:** [More Information Needed]
27
+ - **Language(s) (NLP):** [More Information Needed]
28
+ - **License:** [More Information Needed]
29
+ - **Finetuned from model [optional]:** [More Information Needed]
30
+
31
+ ### Model Sources [optional]
32
+
33
+ <!-- Provide the basic links for the model. -->
34
+
35
+ - **Repository:** [More Information Needed]
36
+ - **Paper [optional]:** [More Information Needed]
37
+ - **Demo [optional]:** [More Information Needed]
38
+
39
+ ## Uses
40
+
41
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
42
+
43
+ ### Direct Use
44
+
45
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
46
+
47
+ [More Information Needed]
48
+
49
+ ### Downstream Use [optional]
50
+
51
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
52
+
53
+ [More Information Needed]
54
+
55
+ ### Out-of-Scope Use
56
+
57
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
58
+
59
+ [More Information Needed]
60
+
61
+ ## Bias, Risks, and Limitations
62
+
63
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
64
+
65
+ [More Information Needed]
66
+
67
+ ### Recommendations
68
+
69
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
70
+
71
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
72
+
73
+ ## How to Get Started with the Model
74
+
75
+ Use the code below to get started with the model.
76
+
77
+ [More Information Needed]
78
+
79
+ ## Training Details
80
+
81
+ ### Training Data
82
+
83
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
84
+
85
+ [More Information Needed]
86
+
87
+ ### Training Procedure
88
+
89
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
90
+
91
+ #### Preprocessing [optional]
92
+
93
+ [More Information Needed]
94
+
95
+
96
+ #### Training Hyperparameters
97
+
98
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
99
+
100
+ #### Speeds, Sizes, Times [optional]
101
+
102
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
103
+
104
+ [More Information Needed]
105
+
106
+ ## Evaluation
107
+
108
+ <!-- This section describes the evaluation protocols and provides the results. -->
109
+
110
+ ### Testing Data, Factors & Metrics
111
+
112
+ #### Testing Data
113
+
114
+ <!-- This should link to a Dataset Card if possible. -->
115
+
116
+ [More Information Needed]
117
+
118
+ #### Factors
119
+
120
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
121
+
122
+ [More Information Needed]
123
+
124
+ #### Metrics
125
+
126
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
127
+
128
+ [More Information Needed]
129
+
130
+ ### Results
131
+
132
+ [More Information Needed]
133
+
134
+ #### Summary
135
+
136
+
137
+
138
+ ## Model Examination [optional]
139
+
140
+ <!-- Relevant interpretability work for the model goes here -->
141
+
142
+ [More Information Needed]
143
+
144
+ ## Environmental Impact
145
+
146
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
147
+
148
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
149
+
150
+ - **Hardware Type:** [More Information Needed]
151
+ - **Hours used:** [More Information Needed]
152
+ - **Cloud Provider:** [More Information Needed]
153
+ - **Compute Region:** [More Information Needed]
154
+ - **Carbon Emitted:** [More Information Needed]
155
+
156
+ ## Technical Specifications [optional]
157
+
158
+ ### Model Architecture and Objective
159
+
160
+ [More Information Needed]
161
+
162
+ ### Compute Infrastructure
163
+
164
+ [More Information Needed]
165
+
166
+ #### Hardware
167
+
168
+ [More Information Needed]
169
+
170
+ #### Software
171
+
172
+ [More Information Needed]
173
+
174
+ ## Citation [optional]
175
+
176
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
177
+
178
+ **BibTeX:**
179
+
180
+ [More Information Needed]
181
+
182
+ **APA:**
183
+
184
+ [More Information Needed]
185
+
186
+ ## Glossary [optional]
187
+
188
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
189
+
190
+ [More Information Needed]
191
+
192
+ ## More Information [optional]
193
+
194
+ [More Information Needed]
195
+
196
+ ## Model Card Authors [optional]
197
+
198
+ [More Information Needed]
199
+
200
+ ## Model Card Contact
201
+
202
+ [More Information Needed]
203
+ ### Framework versions
204
+
205
+ - PEFT 0.18.0
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=22d16h17m48/ft2/adapter_config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": {
3
+ "base_model_class": "DebertaV2ForSequenceClassification",
4
+ "parent_library": "transformers.models.deberta_v2.modeling_deberta_v2"
5
+ },
6
+ "base_model_name_or_path": "microsoft/deberta-v3-base",
7
+ "bias": "none",
8
+ "boft_block_num": 0,
9
+ "boft_block_size": 4,
10
+ "boft_dropout": 0.05,
11
+ "boft_n_butterfly_factor": 2,
12
+ "exclude_modules": null,
13
+ "fan_in_fan_out": false,
14
+ "inference_mode": true,
15
+ "init_weights": true,
16
+ "layers_pattern": null,
17
+ "layers_to_transform": null,
18
+ "modules_to_save": [
19
+ "classifier",
20
+ "pooler"
21
+ ],
22
+ "peft_type": "BOFT",
23
+ "peft_version": "0.18.0",
24
+ "revision": null,
25
+ "target_modules": [
26
+ "value_proj",
27
+ "intermediate.dense",
28
+ "output.dense",
29
+ "query_proj",
30
+ "attention.output.dense",
31
+ "key_proj"
32
+ ],
33
+ "task_type": null
34
+ }
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=22d16h17m48/trainer_state.json ADDED
@@ -0,0 +1,169 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 1200,
3
+ "best_metric": 0.9131120094564884,
4
+ "best_model_checkpoint": "./glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=22d16h17m48/checkpoint-1200",
5
+ "epoch": 8.0,
6
+ "eval_steps": 200,
7
+ "global_step": 1440,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 1.1111111111111112,
14
+ "grad_norm": 5.703500747680664,
15
+ "learning_rate": 0.0006801144205835404,
16
+ "loss": 1.4853,
17
+ "step": 200
18
+ },
19
+ {
20
+ "epoch": 1.1111111111111112,
21
+ "eval_combined_score": 0.8886370993432655,
22
+ "eval_loss": 0.6092813611030579,
23
+ "eval_pearson": 0.8852584362030029,
24
+ "eval_runtime": 0.8185,
25
+ "eval_samples_per_second": 1832.677,
26
+ "eval_spearmanr": 0.892015762483528,
27
+ "eval_steps_per_second": 3.665,
28
+ "step": 200
29
+ },
30
+ {
31
+ "epoch": 2.2222222222222223,
32
+ "grad_norm": 8.369688034057617,
33
+ "learning_rate": 0.0005955104829638303,
34
+ "loss": 0.3485,
35
+ "step": 400
36
+ },
37
+ {
38
+ "epoch": 2.2222222222222223,
39
+ "eval_combined_score": 0.9047925495456136,
40
+ "eval_loss": 0.41716983914375305,
41
+ "eval_pearson": 0.9028346538543701,
42
+ "eval_runtime": 0.7414,
43
+ "eval_samples_per_second": 2023.3,
44
+ "eval_spearmanr": 0.9067504452368571,
45
+ "eval_steps_per_second": 4.047,
46
+ "step": 400
47
+ },
48
+ {
49
+ "epoch": 3.3333333333333335,
50
+ "grad_norm": 1.4970817565917969,
51
+ "learning_rate": 0.0004610272871538184,
52
+ "loss": 0.2625,
53
+ "step": 600
54
+ },
55
+ {
56
+ "epoch": 3.3333333333333335,
57
+ "eval_combined_score": 0.9118021224268875,
58
+ "eval_loss": 0.40821734070777893,
59
+ "eval_pearson": 0.9112259149551392,
60
+ "eval_runtime": 0.7423,
61
+ "eval_samples_per_second": 2020.801,
62
+ "eval_spearmanr": 0.912378329898636,
63
+ "eval_steps_per_second": 4.042,
64
+ "step": 600
65
+ },
66
+ {
67
+ "epoch": 4.444444444444445,
68
+ "grad_norm": 1.1825178861618042,
69
+ "learning_rate": 0.00030398717786598174,
70
+ "loss": 0.2131,
71
+ "step": 800
72
+ },
73
+ {
74
+ "epoch": 4.444444444444445,
75
+ "eval_combined_score": 0.91206781459837,
76
+ "eval_loss": 0.3940591514110565,
77
+ "eval_pearson": 0.9115008115768433,
78
+ "eval_runtime": 0.7425,
79
+ "eval_samples_per_second": 2020.27,
80
+ "eval_spearmanr": 0.912634817619897,
81
+ "eval_steps_per_second": 4.041,
82
+ "step": 800
83
+ },
84
+ {
85
+ "epoch": 5.555555555555555,
86
+ "grad_norm": 1.440523624420166,
87
+ "learning_rate": 0.00015629528624926277,
88
+ "loss": 0.1766,
89
+ "step": 1000
90
+ },
91
+ {
92
+ "epoch": 5.555555555555555,
93
+ "eval_combined_score": 0.9106852877757825,
94
+ "eval_loss": 0.4768920838832855,
95
+ "eval_pearson": 0.9102356433868408,
96
+ "eval_runtime": 0.7412,
97
+ "eval_samples_per_second": 2023.831,
98
+ "eval_spearmanr": 0.9111349321647242,
99
+ "eval_steps_per_second": 4.048,
100
+ "step": 1000
101
+ },
102
+ {
103
+ "epoch": 6.666666666666667,
104
+ "grad_norm": 1.652413249015808,
105
+ "learning_rate": 4.79575082052621e-05,
106
+ "loss": 0.1464,
107
+ "step": 1200
108
+ },
109
+ {
110
+ "epoch": 6.666666666666667,
111
+ "eval_combined_score": 0.9131120094564884,
112
+ "eval_loss": 0.42496126890182495,
113
+ "eval_pearson": 0.9132190942764282,
114
+ "eval_runtime": 0.7406,
115
+ "eval_samples_per_second": 2025.325,
116
+ "eval_spearmanr": 0.9130049246365486,
117
+ "eval_steps_per_second": 4.051,
118
+ "step": 1200
119
+ },
120
+ {
121
+ "epoch": 7.777777777777778,
122
+ "grad_norm": 1.2460709810256958,
123
+ "learning_rate": 9.843417688702115e-07,
124
+ "loss": 0.1413,
125
+ "step": 1400
126
+ },
127
+ {
128
+ "epoch": 7.777777777777778,
129
+ "eval_combined_score": 0.912296316293991,
130
+ "eval_loss": 0.41904354095458984,
131
+ "eval_pearson": 0.9121142029762268,
132
+ "eval_runtime": 0.7419,
133
+ "eval_samples_per_second": 2021.921,
134
+ "eval_spearmanr": 0.9124784296117551,
135
+ "eval_steps_per_second": 4.044,
136
+ "step": 1400
137
+ },
138
+ {
139
+ "epoch": 8.0,
140
+ "step": 1440,
141
+ "total_flos": 3072533598695424.0,
142
+ "train_loss": 0.38917061388492585,
143
+ "train_runtime": 332.8025,
144
+ "train_samples_per_second": 138.196,
145
+ "train_steps_per_second": 4.327
146
+ }
147
+ ],
148
+ "logging_steps": 200,
149
+ "max_steps": 1440,
150
+ "num_input_tokens_seen": 0,
151
+ "num_train_epochs": 8,
152
+ "save_steps": 200,
153
+ "stateful_callbacks": {
154
+ "TrainerControl": {
155
+ "args": {
156
+ "should_epoch_stop": false,
157
+ "should_evaluate": false,
158
+ "should_log": false,
159
+ "should_save": true,
160
+ "should_training_stop": true
161
+ },
162
+ "attributes": {}
163
+ }
164
+ },
165
+ "total_flos": 3072533598695424.0,
166
+ "train_batch_size": 32,
167
+ "trial_name": null,
168
+ "trial_params": null
169
+ }
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h35m28,sd43/STS-B.tsv ADDED
@@ -0,0 +1,1380 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ index prediction
2
+ 0 3.202
3
+ 1 4.144
4
+ 2 5.000
5
+ 3 5.000
6
+ 4 1.959
7
+ 5 1.845
8
+ 6 3.670
9
+ 7 1.581
10
+ 8 2.497
11
+ 9 1.769
12
+ 10 1.769
13
+ 11 5.000
14
+ 12 0.838
15
+ 13 4.977
16
+ 14 2.024
17
+ 15 1.717
18
+ 16 4.724
19
+ 17 2.973
20
+ 18 3.619
21
+ 19 1.198
22
+ 20 1.798
23
+ 21 0.458
24
+ 22 5.000
25
+ 23 5.000
26
+ 24 1.451
27
+ 25 2.835
28
+ 26 1.014
29
+ 27 2.163
30
+ 28 0.818
31
+ 29 2.697
32
+ 30 3.085
33
+ 31 4.681
34
+ 32 1.153
35
+ 33 4.725
36
+ 34 3.492
37
+ 35 0.362
38
+ 36 0.598
39
+ 37 2.192
40
+ 38 4.995
41
+ 39 0.428
42
+ 40 3.884
43
+ 41 2.799
44
+ 42 2.999
45
+ 43 1.001
46
+ 44 1.719
47
+ 45 0.431
48
+ 46 0.169
49
+ 47 0.712
50
+ 48 5.000
51
+ 49 1.935
52
+ 50 0.946
53
+ 51 1.437
54
+ 52 0.365
55
+ 53 0.907
56
+ 54 1.713
57
+ 55 0.208
58
+ 56 3.943
59
+ 57 0.823
60
+ 58 4.448
61
+ 59 0.764
62
+ 60 5.000
63
+ 61 0.000
64
+ 62 4.264
65
+ 63 1.105
66
+ 64 3.671
67
+ 65 0.009
68
+ 66 0.228
69
+ 67 4.191
70
+ 68 0.200
71
+ 69 4.405
72
+ 70 2.056
73
+ 71 4.271
74
+ 72 0.151
75
+ 73 4.184
76
+ 74 4.345
77
+ 75 0.140
78
+ 76 0.638
79
+ 77 0.000
80
+ 78 2.123
81
+ 79 4.250
82
+ 80 0.192
83
+ 81 0.678
84
+ 82 3.045
85
+ 83 0.506
86
+ 84 0.284
87
+ 85 1.558
88
+ 86 3.749
89
+ 87 4.820
90
+ 88 2.992
91
+ 89 1.623
92
+ 90 1.386
93
+ 91 3.596
94
+ 92 0.936
95
+ 93 3.942
96
+ 94 3.233
97
+ 95 4.853
98
+ 96 4.835
99
+ 97 3.439
100
+ 98 2.551
101
+ 99 3.352
102
+ 100 0.000
103
+ 101 0.592
104
+ 102 3.629
105
+ 103 2.851
106
+ 104 4.000
107
+ 105 2.798
108
+ 106 0.087
109
+ 107 2.887
110
+ 108 1.914
111
+ 109 3.732
112
+ 110 3.058
113
+ 111 4.201
114
+ 112 3.269
115
+ 113 4.138
116
+ 114 0.000
117
+ 115 0.026
118
+ 116 4.246
119
+ 117 3.002
120
+ 118 0.147
121
+ 119 0.128
122
+ 120 0.579
123
+ 121 0.344
124
+ 122 0.081
125
+ 123 0.179
126
+ 124 0.000
127
+ 125 4.898
128
+ 126 0.588
129
+ 127 2.281
130
+ 128 5.000
131
+ 129 3.479
132
+ 130 4.964
133
+ 131 2.294
134
+ 132 2.436
135
+ 133 5.000
136
+ 134 3.265
137
+ 135 5.000
138
+ 136 5.000
139
+ 137 1.787
140
+ 138 1.749
141
+ 139 4.930
142
+ 140 5.000
143
+ 141 3.700
144
+ 142 4.459
145
+ 143 3.995
146
+ 144 2.176
147
+ 145 1.417
148
+ 146 4.466
149
+ 147 2.924
150
+ 148 1.672
151
+ 149 0.806
152
+ 150 1.766
153
+ 151 1.830
154
+ 152 3.704
155
+ 153 3.352
156
+ 154 3.967
157
+ 155 3.646
158
+ 156 0.254
159
+ 157 0.959
160
+ 158 4.113
161
+ 159 0.886
162
+ 160 5.000
163
+ 161 2.269
164
+ 162 0.889
165
+ 163 3.231
166
+ 164 4.986
167
+ 165 4.550
168
+ 166 4.254
169
+ 167 3.550
170
+ 168 4.916
171
+ 169 4.405
172
+ 170 1.566
173
+ 171 0.375
174
+ 172 4.623
175
+ 173 5.000
176
+ 174 3.931
177
+ 175 3.373
178
+ 176 2.901
179
+ 177 4.158
180
+ 178 0.182
181
+ 179 0.163
182
+ 180 0.218
183
+ 181 2.781
184
+ 182 3.275
185
+ 183 1.106
186
+ 184 1.399
187
+ 185 0.185
188
+ 186 0.000
189
+ 187 0.291
190
+ 188 0.344
191
+ 189 0.102
192
+ 190 3.674
193
+ 191 4.516
194
+ 192 1.501
195
+ 193 2.890
196
+ 194 3.992
197
+ 195 2.931
198
+ 196 2.932
199
+ 197 0.176
200
+ 198 0.000
201
+ 199 3.131
202
+ 200 0.451
203
+ 201 0.800
204
+ 202 0.262
205
+ 203 0.397
206
+ 204 0.178
207
+ 205 0.525
208
+ 206 3.626
209
+ 207 0.084
210
+ 208 2.206
211
+ 209 0.541
212
+ 210 0.540
213
+ 211 0.212
214
+ 212 0.291
215
+ 213 0.573
216
+ 214 3.193
217
+ 215 0.600
218
+ 216 1.208
219
+ 217 3.474
220
+ 218 3.870
221
+ 219 2.864
222
+ 220 0.774
223
+ 221 0.125
224
+ 222 0.740
225
+ 223 1.180
226
+ 224 0.106
227
+ 225 0.085
228
+ 226 2.823
229
+ 227 2.057
230
+ 228 0.021
231
+ 229 1.683
232
+ 230 0.030
233
+ 231 0.002
234
+ 232 4.878
235
+ 233 2.601
236
+ 234 4.388
237
+ 235 2.642
238
+ 236 0.026
239
+ 237 0.000
240
+ 238 0.000
241
+ 239 0.022
242
+ 240 2.253
243
+ 241 0.000
244
+ 242 0.000
245
+ 243 0.080
246
+ 244 3.400
247
+ 245 0.123
248
+ 246 2.945
249
+ 247 3.397
250
+ 248 0.000
251
+ 249 0.477
252
+ 250 3.978
253
+ 251 4.718
254
+ 252 4.209
255
+ 253 2.511
256
+ 254 0.487
257
+ 255 3.843
258
+ 256 0.340
259
+ 257 4.275
260
+ 258 3.257
261
+ 259 0.033
262
+ 260 5.000
263
+ 261 3.297
264
+ 262 4.286
265
+ 263 4.113
266
+ 264 2.985
267
+ 265 1.393
268
+ 266 2.102
269
+ 267 4.027
270
+ 268 1.465
271
+ 269 5.000
272
+ 270 4.460
273
+ 271 0.258
274
+ 272 3.583
275
+ 273 5.000
276
+ 274 4.202
277
+ 275 5.000
278
+ 276 0.290
279
+ 277 0.000
280
+ 278 4.268
281
+ 279 2.967
282
+ 280 0.338
283
+ 281 2.355
284
+ 282 3.816
285
+ 283 4.072
286
+ 284 3.682
287
+ 285 4.463
288
+ 286 1.133
289
+ 287 3.827
290
+ 288 3.591
291
+ 289 0.960
292
+ 290 5.000
293
+ 291 0.810
294
+ 292 2.350
295
+ 293 4.906
296
+ 294 0.000
297
+ 295 0.438
298
+ 296 1.994
299
+ 297 3.801
300
+ 298 2.897
301
+ 299 2.859
302
+ 300 5.000
303
+ 301 2.480
304
+ 302 1.872
305
+ 303 4.915
306
+ 304 3.534
307
+ 305 4.092
308
+ 306 4.603
309
+ 307 3.893
310
+ 308 3.549
311
+ 309 4.552
312
+ 310 1.602
313
+ 311 0.012
314
+ 312 0.581
315
+ 313 2.880
316
+ 314 3.488
317
+ 315 4.620
318
+ 316 3.802
319
+ 317 3.489
320
+ 318 0.027
321
+ 319 0.113
322
+ 320 3.503
323
+ 321 2.980
324
+ 322 4.420
325
+ 323 2.539
326
+ 324 2.668
327
+ 325 0.175
328
+ 326 2.624
329
+ 327 3.574
330
+ 328 4.719
331
+ 329 3.677
332
+ 330 5.000
333
+ 331 4.808
334
+ 332 3.608
335
+ 333 2.163
336
+ 334 3.594
337
+ 335 1.611
338
+ 336 4.072
339
+ 337 3.983
340
+ 338 0.000
341
+ 339 0.316
342
+ 340 4.222
343
+ 341 4.846
344
+ 342 4.878
345
+ 343 2.108
346
+ 344 3.925
347
+ 345 0.461
348
+ 346 4.403
349
+ 347 0.175
350
+ 348 4.074
351
+ 349 0.107
352
+ 350 4.614
353
+ 351 3.306
354
+ 352 0.918
355
+ 353 4.366
356
+ 354 3.043
357
+ 355 0.158
358
+ 356 2.633
359
+ 357 4.933
360
+ 358 0.054
361
+ 359 4.012
362
+ 360 5.000
363
+ 361 4.560
364
+ 362 2.698
365
+ 363 3.654
366
+ 364 4.561
367
+ 365 0.047
368
+ 366 1.177
369
+ 367 4.170
370
+ 368 4.257
371
+ 369 4.894
372
+ 370 3.473
373
+ 371 4.519
374
+ 372 0.000
375
+ 373 0.459
376
+ 374 1.203
377
+ 375 4.501
378
+ 376 2.376
379
+ 377 0.000
380
+ 378 2.746
381
+ 379 1.621
382
+ 380 3.568
383
+ 381 4.239
384
+ 382 2.586
385
+ 383 0.366
386
+ 384 3.710
387
+ 385 2.273
388
+ 386 5.000
389
+ 387 1.232
390
+ 388 2.953
391
+ 389 2.216
392
+ 390 4.227
393
+ 391 2.074
394
+ 392 0.000
395
+ 393 2.096
396
+ 394 2.623
397
+ 395 0.095
398
+ 396 0.414
399
+ 397 5.000
400
+ 398 3.010
401
+ 399 2.659
402
+ 400 2.576
403
+ 401 0.165
404
+ 402 2.168
405
+ 403 4.208
406
+ 404 3.319
407
+ 405 3.908
408
+ 406 0.138
409
+ 407 3.268
410
+ 408 2.264
411
+ 409 3.246
412
+ 410 4.718
413
+ 411 5.000
414
+ 412 2.013
415
+ 413 4.778
416
+ 414 0.514
417
+ 415 1.212
418
+ 416 0.000
419
+ 417 5.000
420
+ 418 4.422
421
+ 419 3.955
422
+ 420 5.000
423
+ 421 1.708
424
+ 422 3.938
425
+ 423 1.448
426
+ 424 2.335
427
+ 425 3.253
428
+ 426 0.000
429
+ 427 5.000
430
+ 428 2.945
431
+ 429 5.000
432
+ 430 0.644
433
+ 431 5.000
434
+ 432 0.208
435
+ 433 3.728
436
+ 434 0.105
437
+ 435 2.511
438
+ 436 2.770
439
+ 437 5.000
440
+ 438 3.813
441
+ 439 4.674
442
+ 440 3.325
443
+ 441 2.718
444
+ 442 0.000
445
+ 443 2.762
446
+ 444 4.031
447
+ 445 0.000
448
+ 446 1.383
449
+ 447 4.042
450
+ 448 0.452
451
+ 449 1.583
452
+ 450 3.709
453
+ 451 3.049
454
+ 452 0.345
455
+ 453 0.201
456
+ 454 4.496
457
+ 455 2.210
458
+ 456 2.036
459
+ 457 5.000
460
+ 458 1.382
461
+ 459 2.988
462
+ 460 0.128
463
+ 461 0.535
464
+ 462 0.000
465
+ 463 1.967
466
+ 464 2.980
467
+ 465 4.213
468
+ 466 2.161
469
+ 467 4.953
470
+ 468 4.988
471
+ 469 0.299
472
+ 470 2.288
473
+ 471 4.719
474
+ 472 2.042
475
+ 473 4.995
476
+ 474 2.684
477
+ 475 2.772
478
+ 476 0.679
479
+ 477 3.182
480
+ 478 4.267
481
+ 479 0.000
482
+ 480 5.000
483
+ 481 2.650
484
+ 482 3.908
485
+ 483 0.000
486
+ 484 4.625
487
+ 485 0.223
488
+ 486 3.861
489
+ 487 4.719
490
+ 488 3.414
491
+ 489 4.986
492
+ 490 4.120
493
+ 491 4.195
494
+ 492 3.727
495
+ 493 3.729
496
+ 494 3.685
497
+ 495 2.378
498
+ 496 1.111
499
+ 497 5.000
500
+ 498 0.228
501
+ 499 5.000
502
+ 500 1.791
503
+ 501 2.319
504
+ 502 1.464
505
+ 503 2.315
506
+ 504 1.331
507
+ 505 3.242
508
+ 506 3.490
509
+ 507 0.771
510
+ 508 3.157
511
+ 509 0.000
512
+ 510 1.328
513
+ 511 1.262
514
+ 512 2.519
515
+ 513 0.317
516
+ 514 1.187
517
+ 515 0.124
518
+ 516 1.498
519
+ 517 2.178
520
+ 518 1.476
521
+ 519 3.640
522
+ 520 1.590
523
+ 521 2.388
524
+ 522 4.620
525
+ 523 0.914
526
+ 524 2.901
527
+ 525 3.746
528
+ 526 1.746
529
+ 527 0.839
530
+ 528 2.355
531
+ 529 0.762
532
+ 530 0.747
533
+ 531 0.000
534
+ 532 1.320
535
+ 533 0.475
536
+ 534 0.151
537
+ 535 0.399
538
+ 536 2.023
539
+ 537 1.164
540
+ 538 2.050
541
+ 539 2.999
542
+ 540 2.454
543
+ 541 1.033
544
+ 542 1.919
545
+ 543 3.060
546
+ 544 4.457
547
+ 545 5.000
548
+ 546 1.155
549
+ 547 2.378
550
+ 548 3.052
551
+ 549 4.463
552
+ 550 3.962
553
+ 551 1.018
554
+ 552 2.911
555
+ 553 2.502
556
+ 554 1.693
557
+ 555 2.416
558
+ 556 2.770
559
+ 557 1.097
560
+ 558 5.000
561
+ 559 3.956
562
+ 560 4.005
563
+ 561 0.015
564
+ 562 1.258
565
+ 563 0.870
566
+ 564 0.826
567
+ 565 2.580
568
+ 566 5.000
569
+ 567 2.593
570
+ 568 0.390
571
+ 569 1.909
572
+ 570 1.924
573
+ 571 0.943
574
+ 572 0.307
575
+ 573 3.425
576
+ 574 4.674
577
+ 575 2.370
578
+ 576 4.839
579
+ 577 3.338
580
+ 578 2.060
581
+ 579 1.387
582
+ 580 1.299
583
+ 581 3.085
584
+ 582 2.844
585
+ 583 1.373
586
+ 584 2.752
587
+ 585 4.090
588
+ 586 0.604
589
+ 587 5.000
590
+ 588 1.487
591
+ 589 1.609
592
+ 590 1.558
593
+ 591 2.011
594
+ 592 4.815
595
+ 593 1.617
596
+ 594 0.501
597
+ 595 3.516
598
+ 596 2.053
599
+ 597 0.443
600
+ 598 3.991
601
+ 599 4.798
602
+ 600 2.281
603
+ 601 0.280
604
+ 602 3.719
605
+ 603 1.612
606
+ 604 1.045
607
+ 605 4.075
608
+ 606 1.876
609
+ 607 1.293
610
+ 608 3.141
611
+ 609 2.631
612
+ 610 5.000
613
+ 611 2.969
614
+ 612 1.258
615
+ 613 0.719
616
+ 614 3.124
617
+ 615 1.953
618
+ 616 1.611
619
+ 617 0.306
620
+ 618 2.464
621
+ 619 4.318
622
+ 620 0.796
623
+ 621 2.649
624
+ 622 0.818
625
+ 623 5.000
626
+ 624 2.314
627
+ 625 3.144
628
+ 626 2.757
629
+ 627 0.318
630
+ 628 0.648
631
+ 629 2.430
632
+ 630 1.606
633
+ 631 5.000
634
+ 632 3.197
635
+ 633 2.900
636
+ 634 1.474
637
+ 635 2.688
638
+ 636 4.128
639
+ 637 2.395
640
+ 638 3.025
641
+ 639 3.134
642
+ 640 1.175
643
+ 641 1.410
644
+ 642 2.440
645
+ 643 2.623
646
+ 644 1.770
647
+ 645 4.708
648
+ 646 3.428
649
+ 647 4.184
650
+ 648 4.698
651
+ 649 3.508
652
+ 650 1.928
653
+ 651 4.286
654
+ 652 0.862
655
+ 653 1.202
656
+ 654 1.611
657
+ 655 3.266
658
+ 656 4.842
659
+ 657 1.426
660
+ 658 1.916
661
+ 659 0.885
662
+ 660 1.562
663
+ 661 1.483
664
+ 662 0.402
665
+ 663 2.303
666
+ 664 4.878
667
+ 665 1.089
668
+ 666 3.694
669
+ 667 3.016
670
+ 668 3.204
671
+ 669 2.716
672
+ 670 2.139
673
+ 671 4.014
674
+ 672 1.994
675
+ 673 4.877
676
+ 674 4.962
677
+ 675 0.685
678
+ 676 1.697
679
+ 677 3.060
680
+ 678 1.725
681
+ 679 2.849
682
+ 680 3.363
683
+ 681 1.287
684
+ 682 3.134
685
+ 683 2.570
686
+ 684 4.070
687
+ 685 2.064
688
+ 686 4.284
689
+ 687 1.999
690
+ 688 3.935
691
+ 689 4.656
692
+ 690 2.363
693
+ 691 2.359
694
+ 692 4.839
695
+ 693 1.244
696
+ 694 3.522
697
+ 695 1.288
698
+ 696 2.433
699
+ 697 0.561
700
+ 698 1.482
701
+ 699 4.319
702
+ 700 4.035
703
+ 701 2.049
704
+ 702 4.730
705
+ 703 4.671
706
+ 704 2.816
707
+ 705 1.723
708
+ 706 2.249
709
+ 707 0.995
710
+ 708 3.457
711
+ 709 1.969
712
+ 710 4.540
713
+ 711 1.202
714
+ 712 2.095
715
+ 713 1.459
716
+ 714 2.053
717
+ 715 1.955
718
+ 716 1.949
719
+ 717 4.346
720
+ 718 2.395
721
+ 719 5.000
722
+ 720 5.000
723
+ 721 2.783
724
+ 722 2.155
725
+ 723 1.288
726
+ 724 4.410
727
+ 725 0.836
728
+ 726 3.400
729
+ 727 4.427
730
+ 728 3.239
731
+ 729 3.312
732
+ 730 3.270
733
+ 731 1.734
734
+ 732 2.314
735
+ 733 0.294
736
+ 734 2.396
737
+ 735 1.578
738
+ 736 1.770
739
+ 737 4.688
740
+ 738 2.400
741
+ 739 2.734
742
+ 740 4.592
743
+ 741 1.875
744
+ 742 1.393
745
+ 743 3.468
746
+ 744 1.735
747
+ 745 3.456
748
+ 746 0.798
749
+ 747 1.523
750
+ 748 3.862
751
+ 749 1.840
752
+ 750 2.471
753
+ 751 1.687
754
+ 752 2.933
755
+ 753 1.353
756
+ 754 1.361
757
+ 755 4.871
758
+ 756 4.769
759
+ 757 3.652
760
+ 758 1.799
761
+ 759 3.610
762
+ 760 3.704
763
+ 761 0.247
764
+ 762 0.485
765
+ 763 1.829
766
+ 764 3.545
767
+ 765 4.330
768
+ 766 3.253
769
+ 767 2.090
770
+ 768 4.002
771
+ 769 1.617
772
+ 770 2.610
773
+ 771 1.455
774
+ 772 1.443
775
+ 773 4.817
776
+ 774 4.824
777
+ 775 3.843
778
+ 776 1.672
779
+ 777 0.793
780
+ 778 2.361
781
+ 779 0.753
782
+ 780 2.306
783
+ 781 2.417
784
+ 782 5.000
785
+ 783 4.561
786
+ 784 3.464
787
+ 785 0.809
788
+ 786 0.859
789
+ 787 2.778
790
+ 788 5.000
791
+ 789 2.405
792
+ 790 2.508
793
+ 791 2.653
794
+ 792 1.856
795
+ 793 1.619
796
+ 794 2.455
797
+ 795 3.307
798
+ 796 1.770
799
+ 797 4.225
800
+ 798 0.810
801
+ 799 1.335
802
+ 800 1.671
803
+ 801 2.571
804
+ 802 1.485
805
+ 803 0.590
806
+ 804 1.641
807
+ 805 5.000
808
+ 806 2.628
809
+ 807 2.847
810
+ 808 4.021
811
+ 809 2.049
812
+ 810 3.940
813
+ 811 2.072
814
+ 812 1.801
815
+ 813 4.462
816
+ 814 4.580
817
+ 815 1.715
818
+ 816 2.960
819
+ 817 1.659
820
+ 818 3.883
821
+ 819 2.330
822
+ 820 2.156
823
+ 821 2.439
824
+ 822 1.130
825
+ 823 4.295
826
+ 824 2.592
827
+ 825 1.462
828
+ 826 2.907
829
+ 827 0.980
830
+ 828 3.711
831
+ 829 5.000
832
+ 830 2.270
833
+ 831 2.062
834
+ 832 1.272
835
+ 833 4.902
836
+ 834 1.758
837
+ 835 3.491
838
+ 836 4.359
839
+ 837 1.579
840
+ 838 1.617
841
+ 839 4.072
842
+ 840 3.265
843
+ 841 1.706
844
+ 842 4.598
845
+ 843 4.073
846
+ 844 4.352
847
+ 845 4.195
848
+ 846 2.513
849
+ 847 2.117
850
+ 848 1.765
851
+ 849 2.886
852
+ 850 4.766
853
+ 851 1.425
854
+ 852 4.135
855
+ 853 4.835
856
+ 854 5.000
857
+ 855 1.829
858
+ 856 0.784
859
+ 857 2.891
860
+ 858 1.767
861
+ 859 3.168
862
+ 860 4.141
863
+ 861 4.538
864
+ 862 3.577
865
+ 863 0.185
866
+ 864 0.792
867
+ 865 0.954
868
+ 866 1.255
869
+ 867 0.715
870
+ 868 0.727
871
+ 869 4.814
872
+ 870 3.798
873
+ 871 2.252
874
+ 872 2.660
875
+ 873 4.238
876
+ 874 2.085
877
+ 875 3.945
878
+ 876 2.288
879
+ 877 4.551
880
+ 878 4.346
881
+ 879 1.822
882
+ 880 4.101
883
+ 881 3.448
884
+ 882 3.116
885
+ 883 4.431
886
+ 884 1.596
887
+ 885 3.960
888
+ 886 3.262
889
+ 887 4.200
890
+ 888 3.116
891
+ 889 3.082
892
+ 890 5.000
893
+ 891 3.159
894
+ 892 1.547
895
+ 893 3.769
896
+ 894 4.181
897
+ 895 4.371
898
+ 896 3.321
899
+ 897 3.237
900
+ 898 1.258
901
+ 899 3.823
902
+ 900 3.909
903
+ 901 4.081
904
+ 902 3.609
905
+ 903 4.460
906
+ 904 4.756
907
+ 905 2.918
908
+ 906 3.288
909
+ 907 4.175
910
+ 908 3.540
911
+ 909 2.879
912
+ 910 3.780
913
+ 911 2.016
914
+ 912 2.616
915
+ 913 4.261
916
+ 914 3.537
917
+ 915 4.809
918
+ 916 3.934
919
+ 917 2.942
920
+ 918 4.267
921
+ 919 3.893
922
+ 920 3.555
923
+ 921 3.891
924
+ 922 4.971
925
+ 923 4.347
926
+ 924 4.187
927
+ 925 2.260
928
+ 926 3.201
929
+ 927 2.412
930
+ 928 3.947
931
+ 929 3.319
932
+ 930 3.771
933
+ 931 3.567
934
+ 932 4.414
935
+ 933 3.082
936
+ 934 1.927
937
+ 935 2.055
938
+ 936 3.330
939
+ 937 3.331
940
+ 938 3.244
941
+ 939 2.762
942
+ 940 4.303
943
+ 941 3.495
944
+ 942 3.329
945
+ 943 3.146
946
+ 944 3.700
947
+ 945 4.371
948
+ 946 4.057
949
+ 947 4.216
950
+ 948 3.726
951
+ 949 2.141
952
+ 950 3.262
953
+ 951 4.691
954
+ 952 4.279
955
+ 953 3.149
956
+ 954 3.000
957
+ 955 1.823
958
+ 956 3.406
959
+ 957 1.220
960
+ 958 4.292
961
+ 959 3.007
962
+ 960 3.355
963
+ 961 3.858
964
+ 962 4.369
965
+ 963 2.848
966
+ 964 3.668
967
+ 965 3.641
968
+ 966 4.068
969
+ 967 4.272
970
+ 968 3.514
971
+ 969 3.268
972
+ 970 2.696
973
+ 971 3.188
974
+ 972 3.980
975
+ 973 3.129
976
+ 974 3.705
977
+ 975 3.015
978
+ 976 3.165
979
+ 977 4.913
980
+ 978 4.354
981
+ 979 3.805
982
+ 980 2.369
983
+ 981 4.734
984
+ 982 2.872
985
+ 983 2.636
986
+ 984 2.569
987
+ 985 3.144
988
+ 986 4.480
989
+ 987 3.260
990
+ 988 4.143
991
+ 989 4.095
992
+ 990 4.501
993
+ 991 1.789
994
+ 992 3.539
995
+ 993 3.529
996
+ 994 2.558
997
+ 995 2.590
998
+ 996 1.669
999
+ 997 4.206
1000
+ 998 4.293
1001
+ 999 3.008
1002
+ 1000 3.369
1003
+ 1001 3.988
1004
+ 1002 4.010
1005
+ 1003 3.394
1006
+ 1004 1.577
1007
+ 1005 2.684
1008
+ 1006 4.527
1009
+ 1007 3.318
1010
+ 1008 1.853
1011
+ 1009 3.645
1012
+ 1010 4.305
1013
+ 1011 4.128
1014
+ 1012 4.864
1015
+ 1013 2.320
1016
+ 1014 4.256
1017
+ 1015 3.834
1018
+ 1016 4.336
1019
+ 1017 3.178
1020
+ 1018 4.856
1021
+ 1019 3.864
1022
+ 1020 3.315
1023
+ 1021 4.432
1024
+ 1022 4.094
1025
+ 1023 1.866
1026
+ 1024 3.432
1027
+ 1025 3.577
1028
+ 1026 3.285
1029
+ 1027 1.977
1030
+ 1028 3.922
1031
+ 1029 1.902
1032
+ 1030 4.561
1033
+ 1031 4.264
1034
+ 1032 4.219
1035
+ 1033 3.178
1036
+ 1034 5.000
1037
+ 1035 3.529
1038
+ 1036 3.620
1039
+ 1037 4.128
1040
+ 1038 4.545
1041
+ 1039 4.936
1042
+ 1040 3.182
1043
+ 1041 4.308
1044
+ 1042 3.341
1045
+ 1043 3.597
1046
+ 1044 4.111
1047
+ 1045 2.516
1048
+ 1046 1.666
1049
+ 1047 3.019
1050
+ 1048 4.339
1051
+ 1049 4.781
1052
+ 1050 3.576
1053
+ 1051 3.736
1054
+ 1052 4.255
1055
+ 1053 3.057
1056
+ 1054 2.956
1057
+ 1055 4.552
1058
+ 1056 4.218
1059
+ 1057 3.127
1060
+ 1058 4.716
1061
+ 1059 3.541
1062
+ 1060 3.096
1063
+ 1061 3.355
1064
+ 1062 2.820
1065
+ 1063 4.471
1066
+ 1064 2.921
1067
+ 1065 3.101
1068
+ 1066 4.073
1069
+ 1067 2.789
1070
+ 1068 3.011
1071
+ 1069 3.746
1072
+ 1070 5.000
1073
+ 1071 2.657
1074
+ 1072 3.195
1075
+ 1073 0.842
1076
+ 1074 3.202
1077
+ 1075 2.264
1078
+ 1076 3.784
1079
+ 1077 3.893
1080
+ 1078 4.018
1081
+ 1079 2.028
1082
+ 1080 4.034
1083
+ 1081 3.493
1084
+ 1082 3.921
1085
+ 1083 0.974
1086
+ 1084 3.719
1087
+ 1085 3.924
1088
+ 1086 2.496
1089
+ 1087 3.976
1090
+ 1088 1.535
1091
+ 1089 4.048
1092
+ 1090 3.405
1093
+ 1091 4.851
1094
+ 1092 3.686
1095
+ 1093 3.867
1096
+ 1094 3.376
1097
+ 1095 3.667
1098
+ 1096 3.514
1099
+ 1097 3.906
1100
+ 1098 3.113
1101
+ 1099 2.082
1102
+ 1100 2.732
1103
+ 1101 3.844
1104
+ 1102 3.440
1105
+ 1103 3.589
1106
+ 1104 3.765
1107
+ 1105 3.651
1108
+ 1106 3.183
1109
+ 1107 1.810
1110
+ 1108 3.551
1111
+ 1109 2.665
1112
+ 1110 1.139
1113
+ 1111 4.766
1114
+ 1112 3.499
1115
+ 1113 4.803
1116
+ 1114 3.898
1117
+ 1115 3.675
1118
+ 1116 2.772
1119
+ 1117 3.980
1120
+ 1118 4.737
1121
+ 1119 3.492
1122
+ 1120 3.012
1123
+ 1121 0.883
1124
+ 1122 4.228
1125
+ 1123 3.310
1126
+ 1124 3.385
1127
+ 1125 4.663
1128
+ 1126 3.628
1129
+ 1127 4.499
1130
+ 1128 4.312
1131
+ 1129 1.718
1132
+ 1130 1.273
1133
+ 1131 3.997
1134
+ 1132 4.915
1135
+ 1133 0.655
1136
+ 1134 2.740
1137
+ 1135 2.460
1138
+ 1136 2.492
1139
+ 1137 3.275
1140
+ 1138 0.656
1141
+ 1139 4.745
1142
+ 1140 1.386
1143
+ 1141 0.718
1144
+ 1142 2.326
1145
+ 1143 4.267
1146
+ 1144 3.690
1147
+ 1145 2.109
1148
+ 1146 1.136
1149
+ 1147 4.778
1150
+ 1148 4.184
1151
+ 1149 4.882
1152
+ 1150 0.000
1153
+ 1151 0.157
1154
+ 1152 5.000
1155
+ 1153 4.951
1156
+ 1154 5.000
1157
+ 1155 3.974
1158
+ 1156 4.241
1159
+ 1157 0.056
1160
+ 1158 4.716
1161
+ 1159 1.325
1162
+ 1160 4.006
1163
+ 1161 3.517
1164
+ 1162 0.796
1165
+ 1163 4.888
1166
+ 1164 4.816
1167
+ 1165 0.557
1168
+ 1166 3.960
1169
+ 1167 1.547
1170
+ 1168 3.355
1171
+ 1169 0.272
1172
+ 1170 2.415
1173
+ 1171 4.512
1174
+ 1172 2.970
1175
+ 1173 3.818
1176
+ 1174 0.138
1177
+ 1175 0.003
1178
+ 1176 5.000
1179
+ 1177 1.586
1180
+ 1178 3.136
1181
+ 1179 0.428
1182
+ 1180 0.799
1183
+ 1181 3.925
1184
+ 1182 1.691
1185
+ 1183 2.637
1186
+ 1184 0.467
1187
+ 1185 4.377
1188
+ 1186 5.000
1189
+ 1187 3.936
1190
+ 1188 3.268
1191
+ 1189 4.734
1192
+ 1190 3.653
1193
+ 1191 4.418
1194
+ 1192 1.374
1195
+ 1193 1.980
1196
+ 1194 1.051
1197
+ 1195 1.829
1198
+ 1196 4.766
1199
+ 1197 4.557
1200
+ 1198 3.601
1201
+ 1199 3.728
1202
+ 1200 2.795
1203
+ 1201 0.963
1204
+ 1202 3.313
1205
+ 1203 4.821
1206
+ 1204 4.183
1207
+ 1205 2.956
1208
+ 1206 5.000
1209
+ 1207 4.136
1210
+ 1208 2.002
1211
+ 1209 2.206
1212
+ 1210 2.497
1213
+ 1211 4.160
1214
+ 1212 4.268
1215
+ 1213 2.470
1216
+ 1214 5.000
1217
+ 1215 4.455
1218
+ 1216 1.177
1219
+ 1217 2.572
1220
+ 1218 1.456
1221
+ 1219 1.813
1222
+ 1220 1.456
1223
+ 1221 3.096
1224
+ 1222 4.208
1225
+ 1223 3.155
1226
+ 1224 1.448
1227
+ 1225 3.279
1228
+ 1226 3.107
1229
+ 1227 4.923
1230
+ 1228 4.632
1231
+ 1229 3.076
1232
+ 1230 2.664
1233
+ 1231 3.278
1234
+ 1232 3.605
1235
+ 1233 4.888
1236
+ 1234 4.243
1237
+ 1235 2.304
1238
+ 1236 2.392
1239
+ 1237 0.871
1240
+ 1238 1.792
1241
+ 1239 1.130
1242
+ 1240 1.007
1243
+ 1241 1.962
1244
+ 1242 2.486
1245
+ 1243 2.748
1246
+ 1244 0.000
1247
+ 1245 5.000
1248
+ 1246 4.196
1249
+ 1247 1.075
1250
+ 1248 0.146
1251
+ 1249 4.841
1252
+ 1250 4.497
1253
+ 1251 3.887
1254
+ 1252 4.423
1255
+ 1253 2.524
1256
+ 1254 0.000
1257
+ 1255 4.701
1258
+ 1256 2.038
1259
+ 1257 4.048
1260
+ 1258 3.932
1261
+ 1259 4.680
1262
+ 1260 3.472
1263
+ 1261 0.945
1264
+ 1262 4.036
1265
+ 1263 1.526
1266
+ 1264 3.575
1267
+ 1265 3.996
1268
+ 1266 4.136
1269
+ 1267 4.791
1270
+ 1268 1.291
1271
+ 1269 1.069
1272
+ 1270 3.075
1273
+ 1271 3.533
1274
+ 1272 1.103
1275
+ 1273 3.990
1276
+ 1274 4.361
1277
+ 1275 2.267
1278
+ 1276 0.533
1279
+ 1277 3.630
1280
+ 1278 0.835
1281
+ 1279 4.658
1282
+ 1280 0.260
1283
+ 1281 0.175
1284
+ 1282 3.279
1285
+ 1283 4.890
1286
+ 1284 1.583
1287
+ 1285 4.407
1288
+ 1286 4.382
1289
+ 1287 1.999
1290
+ 1288 0.000
1291
+ 1289 2.085
1292
+ 1290 5.000
1293
+ 1291 0.554
1294
+ 1292 4.175
1295
+ 1293 3.415
1296
+ 1294 1.710
1297
+ 1295 0.409
1298
+ 1296 5.000
1299
+ 1297 1.421
1300
+ 1298 0.760
1301
+ 1299 4.656
1302
+ 1300 2.741
1303
+ 1301 2.086
1304
+ 1302 2.610
1305
+ 1303 2.093
1306
+ 1304 1.969
1307
+ 1305 0.301
1308
+ 1306 1.994
1309
+ 1307 0.686
1310
+ 1308 2.395
1311
+ 1309 3.122
1312
+ 1310 3.386
1313
+ 1311 4.859
1314
+ 1312 0.709
1315
+ 1313 3.252
1316
+ 1314 2.718
1317
+ 1315 4.765
1318
+ 1316 4.744
1319
+ 1317 5.000
1320
+ 1318 2.068
1321
+ 1319 0.039
1322
+ 1320 5.000
1323
+ 1321 1.591
1324
+ 1322 5.000
1325
+ 1323 2.722
1326
+ 1324 4.965
1327
+ 1325 4.794
1328
+ 1326 5.000
1329
+ 1327 3.437
1330
+ 1328 0.536
1331
+ 1329 4.824
1332
+ 1330 4.556
1333
+ 1331 2.562
1334
+ 1332 1.575
1335
+ 1333 1.711
1336
+ 1334 2.474
1337
+ 1335 4.635
1338
+ 1336 4.481
1339
+ 1337 4.560
1340
+ 1338 0.435
1341
+ 1339 1.889
1342
+ 1340 3.705
1343
+ 1341 4.432
1344
+ 1342 2.744
1345
+ 1343 0.391
1346
+ 1344 0.681
1347
+ 1345 2.304
1348
+ 1346 4.509
1349
+ 1347 5.000
1350
+ 1348 5.000
1351
+ 1349 2.265
1352
+ 1350 5.000
1353
+ 1351 5.000
1354
+ 1352 5.000
1355
+ 1353 5.000
1356
+ 1354 4.471
1357
+ 1355 2.511
1358
+ 1356 2.164
1359
+ 1357 5.000
1360
+ 1358 3.751
1361
+ 1359 2.908
1362
+ 1360 0.684
1363
+ 1361 3.919
1364
+ 1362 1.322
1365
+ 1363 2.104
1366
+ 1364 5.000
1367
+ 1365 1.976
1368
+ 1366 1.596
1369
+ 1367 2.349
1370
+ 1368 0.236
1371
+ 1369 0.443
1372
+ 1370 0.970
1373
+ 1371 1.224
1374
+ 1372 1.615
1375
+ 1373 1.946
1376
+ 1374 0.466
1377
+ 1375 1.084
1378
+ 1376 1.709
1379
+ 1377 0.055
1380
+ 1378 1.001
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h35m28,sd43/all_results.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 8.0,
3
+ "eval_combined_score": 0.9120661209326952,
4
+ "eval_loss": 0.4218188524246216,
5
+ "eval_pearson": 0.9118423461914062,
6
+ "eval_runtime": 1.0156,
7
+ "eval_samples": 1500,
8
+ "eval_samples_per_second": 1476.987,
9
+ "eval_spearmanr": 0.9122898956739842,
10
+ "eval_steps_per_second": 2.954
11
+ }
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h35m28,sd43/eval_results.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 8.0,
3
+ "eval_combined_score": 0.9120661209326952,
4
+ "eval_loss": 0.4218188524246216,
5
+ "eval_pearson": 0.9118423461914062,
6
+ "eval_runtime": 1.0156,
7
+ "eval_samples": 1500,
8
+ "eval_samples_per_second": 1476.987,
9
+ "eval_spearmanr": 0.9122898956739842,
10
+ "eval_steps_per_second": 2.954
11
+ }
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h35m28,sd43/ft/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[MASK]": 128000
3
+ }
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h35m28,sd43/ft/special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": {
9
+ "content": "[UNK]",
10
+ "lstrip": false,
11
+ "normalized": true,
12
+ "rstrip": false,
13
+ "single_word": false
14
+ }
15
+ }
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h35m28,sd43/ft/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h35m28,sd43/ft/tokenizer_config.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[CLS]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[SEP]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[UNK]",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "128000": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "[CLS]",
45
+ "clean_up_tokenization_spaces": false,
46
+ "cls_token": "[CLS]",
47
+ "do_lower_case": false,
48
+ "eos_token": "[SEP]",
49
+ "extra_special_tokens": {},
50
+ "mask_token": "[MASK]",
51
+ "model_max_length": 512,
52
+ "pad_token": "[PAD]",
53
+ "padding_side": "right",
54
+ "sep_token": "[SEP]",
55
+ "sp_model_kwargs": {},
56
+ "split_by_punct": false,
57
+ "tokenizer_class": "DebertaV2Tokenizer",
58
+ "unk_token": "[UNK]",
59
+ "vocab_type": "spm"
60
+ }
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h35m28,sd43/ft2/README.md ADDED
@@ -0,0 +1,205 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: microsoft/deberta-v3-base
3
+ library_name: peft
4
+ tags:
5
+ - base_model:adapter:microsoft/deberta-v3-base
6
+ - transformers
7
+ ---
8
+
9
+ # Model Card for Model ID
10
+
11
+ <!-- Provide a quick summary of what the model is/does. -->
12
+
13
+
14
+
15
+ ## Model Details
16
+
17
+ ### Model Description
18
+
19
+ <!-- Provide a longer summary of what this model is. -->
20
+
21
+
22
+
23
+ - **Developed by:** [More Information Needed]
24
+ - **Funded by [optional]:** [More Information Needed]
25
+ - **Shared by [optional]:** [More Information Needed]
26
+ - **Model type:** [More Information Needed]
27
+ - **Language(s) (NLP):** [More Information Needed]
28
+ - **License:** [More Information Needed]
29
+ - **Finetuned from model [optional]:** [More Information Needed]
30
+
31
+ ### Model Sources [optional]
32
+
33
+ <!-- Provide the basic links for the model. -->
34
+
35
+ - **Repository:** [More Information Needed]
36
+ - **Paper [optional]:** [More Information Needed]
37
+ - **Demo [optional]:** [More Information Needed]
38
+
39
+ ## Uses
40
+
41
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
42
+
43
+ ### Direct Use
44
+
45
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
46
+
47
+ [More Information Needed]
48
+
49
+ ### Downstream Use [optional]
50
+
51
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
52
+
53
+ [More Information Needed]
54
+
55
+ ### Out-of-Scope Use
56
+
57
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
58
+
59
+ [More Information Needed]
60
+
61
+ ## Bias, Risks, and Limitations
62
+
63
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
64
+
65
+ [More Information Needed]
66
+
67
+ ### Recommendations
68
+
69
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
70
+
71
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
72
+
73
+ ## How to Get Started with the Model
74
+
75
+ Use the code below to get started with the model.
76
+
77
+ [More Information Needed]
78
+
79
+ ## Training Details
80
+
81
+ ### Training Data
82
+
83
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
84
+
85
+ [More Information Needed]
86
+
87
+ ### Training Procedure
88
+
89
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
90
+
91
+ #### Preprocessing [optional]
92
+
93
+ [More Information Needed]
94
+
95
+
96
+ #### Training Hyperparameters
97
+
98
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
99
+
100
+ #### Speeds, Sizes, Times [optional]
101
+
102
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
103
+
104
+ [More Information Needed]
105
+
106
+ ## Evaluation
107
+
108
+ <!-- This section describes the evaluation protocols and provides the results. -->
109
+
110
+ ### Testing Data, Factors & Metrics
111
+
112
+ #### Testing Data
113
+
114
+ <!-- This should link to a Dataset Card if possible. -->
115
+
116
+ [More Information Needed]
117
+
118
+ #### Factors
119
+
120
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
121
+
122
+ [More Information Needed]
123
+
124
+ #### Metrics
125
+
126
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
127
+
128
+ [More Information Needed]
129
+
130
+ ### Results
131
+
132
+ [More Information Needed]
133
+
134
+ #### Summary
135
+
136
+
137
+
138
+ ## Model Examination [optional]
139
+
140
+ <!-- Relevant interpretability work for the model goes here -->
141
+
142
+ [More Information Needed]
143
+
144
+ ## Environmental Impact
145
+
146
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
147
+
148
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
149
+
150
+ - **Hardware Type:** [More Information Needed]
151
+ - **Hours used:** [More Information Needed]
152
+ - **Cloud Provider:** [More Information Needed]
153
+ - **Compute Region:** [More Information Needed]
154
+ - **Carbon Emitted:** [More Information Needed]
155
+
156
+ ## Technical Specifications [optional]
157
+
158
+ ### Model Architecture and Objective
159
+
160
+ [More Information Needed]
161
+
162
+ ### Compute Infrastructure
163
+
164
+ [More Information Needed]
165
+
166
+ #### Hardware
167
+
168
+ [More Information Needed]
169
+
170
+ #### Software
171
+
172
+ [More Information Needed]
173
+
174
+ ## Citation [optional]
175
+
176
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
177
+
178
+ **BibTeX:**
179
+
180
+ [More Information Needed]
181
+
182
+ **APA:**
183
+
184
+ [More Information Needed]
185
+
186
+ ## Glossary [optional]
187
+
188
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
189
+
190
+ [More Information Needed]
191
+
192
+ ## More Information [optional]
193
+
194
+ [More Information Needed]
195
+
196
+ ## Model Card Authors [optional]
197
+
198
+ [More Information Needed]
199
+
200
+ ## Model Card Contact
201
+
202
+ [More Information Needed]
203
+ ### Framework versions
204
+
205
+ - PEFT 0.18.0
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h35m28,sd43/ft2/adapter_config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": {
3
+ "base_model_class": "DebertaV2ForSequenceClassification",
4
+ "parent_library": "transformers.models.deberta_v2.modeling_deberta_v2"
5
+ },
6
+ "base_model_name_or_path": "microsoft/deberta-v3-base",
7
+ "bias": "none",
8
+ "boft_block_num": 0,
9
+ "boft_block_size": 4,
10
+ "boft_dropout": 0.05,
11
+ "boft_n_butterfly_factor": 2,
12
+ "exclude_modules": null,
13
+ "fan_in_fan_out": false,
14
+ "inference_mode": true,
15
+ "init_weights": true,
16
+ "layers_pattern": null,
17
+ "layers_to_transform": null,
18
+ "modules_to_save": [
19
+ "classifier",
20
+ "pooler"
21
+ ],
22
+ "peft_type": "BOFT",
23
+ "peft_version": "0.18.0",
24
+ "revision": null,
25
+ "target_modules": [
26
+ "query_proj",
27
+ "value_proj",
28
+ "output.dense",
29
+ "key_proj",
30
+ "intermediate.dense",
31
+ "attention.output.dense"
32
+ ],
33
+ "task_type": null
34
+ }
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h35m28,sd43/trainer_state.json ADDED
@@ -0,0 +1,169 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 1200,
3
+ "best_metric": 0.9120661209326952,
4
+ "best_model_checkpoint": "./glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h35m28,sd43/checkpoint-1200",
5
+ "epoch": 8.0,
6
+ "eval_steps": 200,
7
+ "global_step": 1440,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 1.1111111111111112,
14
+ "grad_norm": 1.7553361654281616,
15
+ "learning_rate": 0.0006801144205835404,
16
+ "loss": 1.3664,
17
+ "step": 200
18
+ },
19
+ {
20
+ "epoch": 1.1111111111111112,
21
+ "eval_combined_score": 0.8862456234230243,
22
+ "eval_loss": 0.5907981991767883,
23
+ "eval_pearson": 0.8840484619140625,
24
+ "eval_runtime": 1.0079,
25
+ "eval_samples_per_second": 1488.191,
26
+ "eval_spearmanr": 0.8884427849319861,
27
+ "eval_steps_per_second": 2.976,
28
+ "step": 200
29
+ },
30
+ {
31
+ "epoch": 2.2222222222222223,
32
+ "grad_norm": 2.1135005950927734,
33
+ "learning_rate": 0.0005955104829638303,
34
+ "loss": 0.3475,
35
+ "step": 400
36
+ },
37
+ {
38
+ "epoch": 2.2222222222222223,
39
+ "eval_combined_score": 0.9062500854408247,
40
+ "eval_loss": 0.45917749404907227,
41
+ "eval_pearson": 0.9043642282485962,
42
+ "eval_runtime": 0.8397,
43
+ "eval_samples_per_second": 1786.311,
44
+ "eval_spearmanr": 0.9081359426330533,
45
+ "eval_steps_per_second": 3.573,
46
+ "step": 400
47
+ },
48
+ {
49
+ "epoch": 3.3333333333333335,
50
+ "grad_norm": 6.376646995544434,
51
+ "learning_rate": 0.0004610272871538184,
52
+ "loss": 0.2666,
53
+ "step": 600
54
+ },
55
+ {
56
+ "epoch": 3.3333333333333335,
57
+ "eval_combined_score": 0.9112185982157557,
58
+ "eval_loss": 0.503534197807312,
59
+ "eval_pearson": 0.9096840620040894,
60
+ "eval_runtime": 0.9756,
61
+ "eval_samples_per_second": 1537.485,
62
+ "eval_spearmanr": 0.9127531344274221,
63
+ "eval_steps_per_second": 3.075,
64
+ "step": 600
65
+ },
66
+ {
67
+ "epoch": 4.444444444444445,
68
+ "grad_norm": 1.1739941835403442,
69
+ "learning_rate": 0.00030398717786598174,
70
+ "loss": 0.2037,
71
+ "step": 800
72
+ },
73
+ {
74
+ "epoch": 4.444444444444445,
75
+ "eval_combined_score": 0.91115951544696,
76
+ "eval_loss": 0.39383837580680847,
77
+ "eval_pearson": 0.9108015894889832,
78
+ "eval_runtime": 0.882,
79
+ "eval_samples_per_second": 1700.608,
80
+ "eval_spearmanr": 0.9115174414049371,
81
+ "eval_steps_per_second": 3.401,
82
+ "step": 800
83
+ },
84
+ {
85
+ "epoch": 5.555555555555555,
86
+ "grad_norm": 1.0420633554458618,
87
+ "learning_rate": 0.00015629528624926277,
88
+ "loss": 0.1704,
89
+ "step": 1000
90
+ },
91
+ {
92
+ "epoch": 5.555555555555555,
93
+ "eval_combined_score": 0.910017263887517,
94
+ "eval_loss": 0.4557992219924927,
95
+ "eval_pearson": 0.909343421459198,
96
+ "eval_runtime": 0.969,
97
+ "eval_samples_per_second": 1548.029,
98
+ "eval_spearmanr": 0.910691106315836,
99
+ "eval_steps_per_second": 3.096,
100
+ "step": 1000
101
+ },
102
+ {
103
+ "epoch": 6.666666666666667,
104
+ "grad_norm": 1.1293386220932007,
105
+ "learning_rate": 4.79575082052621e-05,
106
+ "loss": 0.1427,
107
+ "step": 1200
108
+ },
109
+ {
110
+ "epoch": 6.666666666666667,
111
+ "eval_combined_score": 0.9120661209326952,
112
+ "eval_loss": 0.4218188524246216,
113
+ "eval_pearson": 0.9118423461914062,
114
+ "eval_runtime": 0.8308,
115
+ "eval_samples_per_second": 1805.384,
116
+ "eval_spearmanr": 0.9122898956739842,
117
+ "eval_steps_per_second": 3.611,
118
+ "step": 1200
119
+ },
120
+ {
121
+ "epoch": 7.777777777777778,
122
+ "grad_norm": 3.8367955684661865,
123
+ "learning_rate": 9.843417688702115e-07,
124
+ "loss": 0.1384,
125
+ "step": 1400
126
+ },
127
+ {
128
+ "epoch": 7.777777777777778,
129
+ "eval_combined_score": 0.9115026493952896,
130
+ "eval_loss": 0.4243917763233185,
131
+ "eval_pearson": 0.9111843109130859,
132
+ "eval_runtime": 0.8314,
133
+ "eval_samples_per_second": 1804.171,
134
+ "eval_spearmanr": 0.9118209878774931,
135
+ "eval_steps_per_second": 3.608,
136
+ "step": 1400
137
+ },
138
+ {
139
+ "epoch": 8.0,
140
+ "step": 1440,
141
+ "total_flos": 3072533598695424.0,
142
+ "train_loss": 0.37011299994256763,
143
+ "train_runtime": 408.166,
144
+ "train_samples_per_second": 112.68,
145
+ "train_steps_per_second": 3.528
146
+ }
147
+ ],
148
+ "logging_steps": 200,
149
+ "max_steps": 1440,
150
+ "num_input_tokens_seen": 0,
151
+ "num_train_epochs": 8,
152
+ "save_steps": 200,
153
+ "stateful_callbacks": {
154
+ "TrainerControl": {
155
+ "args": {
156
+ "should_epoch_stop": false,
157
+ "should_evaluate": false,
158
+ "should_log": false,
159
+ "should_save": true,
160
+ "should_training_stop": true
161
+ },
162
+ "attributes": {}
163
+ }
164
+ },
165
+ "total_flos": 3072533598695424.0,
166
+ "train_batch_size": 32,
167
+ "trial_name": null,
168
+ "trial_params": null
169
+ }
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h42m49,sd44/STS-B.tsv ADDED
@@ -0,0 +1,1380 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ index prediction
2
+ 0 3.582
3
+ 1 4.442
4
+ 2 5.000
5
+ 3 5.000
6
+ 4 2.129
7
+ 5 1.917
8
+ 6 3.794
9
+ 7 1.664
10
+ 8 2.549
11
+ 9 1.871
12
+ 10 1.871
13
+ 11 5.000
14
+ 12 0.718
15
+ 13 5.000
16
+ 14 1.994
17
+ 15 1.820
18
+ 16 4.564
19
+ 17 2.940
20
+ 18 3.931
21
+ 19 1.101
22
+ 20 1.832
23
+ 21 0.447
24
+ 22 5.000
25
+ 23 5.000
26
+ 24 1.521
27
+ 25 2.808
28
+ 26 0.954
29
+ 27 2.214
30
+ 28 0.835
31
+ 29 2.808
32
+ 30 3.260
33
+ 31 4.736
34
+ 32 1.197
35
+ 33 4.607
36
+ 34 3.516
37
+ 35 0.421
38
+ 36 0.619
39
+ 37 2.136
40
+ 38 5.000
41
+ 39 0.654
42
+ 40 4.006
43
+ 41 2.798
44
+ 42 3.142
45
+ 43 0.831
46
+ 44 1.758
47
+ 45 0.581
48
+ 46 0.214
49
+ 47 0.839
50
+ 48 4.964
51
+ 49 1.904
52
+ 50 0.942
53
+ 51 1.612
54
+ 52 0.408
55
+ 53 0.844
56
+ 54 1.848
57
+ 55 0.331
58
+ 56 4.041
59
+ 57 0.861
60
+ 58 4.259
61
+ 59 0.814
62
+ 60 5.000
63
+ 61 0.000
64
+ 62 4.190
65
+ 63 0.836
66
+ 64 3.919
67
+ 65 0.059
68
+ 66 0.360
69
+ 67 4.197
70
+ 68 0.197
71
+ 69 4.463
72
+ 70 2.274
73
+ 71 4.110
74
+ 72 0.087
75
+ 73 4.273
76
+ 74 3.983
77
+ 75 0.217
78
+ 76 0.365
79
+ 77 0.000
80
+ 78 2.591
81
+ 79 4.170
82
+ 80 0.232
83
+ 81 0.723
84
+ 82 3.137
85
+ 83 0.410
86
+ 84 0.316
87
+ 85 1.591
88
+ 86 3.686
89
+ 87 4.711
90
+ 88 3.292
91
+ 89 1.587
92
+ 90 1.486
93
+ 91 3.595
94
+ 92 0.894
95
+ 93 3.895
96
+ 94 2.921
97
+ 95 4.756
98
+ 96 4.734
99
+ 97 4.010
100
+ 98 2.585
101
+ 99 3.326
102
+ 100 0.000
103
+ 101 0.704
104
+ 102 3.506
105
+ 103 3.009
106
+ 104 4.002
107
+ 105 2.944
108
+ 106 0.049
109
+ 107 3.217
110
+ 108 2.098
111
+ 109 3.549
112
+ 110 2.892
113
+ 111 4.027
114
+ 112 3.333
115
+ 113 4.092
116
+ 114 0.000
117
+ 115 0.000
118
+ 116 4.166
119
+ 117 3.041
120
+ 118 0.032
121
+ 119 0.139
122
+ 120 0.663
123
+ 121 0.192
124
+ 122 0.000
125
+ 123 0.379
126
+ 124 0.000
127
+ 125 4.981
128
+ 126 0.651
129
+ 127 2.296
130
+ 128 5.000
131
+ 129 3.881
132
+ 130 4.949
133
+ 131 2.348
134
+ 132 2.969
135
+ 133 5.000
136
+ 134 3.469
137
+ 135 5.000
138
+ 136 5.000
139
+ 137 1.900
140
+ 138 1.840
141
+ 139 4.836
142
+ 140 5.000
143
+ 141 3.802
144
+ 142 4.371
145
+ 143 3.832
146
+ 144 2.169
147
+ 145 1.497
148
+ 146 4.401
149
+ 147 2.970
150
+ 148 1.766
151
+ 149 0.729
152
+ 150 1.945
153
+ 151 1.931
154
+ 152 3.647
155
+ 153 3.437
156
+ 154 3.562
157
+ 155 3.713
158
+ 156 0.312
159
+ 157 0.765
160
+ 158 4.147
161
+ 159 0.764
162
+ 160 5.000
163
+ 161 2.301
164
+ 162 0.551
165
+ 163 3.386
166
+ 164 4.916
167
+ 165 4.051
168
+ 166 4.427
169
+ 167 3.837
170
+ 168 4.708
171
+ 169 4.626
172
+ 170 1.629
173
+ 171 0.569
174
+ 172 4.393
175
+ 173 4.867
176
+ 174 4.098
177
+ 175 3.427
178
+ 176 3.146
179
+ 177 4.256
180
+ 178 0.224
181
+ 179 0.383
182
+ 180 0.448
183
+ 181 3.080
184
+ 182 3.464
185
+ 183 0.981
186
+ 184 1.703
187
+ 185 0.196
188
+ 186 0.000
189
+ 187 0.330
190
+ 188 0.451
191
+ 189 0.158
192
+ 190 3.817
193
+ 191 4.311
194
+ 192 1.486
195
+ 193 3.243
196
+ 194 3.871
197
+ 195 3.402
198
+ 196 3.287
199
+ 197 0.326
200
+ 198 0.022
201
+ 199 3.126
202
+ 200 0.562
203
+ 201 0.679
204
+ 202 0.213
205
+ 203 0.535
206
+ 204 0.251
207
+ 205 0.499
208
+ 206 3.519
209
+ 207 0.073
210
+ 208 2.272
211
+ 209 0.706
212
+ 210 0.663
213
+ 211 0.259
214
+ 212 0.318
215
+ 213 0.798
216
+ 214 3.134
217
+ 215 0.700
218
+ 216 1.423
219
+ 217 3.472
220
+ 218 3.735
221
+ 219 3.075
222
+ 220 0.637
223
+ 221 0.098
224
+ 222 0.954
225
+ 223 1.505
226
+ 224 0.178
227
+ 225 0.155
228
+ 226 2.877
229
+ 227 2.605
230
+ 228 0.000
231
+ 229 1.781
232
+ 230 0.000
233
+ 231 0.003
234
+ 232 4.897
235
+ 233 2.574
236
+ 234 4.273
237
+ 235 2.351
238
+ 236 0.069
239
+ 237 0.000
240
+ 238 0.000
241
+ 239 0.038
242
+ 240 3.135
243
+ 241 0.000
244
+ 242 0.021
245
+ 243 0.006
246
+ 244 3.561
247
+ 245 0.160
248
+ 246 2.955
249
+ 247 3.357
250
+ 248 0.000
251
+ 249 0.570
252
+ 250 3.794
253
+ 251 4.447
254
+ 252 3.966
255
+ 253 2.427
256
+ 254 0.521
257
+ 255 3.755
258
+ 256 0.469
259
+ 257 4.337
260
+ 258 2.996
261
+ 259 0.020
262
+ 260 5.000
263
+ 261 3.384
264
+ 262 4.203
265
+ 263 4.231
266
+ 264 3.152
267
+ 265 1.180
268
+ 266 2.212
269
+ 267 3.708
270
+ 268 1.150
271
+ 269 5.000
272
+ 270 4.265
273
+ 271 0.249
274
+ 272 3.698
275
+ 273 5.000
276
+ 274 4.084
277
+ 275 5.000
278
+ 276 0.431
279
+ 277 0.000
280
+ 278 4.357
281
+ 279 2.935
282
+ 280 0.507
283
+ 281 2.169
284
+ 282 3.716
285
+ 283 4.187
286
+ 284 3.811
287
+ 285 4.348
288
+ 286 1.270
289
+ 287 3.747
290
+ 288 3.403
291
+ 289 0.917
292
+ 290 5.000
293
+ 291 0.739
294
+ 292 2.419
295
+ 293 4.875
296
+ 294 0.000
297
+ 295 0.405
298
+ 296 1.855
299
+ 297 3.727
300
+ 298 2.825
301
+ 299 2.913
302
+ 300 5.000
303
+ 301 2.984
304
+ 302 1.526
305
+ 303 4.867
306
+ 304 3.665
307
+ 305 4.387
308
+ 306 4.567
309
+ 307 3.997
310
+ 308 3.921
311
+ 309 4.376
312
+ 310 1.966
313
+ 311 0.132
314
+ 312 0.723
315
+ 313 2.856
316
+ 314 3.557
317
+ 315 4.609
318
+ 316 3.656
319
+ 317 3.517
320
+ 318 0.000
321
+ 319 0.106
322
+ 320 3.575
323
+ 321 3.171
324
+ 322 4.395
325
+ 323 2.790
326
+ 324 2.574
327
+ 325 0.245
328
+ 326 2.467
329
+ 327 3.629
330
+ 328 4.620
331
+ 329 3.743
332
+ 330 5.000
333
+ 331 4.871
334
+ 332 3.862
335
+ 333 2.272
336
+ 334 3.678
337
+ 335 1.729
338
+ 336 3.938
339
+ 337 4.122
340
+ 338 0.000
341
+ 339 0.527
342
+ 340 4.021
343
+ 341 4.807
344
+ 342 4.858
345
+ 343 2.200
346
+ 344 3.595
347
+ 345 0.265
348
+ 346 4.388
349
+ 347 0.164
350
+ 348 4.223
351
+ 349 0.162
352
+ 350 4.510
353
+ 351 3.683
354
+ 352 0.973
355
+ 353 4.309
356
+ 354 2.927
357
+ 355 0.086
358
+ 356 2.463
359
+ 357 4.908
360
+ 358 0.259
361
+ 359 4.039
362
+ 360 5.000
363
+ 361 4.574
364
+ 362 2.493
365
+ 363 3.774
366
+ 364 4.790
367
+ 365 0.000
368
+ 366 1.177
369
+ 367 4.093
370
+ 368 4.270
371
+ 369 4.915
372
+ 370 3.463
373
+ 371 4.647
374
+ 372 0.025
375
+ 373 0.853
376
+ 374 1.158
377
+ 375 4.356
378
+ 376 2.234
379
+ 377 0.000
380
+ 378 2.334
381
+ 379 1.706
382
+ 380 3.193
383
+ 381 4.305
384
+ 382 2.615
385
+ 383 0.429
386
+ 384 3.704
387
+ 385 2.210
388
+ 386 5.000
389
+ 387 1.316
390
+ 388 2.706
391
+ 389 1.963
392
+ 390 3.962
393
+ 391 2.133
394
+ 392 0.000
395
+ 393 2.014
396
+ 394 2.668
397
+ 395 0.144
398
+ 396 0.154
399
+ 397 5.000
400
+ 398 2.931
401
+ 399 2.604
402
+ 400 2.494
403
+ 401 0.190
404
+ 402 2.295
405
+ 403 4.191
406
+ 404 3.429
407
+ 405 4.040
408
+ 406 0.293
409
+ 407 2.987
410
+ 408 2.468
411
+ 409 3.365
412
+ 410 4.811
413
+ 411 5.000
414
+ 412 1.882
415
+ 413 4.833
416
+ 414 0.526
417
+ 415 1.447
418
+ 416 0.000
419
+ 417 4.934
420
+ 418 4.423
421
+ 419 3.826
422
+ 420 5.000
423
+ 421 1.645
424
+ 422 4.239
425
+ 423 1.394
426
+ 424 2.243
427
+ 425 3.371
428
+ 426 0.000
429
+ 427 5.000
430
+ 428 2.828
431
+ 429 5.000
432
+ 430 0.588
433
+ 431 5.000
434
+ 432 0.293
435
+ 433 3.751
436
+ 434 0.225
437
+ 435 2.521
438
+ 436 2.595
439
+ 437 5.000
440
+ 438 3.516
441
+ 439 4.735
442
+ 440 3.570
443
+ 441 3.065
444
+ 442 0.000
445
+ 443 3.409
446
+ 444 3.890
447
+ 445 0.000
448
+ 446 0.946
449
+ 447 3.936
450
+ 448 0.477
451
+ 449 1.657
452
+ 450 3.661
453
+ 451 2.916
454
+ 452 0.303
455
+ 453 0.252
456
+ 454 4.492
457
+ 455 2.630
458
+ 456 1.979
459
+ 457 5.000
460
+ 458 1.326
461
+ 459 2.923
462
+ 460 0.139
463
+ 461 0.675
464
+ 462 0.000
465
+ 463 1.942
466
+ 464 3.154
467
+ 465 4.371
468
+ 466 2.047
469
+ 467 4.925
470
+ 468 4.946
471
+ 469 0.298
472
+ 470 2.348
473
+ 471 4.560
474
+ 472 2.096
475
+ 473 4.881
476
+ 474 2.865
477
+ 475 2.504
478
+ 476 0.930
479
+ 477 3.677
480
+ 478 4.346
481
+ 479 0.127
482
+ 480 4.987
483
+ 481 2.509
484
+ 482 4.097
485
+ 483 0.000
486
+ 484 4.528
487
+ 485 0.392
488
+ 486 3.800
489
+ 487 4.592
490
+ 488 3.576
491
+ 489 4.981
492
+ 490 3.964
493
+ 491 4.274
494
+ 492 3.694
495
+ 493 3.534
496
+ 494 3.627
497
+ 495 2.540
498
+ 496 1.297
499
+ 497 5.000
500
+ 498 0.243
501
+ 499 5.000
502
+ 500 1.932
503
+ 501 2.785
504
+ 502 1.409
505
+ 503 2.211
506
+ 504 1.258
507
+ 505 3.291
508
+ 506 3.506
509
+ 507 0.753
510
+ 508 2.967
511
+ 509 0.000
512
+ 510 1.200
513
+ 511 1.384
514
+ 512 2.538
515
+ 513 0.355
516
+ 514 1.206
517
+ 515 0.246
518
+ 516 1.398
519
+ 517 2.212
520
+ 518 1.369
521
+ 519 4.003
522
+ 520 1.563
523
+ 521 2.555
524
+ 522 4.591
525
+ 523 0.966
526
+ 524 3.012
527
+ 525 3.635
528
+ 526 1.758
529
+ 527 0.839
530
+ 528 2.300
531
+ 529 0.600
532
+ 530 0.709
533
+ 531 0.000
534
+ 532 0.931
535
+ 533 0.485
536
+ 534 0.182
537
+ 535 0.413
538
+ 536 2.136
539
+ 537 1.229
540
+ 538 2.326
541
+ 539 3.184
542
+ 540 2.415
543
+ 541 1.076
544
+ 542 2.069
545
+ 543 3.323
546
+ 544 4.548
547
+ 545 5.000
548
+ 546 1.093
549
+ 547 2.394
550
+ 548 3.205
551
+ 549 4.306
552
+ 550 4.031
553
+ 551 0.955
554
+ 552 3.171
555
+ 553 2.445
556
+ 554 1.875
557
+ 555 2.532
558
+ 556 2.782
559
+ 557 0.981
560
+ 558 5.000
561
+ 559 3.893
562
+ 560 4.219
563
+ 561 0.046
564
+ 562 1.416
565
+ 563 1.206
566
+ 564 0.829
567
+ 565 2.235
568
+ 566 5.000
569
+ 567 2.892
570
+ 568 0.479
571
+ 569 2.015
572
+ 570 1.989
573
+ 571 0.840
574
+ 572 0.450
575
+ 573 3.481
576
+ 574 4.820
577
+ 575 2.383
578
+ 576 4.931
579
+ 577 3.411
580
+ 578 2.215
581
+ 579 1.260
582
+ 580 1.328
583
+ 581 2.919
584
+ 582 2.787
585
+ 583 1.056
586
+ 584 3.069
587
+ 585 4.141
588
+ 586 0.492
589
+ 587 5.000
590
+ 588 1.160
591
+ 589 1.466
592
+ 590 1.663
593
+ 591 2.185
594
+ 592 4.885
595
+ 593 1.145
596
+ 594 0.681
597
+ 595 3.433
598
+ 596 2.060
599
+ 597 0.334
600
+ 598 3.839
601
+ 599 4.794
602
+ 600 2.345
603
+ 601 0.273
604
+ 602 3.657
605
+ 603 1.634
606
+ 604 0.897
607
+ 605 3.994
608
+ 606 1.846
609
+ 607 1.596
610
+ 608 3.256
611
+ 609 2.588
612
+ 610 5.000
613
+ 611 2.884
614
+ 612 1.045
615
+ 613 0.689
616
+ 614 3.487
617
+ 615 2.102
618
+ 616 2.087
619
+ 617 0.407
620
+ 618 2.753
621
+ 619 4.623
622
+ 620 0.407
623
+ 621 2.640
624
+ 622 0.719
625
+ 623 5.000
626
+ 624 2.657
627
+ 625 2.746
628
+ 626 2.922
629
+ 627 0.263
630
+ 628 0.779
631
+ 629 2.307
632
+ 630 1.663
633
+ 631 5.000
634
+ 632 3.243
635
+ 633 2.888
636
+ 634 1.470
637
+ 635 2.521
638
+ 636 4.162
639
+ 637 2.252
640
+ 638 2.841
641
+ 639 2.893
642
+ 640 0.800
643
+ 641 1.388
644
+ 642 2.545
645
+ 643 2.927
646
+ 644 1.855
647
+ 645 4.534
648
+ 646 3.266
649
+ 647 3.943
650
+ 648 4.987
651
+ 649 3.365
652
+ 650 2.074
653
+ 651 4.486
654
+ 652 0.795
655
+ 653 1.252
656
+ 654 1.724
657
+ 655 3.409
658
+ 656 4.742
659
+ 657 1.579
660
+ 658 1.791
661
+ 659 1.213
662
+ 660 1.576
663
+ 661 1.605
664
+ 662 0.638
665
+ 663 2.152
666
+ 664 5.000
667
+ 665 1.130
668
+ 666 3.763
669
+ 667 3.264
670
+ 668 3.259
671
+ 669 2.631
672
+ 670 2.561
673
+ 671 3.978
674
+ 672 1.884
675
+ 673 4.896
676
+ 674 4.899
677
+ 675 0.742
678
+ 676 2.012
679
+ 677 2.938
680
+ 678 1.603
681
+ 679 2.633
682
+ 680 3.456
683
+ 681 1.394
684
+ 682 3.088
685
+ 683 2.432
686
+ 684 3.881
687
+ 685 2.061
688
+ 686 4.146
689
+ 687 1.932
690
+ 688 3.913
691
+ 689 4.603
692
+ 690 2.311
693
+ 691 2.513
694
+ 692 4.894
695
+ 693 1.474
696
+ 694 3.510
697
+ 695 1.530
698
+ 696 2.760
699
+ 697 0.607
700
+ 698 1.360
701
+ 699 4.470
702
+ 700 4.235
703
+ 701 2.125
704
+ 702 4.881
705
+ 703 4.659
706
+ 704 2.924
707
+ 705 1.793
708
+ 706 2.312
709
+ 707 0.987
710
+ 708 3.574
711
+ 709 1.853
712
+ 710 4.347
713
+ 711 1.242
714
+ 712 2.070
715
+ 713 1.450
716
+ 714 2.006
717
+ 715 1.834
718
+ 716 2.186
719
+ 717 4.294
720
+ 718 2.450
721
+ 719 5.000
722
+ 720 5.000
723
+ 721 3.116
724
+ 722 2.263
725
+ 723 1.382
726
+ 724 4.174
727
+ 725 1.037
728
+ 726 3.670
729
+ 727 4.363
730
+ 728 2.765
731
+ 729 3.430
732
+ 730 3.304
733
+ 731 1.930
734
+ 732 2.527
735
+ 733 0.350
736
+ 734 2.683
737
+ 735 1.637
738
+ 736 1.993
739
+ 737 4.663
740
+ 738 2.335
741
+ 739 2.820
742
+ 740 4.690
743
+ 741 1.971
744
+ 742 1.679
745
+ 743 3.579
746
+ 744 1.950
747
+ 745 3.341
748
+ 746 0.934
749
+ 747 1.410
750
+ 748 3.925
751
+ 749 1.381
752
+ 750 2.001
753
+ 751 2.000
754
+ 752 2.591
755
+ 753 1.145
756
+ 754 1.647
757
+ 755 4.797
758
+ 756 4.740
759
+ 757 4.014
760
+ 758 1.960
761
+ 759 4.041
762
+ 760 3.292
763
+ 761 0.431
764
+ 762 0.586
765
+ 763 1.857
766
+ 764 3.455
767
+ 765 4.117
768
+ 766 3.698
769
+ 767 2.260
770
+ 768 3.993
771
+ 769 1.650
772
+ 770 2.461
773
+ 771 1.602
774
+ 772 1.673
775
+ 773 4.771
776
+ 774 4.731
777
+ 775 4.030
778
+ 776 1.694
779
+ 777 0.561
780
+ 778 1.902
781
+ 779 0.940
782
+ 780 2.279
783
+ 781 2.608
784
+ 782 5.000
785
+ 783 4.560
786
+ 784 2.907
787
+ 785 0.815
788
+ 786 1.005
789
+ 787 3.006
790
+ 788 5.000
791
+ 789 2.399
792
+ 790 2.935
793
+ 791 2.737
794
+ 792 1.823
795
+ 793 1.725
796
+ 794 2.789
797
+ 795 3.341
798
+ 796 1.722
799
+ 797 4.287
800
+ 798 0.790
801
+ 799 1.437
802
+ 800 1.806
803
+ 801 2.287
804
+ 802 1.355
805
+ 803 0.690
806
+ 804 1.640
807
+ 805 5.000
808
+ 806 3.143
809
+ 807 2.784
810
+ 808 3.937
811
+ 809 2.277
812
+ 810 3.767
813
+ 811 2.078
814
+ 812 2.175
815
+ 813 4.318
816
+ 814 4.558
817
+ 815 1.319
818
+ 816 3.431
819
+ 817 1.809
820
+ 818 3.479
821
+ 819 2.185
822
+ 820 2.647
823
+ 821 2.161
824
+ 822 1.267
825
+ 823 4.123
826
+ 824 2.546
827
+ 825 1.884
828
+ 826 3.130
829
+ 827 1.016
830
+ 828 3.834
831
+ 829 5.000
832
+ 830 1.728
833
+ 831 2.006
834
+ 832 1.369
835
+ 833 4.692
836
+ 834 1.932
837
+ 835 2.755
838
+ 836 4.532
839
+ 837 1.540
840
+ 838 1.711
841
+ 839 3.741
842
+ 840 3.319
843
+ 841 1.993
844
+ 842 4.452
845
+ 843 3.852
846
+ 844 4.320
847
+ 845 4.255
848
+ 846 2.836
849
+ 847 2.303
850
+ 848 1.748
851
+ 849 3.368
852
+ 850 4.681
853
+ 851 1.160
854
+ 852 4.162
855
+ 853 4.691
856
+ 854 5.000
857
+ 855 1.846
858
+ 856 0.601
859
+ 857 2.886
860
+ 858 1.747
861
+ 859 3.225
862
+ 860 3.841
863
+ 861 4.778
864
+ 862 3.844
865
+ 863 0.477
866
+ 864 0.924
867
+ 865 0.820
868
+ 866 1.032
869
+ 867 0.619
870
+ 868 0.816
871
+ 869 4.570
872
+ 870 3.861
873
+ 871 2.512
874
+ 872 2.788
875
+ 873 4.249
876
+ 874 1.990
877
+ 875 4.226
878
+ 876 2.192
879
+ 877 4.486
880
+ 878 4.448
881
+ 879 1.978
882
+ 880 4.163
883
+ 881 3.379
884
+ 882 3.086
885
+ 883 4.549
886
+ 884 1.905
887
+ 885 4.057
888
+ 886 3.266
889
+ 887 4.409
890
+ 888 3.186
891
+ 889 3.171
892
+ 890 5.000
893
+ 891 3.319
894
+ 892 1.958
895
+ 893 3.946
896
+ 894 4.129
897
+ 895 4.512
898
+ 896 3.439
899
+ 897 3.192
900
+ 898 1.350
901
+ 899 3.754
902
+ 900 3.987
903
+ 901 4.244
904
+ 902 3.460
905
+ 903 4.498
906
+ 904 4.688
907
+ 905 3.157
908
+ 906 3.300
909
+ 907 4.121
910
+ 908 3.616
911
+ 909 3.014
912
+ 910 3.772
913
+ 911 2.335
914
+ 912 2.622
915
+ 913 4.410
916
+ 914 3.592
917
+ 915 4.838
918
+ 916 4.072
919
+ 917 2.846
920
+ 918 4.394
921
+ 919 3.938
922
+ 920 3.604
923
+ 921 3.924
924
+ 922 5.000
925
+ 923 4.267
926
+ 924 4.010
927
+ 925 2.489
928
+ 926 3.305
929
+ 927 2.459
930
+ 928 3.708
931
+ 929 3.284
932
+ 930 3.967
933
+ 931 3.507
934
+ 932 4.232
935
+ 933 3.053
936
+ 934 1.948
937
+ 935 2.229
938
+ 936 3.705
939
+ 937 3.348
940
+ 938 3.067
941
+ 939 2.614
942
+ 940 4.179
943
+ 941 3.741
944
+ 942 3.329
945
+ 943 3.182
946
+ 944 3.832
947
+ 945 4.510
948
+ 946 4.007
949
+ 947 4.082
950
+ 948 3.701
951
+ 949 2.154
952
+ 950 3.262
953
+ 951 4.629
954
+ 952 4.464
955
+ 953 3.215
956
+ 954 3.184
957
+ 955 2.012
958
+ 956 3.433
959
+ 957 1.406
960
+ 958 4.295
961
+ 959 3.259
962
+ 960 3.359
963
+ 961 3.631
964
+ 962 4.475
965
+ 963 3.020
966
+ 964 3.388
967
+ 965 3.606
968
+ 966 4.083
969
+ 967 4.043
970
+ 968 3.622
971
+ 969 3.363
972
+ 970 2.607
973
+ 971 3.320
974
+ 972 3.861
975
+ 973 3.264
976
+ 974 3.723
977
+ 975 2.582
978
+ 976 3.322
979
+ 977 4.916
980
+ 978 4.234
981
+ 979 3.910
982
+ 980 2.260
983
+ 981 4.561
984
+ 982 2.931
985
+ 983 2.962
986
+ 984 2.603
987
+ 985 3.362
988
+ 986 4.442
989
+ 987 3.324
990
+ 988 4.072
991
+ 989 4.010
992
+ 990 4.554
993
+ 991 1.782
994
+ 992 3.710
995
+ 993 3.444
996
+ 994 2.795
997
+ 995 3.174
998
+ 996 1.931
999
+ 997 4.340
1000
+ 998 4.162
1001
+ 999 3.218
1002
+ 1000 3.290
1003
+ 1001 4.086
1004
+ 1002 4.170
1005
+ 1003 3.515
1006
+ 1004 1.566
1007
+ 1005 2.695
1008
+ 1006 4.470
1009
+ 1007 3.314
1010
+ 1008 1.908
1011
+ 1009 3.631
1012
+ 1010 4.233
1013
+ 1011 4.276
1014
+ 1012 4.806
1015
+ 1013 2.430
1016
+ 1014 4.373
1017
+ 1015 3.940
1018
+ 1016 4.367
1019
+ 1017 3.056
1020
+ 1018 4.846
1021
+ 1019 4.213
1022
+ 1020 3.128
1023
+ 1021 4.462
1024
+ 1022 4.176
1025
+ 1023 1.829
1026
+ 1024 3.437
1027
+ 1025 3.572
1028
+ 1026 3.385
1029
+ 1027 1.906
1030
+ 1028 3.994
1031
+ 1029 1.879
1032
+ 1030 4.800
1033
+ 1031 4.145
1034
+ 1032 4.288
1035
+ 1033 3.149
1036
+ 1034 4.957
1037
+ 1035 3.685
1038
+ 1036 3.267
1039
+ 1037 3.917
1040
+ 1038 4.617
1041
+ 1039 4.856
1042
+ 1040 3.295
1043
+ 1041 4.256
1044
+ 1042 3.308
1045
+ 1043 3.591
1046
+ 1044 4.058
1047
+ 1045 2.823
1048
+ 1046 1.618
1049
+ 1047 3.052
1050
+ 1048 4.326
1051
+ 1049 4.872
1052
+ 1050 3.857
1053
+ 1051 3.852
1054
+ 1052 4.045
1055
+ 1053 3.224
1056
+ 1054 3.048
1057
+ 1055 4.409
1058
+ 1056 4.185
1059
+ 1057 3.209
1060
+ 1058 4.672
1061
+ 1059 3.591
1062
+ 1060 2.944
1063
+ 1061 3.518
1064
+ 1062 2.987
1065
+ 1063 4.480
1066
+ 1064 3.021
1067
+ 1065 2.644
1068
+ 1066 3.942
1069
+ 1067 2.794
1070
+ 1068 3.392
1071
+ 1069 3.697
1072
+ 1070 5.000
1073
+ 1071 2.899
1074
+ 1072 2.982
1075
+ 1073 1.015
1076
+ 1074 3.129
1077
+ 1075 2.099
1078
+ 1076 3.679
1079
+ 1077 3.809
1080
+ 1078 3.925
1081
+ 1079 2.077
1082
+ 1080 4.042
1083
+ 1081 3.504
1084
+ 1082 3.913
1085
+ 1083 1.105
1086
+ 1084 3.898
1087
+ 1085 3.898
1088
+ 1086 2.294
1089
+ 1087 4.205
1090
+ 1088 1.342
1091
+ 1089 4.087
1092
+ 1090 3.370
1093
+ 1091 4.763
1094
+ 1092 4.157
1095
+ 1093 3.851
1096
+ 1094 3.406
1097
+ 1095 3.603
1098
+ 1096 3.530
1099
+ 1097 3.940
1100
+ 1098 3.269
1101
+ 1099 1.930
1102
+ 1100 2.926
1103
+ 1101 3.796
1104
+ 1102 3.666
1105
+ 1103 3.777
1106
+ 1104 3.831
1107
+ 1105 3.560
1108
+ 1106 3.527
1109
+ 1107 2.129
1110
+ 1108 3.696
1111
+ 1109 2.413
1112
+ 1110 1.198
1113
+ 1111 4.597
1114
+ 1112 3.500
1115
+ 1113 4.793
1116
+ 1114 3.891
1117
+ 1115 3.828
1118
+ 1116 2.716
1119
+ 1117 3.873
1120
+ 1118 4.646
1121
+ 1119 3.252
1122
+ 1120 2.588
1123
+ 1121 0.966
1124
+ 1122 4.071
1125
+ 1123 3.366
1126
+ 1124 3.463
1127
+ 1125 4.742
1128
+ 1126 3.780
1129
+ 1127 4.508
1130
+ 1128 4.300
1131
+ 1129 2.088
1132
+ 1130 1.514
1133
+ 1131 4.232
1134
+ 1132 4.942
1135
+ 1133 0.785
1136
+ 1134 2.614
1137
+ 1135 2.557
1138
+ 1136 2.700
1139
+ 1137 3.451
1140
+ 1138 0.592
1141
+ 1139 4.953
1142
+ 1140 1.306
1143
+ 1141 0.820
1144
+ 1142 2.419
1145
+ 1143 4.372
1146
+ 1144 3.877
1147
+ 1145 2.097
1148
+ 1146 1.069
1149
+ 1147 4.691
1150
+ 1148 3.956
1151
+ 1149 5.000
1152
+ 1150 0.037
1153
+ 1151 0.226
1154
+ 1152 5.000
1155
+ 1153 5.000
1156
+ 1154 5.000
1157
+ 1155 3.803
1158
+ 1156 4.434
1159
+ 1157 0.110
1160
+ 1158 4.722
1161
+ 1159 1.156
1162
+ 1160 4.102
1163
+ 1161 3.673
1164
+ 1162 0.794
1165
+ 1163 4.970
1166
+ 1164 4.830
1167
+ 1165 0.957
1168
+ 1166 4.117
1169
+ 1167 1.126
1170
+ 1168 3.386
1171
+ 1169 0.393
1172
+ 1170 2.479
1173
+ 1171 4.606
1174
+ 1172 3.063
1175
+ 1173 3.946
1176
+ 1174 0.252
1177
+ 1175 0.000
1178
+ 1176 5.000
1179
+ 1177 1.495
1180
+ 1178 3.192
1181
+ 1179 0.463
1182
+ 1180 0.669
1183
+ 1181 3.893
1184
+ 1182 1.798
1185
+ 1183 2.904
1186
+ 1184 0.480
1187
+ 1185 4.614
1188
+ 1186 5.000
1189
+ 1187 3.804
1190
+ 1188 3.253
1191
+ 1189 4.894
1192
+ 1190 3.587
1193
+ 1191 4.377
1194
+ 1192 1.589
1195
+ 1193 1.971
1196
+ 1194 1.041
1197
+ 1195 2.381
1198
+ 1196 4.911
1199
+ 1197 4.588
1200
+ 1198 3.688
1201
+ 1199 3.853
1202
+ 1200 2.551
1203
+ 1201 0.755
1204
+ 1202 3.141
1205
+ 1203 4.716
1206
+ 1204 4.233
1207
+ 1205 3.068
1208
+ 1206 5.000
1209
+ 1207 4.170
1210
+ 1208 2.129
1211
+ 1209 2.043
1212
+ 1210 2.555
1213
+ 1211 4.424
1214
+ 1212 4.363
1215
+ 1213 2.350
1216
+ 1214 5.000
1217
+ 1215 4.466
1218
+ 1216 1.118
1219
+ 1217 1.932
1220
+ 1218 1.536
1221
+ 1219 1.616
1222
+ 1220 1.701
1223
+ 1221 3.263
1224
+ 1222 4.161
1225
+ 1223 3.284
1226
+ 1224 1.640
1227
+ 1225 3.370
1228
+ 1226 3.222
1229
+ 1227 4.917
1230
+ 1228 4.782
1231
+ 1229 3.093
1232
+ 1230 3.027
1233
+ 1231 3.524
1234
+ 1232 3.513
1235
+ 1233 4.701
1236
+ 1234 4.204
1237
+ 1235 2.361
1238
+ 1236 2.444
1239
+ 1237 1.037
1240
+ 1238 1.640
1241
+ 1239 1.052
1242
+ 1240 1.010
1243
+ 1241 1.731
1244
+ 1242 2.394
1245
+ 1243 2.523
1246
+ 1244 0.073
1247
+ 1245 5.000
1248
+ 1246 4.216
1249
+ 1247 1.600
1250
+ 1248 0.318
1251
+ 1249 5.000
1252
+ 1250 4.505
1253
+ 1251 3.760
1254
+ 1252 4.734
1255
+ 1253 2.625
1256
+ 1254 0.000
1257
+ 1255 4.858
1258
+ 1256 2.077
1259
+ 1257 4.281
1260
+ 1258 4.066
1261
+ 1259 4.716
1262
+ 1260 3.466
1263
+ 1261 1.213
1264
+ 1262 4.306
1265
+ 1263 1.300
1266
+ 1264 3.395
1267
+ 1265 3.957
1268
+ 1266 4.079
1269
+ 1267 4.765
1270
+ 1268 1.192
1271
+ 1269 1.004
1272
+ 1270 3.143
1273
+ 1271 3.629
1274
+ 1272 0.846
1275
+ 1273 4.243
1276
+ 1274 4.290
1277
+ 1275 2.312
1278
+ 1276 0.408
1279
+ 1277 3.561
1280
+ 1278 0.884
1281
+ 1279 4.766
1282
+ 1280 0.386
1283
+ 1281 0.352
1284
+ 1282 3.139
1285
+ 1283 4.902
1286
+ 1284 1.706
1287
+ 1285 4.489
1288
+ 1286 4.208
1289
+ 1287 1.161
1290
+ 1288 0.000
1291
+ 1289 2.369
1292
+ 1290 5.000
1293
+ 1291 0.643
1294
+ 1292 4.212
1295
+ 1293 3.570
1296
+ 1294 1.583
1297
+ 1295 0.400
1298
+ 1296 5.000
1299
+ 1297 1.374
1300
+ 1298 0.947
1301
+ 1299 4.759
1302
+ 1300 2.738
1303
+ 1301 2.182
1304
+ 1302 2.828
1305
+ 1303 2.173
1306
+ 1304 2.174
1307
+ 1305 0.491
1308
+ 1306 2.154
1309
+ 1307 0.687
1310
+ 1308 2.400
1311
+ 1309 3.036
1312
+ 1310 3.707
1313
+ 1311 4.974
1314
+ 1312 0.993
1315
+ 1313 3.212
1316
+ 1314 2.835
1317
+ 1315 4.841
1318
+ 1316 4.965
1319
+ 1317 5.000
1320
+ 1318 2.208
1321
+ 1319 0.098
1322
+ 1320 5.000
1323
+ 1321 1.787
1324
+ 1322 5.000
1325
+ 1323 2.681
1326
+ 1324 5.000
1327
+ 1325 4.819
1328
+ 1326 5.000
1329
+ 1327 3.629
1330
+ 1328 0.729
1331
+ 1329 4.922
1332
+ 1330 4.652
1333
+ 1331 2.450
1334
+ 1332 1.818
1335
+ 1333 1.716
1336
+ 1334 2.588
1337
+ 1335 4.846
1338
+ 1336 4.678
1339
+ 1337 4.595
1340
+ 1338 0.469
1341
+ 1339 1.944
1342
+ 1340 3.777
1343
+ 1341 4.592
1344
+ 1342 2.585
1345
+ 1343 0.540
1346
+ 1344 0.837
1347
+ 1345 2.435
1348
+ 1346 4.633
1349
+ 1347 5.000
1350
+ 1348 5.000
1351
+ 1349 2.370
1352
+ 1350 5.000
1353
+ 1351 5.000
1354
+ 1352 5.000
1355
+ 1353 5.000
1356
+ 1354 4.499
1357
+ 1355 2.496
1358
+ 1356 2.151
1359
+ 1357 5.000
1360
+ 1358 3.967
1361
+ 1359 3.023
1362
+ 1360 0.623
1363
+ 1361 3.937
1364
+ 1362 1.288
1365
+ 1363 2.109
1366
+ 1364 5.000
1367
+ 1365 1.949
1368
+ 1366 1.635
1369
+ 1367 2.381
1370
+ 1368 0.239
1371
+ 1369 0.476
1372
+ 1370 0.927
1373
+ 1371 1.277
1374
+ 1372 1.741
1375
+ 1373 1.975
1376
+ 1374 0.469
1377
+ 1375 1.196
1378
+ 1376 1.955
1379
+ 1377 0.149
1380
+ 1378 1.198
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h42m49,sd44/all_results.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 8.0,
3
+ "eval_combined_score": 0.9115528104293169,
4
+ "eval_loss": 0.43174925446510315,
5
+ "eval_pearson": 0.9114007949829102,
6
+ "eval_runtime": 0.84,
7
+ "eval_samples": 1500,
8
+ "eval_samples_per_second": 1785.775,
9
+ "eval_spearmanr": 0.9117048258757235,
10
+ "eval_steps_per_second": 3.572
11
+ }
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h42m49,sd44/eval_results.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 8.0,
3
+ "eval_combined_score": 0.9115528104293169,
4
+ "eval_loss": 0.43174925446510315,
5
+ "eval_pearson": 0.9114007949829102,
6
+ "eval_runtime": 0.84,
7
+ "eval_samples": 1500,
8
+ "eval_samples_per_second": 1785.775,
9
+ "eval_spearmanr": 0.9117048258757235,
10
+ "eval_steps_per_second": 3.572
11
+ }
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h42m49,sd44/ft/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[MASK]": 128000
3
+ }
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h42m49,sd44/ft/special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": {
9
+ "content": "[UNK]",
10
+ "lstrip": false,
11
+ "normalized": true,
12
+ "rstrip": false,
13
+ "single_word": false
14
+ }
15
+ }
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h42m49,sd44/ft/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h42m49,sd44/ft/tokenizer_config.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[CLS]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[SEP]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[UNK]",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "128000": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "[CLS]",
45
+ "clean_up_tokenization_spaces": false,
46
+ "cls_token": "[CLS]",
47
+ "do_lower_case": false,
48
+ "eos_token": "[SEP]",
49
+ "extra_special_tokens": {},
50
+ "mask_token": "[MASK]",
51
+ "model_max_length": 512,
52
+ "pad_token": "[PAD]",
53
+ "padding_side": "right",
54
+ "sep_token": "[SEP]",
55
+ "sp_model_kwargs": {},
56
+ "split_by_punct": false,
57
+ "tokenizer_class": "DebertaV2Tokenizer",
58
+ "unk_token": "[UNK]",
59
+ "vocab_type": "spm"
60
+ }
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h42m49,sd44/ft2/README.md ADDED
@@ -0,0 +1,205 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: microsoft/deberta-v3-base
3
+ library_name: peft
4
+ tags:
5
+ - base_model:adapter:microsoft/deberta-v3-base
6
+ - transformers
7
+ ---
8
+
9
+ # Model Card for Model ID
10
+
11
+ <!-- Provide a quick summary of what the model is/does. -->
12
+
13
+
14
+
15
+ ## Model Details
16
+
17
+ ### Model Description
18
+
19
+ <!-- Provide a longer summary of what this model is. -->
20
+
21
+
22
+
23
+ - **Developed by:** [More Information Needed]
24
+ - **Funded by [optional]:** [More Information Needed]
25
+ - **Shared by [optional]:** [More Information Needed]
26
+ - **Model type:** [More Information Needed]
27
+ - **Language(s) (NLP):** [More Information Needed]
28
+ - **License:** [More Information Needed]
29
+ - **Finetuned from model [optional]:** [More Information Needed]
30
+
31
+ ### Model Sources [optional]
32
+
33
+ <!-- Provide the basic links for the model. -->
34
+
35
+ - **Repository:** [More Information Needed]
36
+ - **Paper [optional]:** [More Information Needed]
37
+ - **Demo [optional]:** [More Information Needed]
38
+
39
+ ## Uses
40
+
41
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
42
+
43
+ ### Direct Use
44
+
45
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
46
+
47
+ [More Information Needed]
48
+
49
+ ### Downstream Use [optional]
50
+
51
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
52
+
53
+ [More Information Needed]
54
+
55
+ ### Out-of-Scope Use
56
+
57
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
58
+
59
+ [More Information Needed]
60
+
61
+ ## Bias, Risks, and Limitations
62
+
63
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
64
+
65
+ [More Information Needed]
66
+
67
+ ### Recommendations
68
+
69
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
70
+
71
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
72
+
73
+ ## How to Get Started with the Model
74
+
75
+ Use the code below to get started with the model.
76
+
77
+ [More Information Needed]
78
+
79
+ ## Training Details
80
+
81
+ ### Training Data
82
+
83
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
84
+
85
+ [More Information Needed]
86
+
87
+ ### Training Procedure
88
+
89
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
90
+
91
+ #### Preprocessing [optional]
92
+
93
+ [More Information Needed]
94
+
95
+
96
+ #### Training Hyperparameters
97
+
98
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
99
+
100
+ #### Speeds, Sizes, Times [optional]
101
+
102
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
103
+
104
+ [More Information Needed]
105
+
106
+ ## Evaluation
107
+
108
+ <!-- This section describes the evaluation protocols and provides the results. -->
109
+
110
+ ### Testing Data, Factors & Metrics
111
+
112
+ #### Testing Data
113
+
114
+ <!-- This should link to a Dataset Card if possible. -->
115
+
116
+ [More Information Needed]
117
+
118
+ #### Factors
119
+
120
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
121
+
122
+ [More Information Needed]
123
+
124
+ #### Metrics
125
+
126
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
127
+
128
+ [More Information Needed]
129
+
130
+ ### Results
131
+
132
+ [More Information Needed]
133
+
134
+ #### Summary
135
+
136
+
137
+
138
+ ## Model Examination [optional]
139
+
140
+ <!-- Relevant interpretability work for the model goes here -->
141
+
142
+ [More Information Needed]
143
+
144
+ ## Environmental Impact
145
+
146
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
147
+
148
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
149
+
150
+ - **Hardware Type:** [More Information Needed]
151
+ - **Hours used:** [More Information Needed]
152
+ - **Cloud Provider:** [More Information Needed]
153
+ - **Compute Region:** [More Information Needed]
154
+ - **Carbon Emitted:** [More Information Needed]
155
+
156
+ ## Technical Specifications [optional]
157
+
158
+ ### Model Architecture and Objective
159
+
160
+ [More Information Needed]
161
+
162
+ ### Compute Infrastructure
163
+
164
+ [More Information Needed]
165
+
166
+ #### Hardware
167
+
168
+ [More Information Needed]
169
+
170
+ #### Software
171
+
172
+ [More Information Needed]
173
+
174
+ ## Citation [optional]
175
+
176
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
177
+
178
+ **BibTeX:**
179
+
180
+ [More Information Needed]
181
+
182
+ **APA:**
183
+
184
+ [More Information Needed]
185
+
186
+ ## Glossary [optional]
187
+
188
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
189
+
190
+ [More Information Needed]
191
+
192
+ ## More Information [optional]
193
+
194
+ [More Information Needed]
195
+
196
+ ## Model Card Authors [optional]
197
+
198
+ [More Information Needed]
199
+
200
+ ## Model Card Contact
201
+
202
+ [More Information Needed]
203
+ ### Framework versions
204
+
205
+ - PEFT 0.18.0
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h42m49,sd44/ft2/adapter_config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": {
3
+ "base_model_class": "DebertaV2ForSequenceClassification",
4
+ "parent_library": "transformers.models.deberta_v2.modeling_deberta_v2"
5
+ },
6
+ "base_model_name_or_path": "microsoft/deberta-v3-base",
7
+ "bias": "none",
8
+ "boft_block_num": 0,
9
+ "boft_block_size": 4,
10
+ "boft_dropout": 0.05,
11
+ "boft_n_butterfly_factor": 2,
12
+ "exclude_modules": null,
13
+ "fan_in_fan_out": false,
14
+ "inference_mode": true,
15
+ "init_weights": true,
16
+ "layers_pattern": null,
17
+ "layers_to_transform": null,
18
+ "modules_to_save": [
19
+ "classifier",
20
+ "pooler"
21
+ ],
22
+ "peft_type": "BOFT",
23
+ "peft_version": "0.18.0",
24
+ "revision": null,
25
+ "target_modules": [
26
+ "attention.output.dense",
27
+ "query_proj",
28
+ "key_proj",
29
+ "intermediate.dense",
30
+ "value_proj",
31
+ "output.dense"
32
+ ],
33
+ "task_type": null
34
+ }
reproduction/glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h42m49,sd44/trainer_state.json ADDED
@@ -0,0 +1,169 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 1200,
3
+ "best_metric": 0.9115528104293169,
4
+ "best_model_checkpoint": "./glue_expBOFT/stsb/dr0.05,mlr7e-04,clr7e-04,ep=8.0t=25d23h42m49,sd44/checkpoint-1200",
5
+ "epoch": 8.0,
6
+ "eval_steps": 200,
7
+ "global_step": 1440,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 1.1111111111111112,
14
+ "grad_norm": 8.726420402526855,
15
+ "learning_rate": 0.0006801144205835404,
16
+ "loss": 1.3406,
17
+ "step": 200
18
+ },
19
+ {
20
+ "epoch": 1.1111111111111112,
21
+ "eval_combined_score": 0.8943649938364275,
22
+ "eval_loss": 0.590000569820404,
23
+ "eval_pearson": 0.8908621072769165,
24
+ "eval_runtime": 1.2083,
25
+ "eval_samples_per_second": 1241.456,
26
+ "eval_spearmanr": 0.8978678803959387,
27
+ "eval_steps_per_second": 2.483,
28
+ "step": 200
29
+ },
30
+ {
31
+ "epoch": 2.2222222222222223,
32
+ "grad_norm": 1.9106940031051636,
33
+ "learning_rate": 0.0005955104829638303,
34
+ "loss": 0.3641,
35
+ "step": 400
36
+ },
37
+ {
38
+ "epoch": 2.2222222222222223,
39
+ "eval_combined_score": 0.9049564374413974,
40
+ "eval_loss": 0.4645957052707672,
41
+ "eval_pearson": 0.9030255079269409,
42
+ "eval_runtime": 0.9964,
43
+ "eval_samples_per_second": 1505.425,
44
+ "eval_spearmanr": 0.9068873669558538,
45
+ "eval_steps_per_second": 3.011,
46
+ "step": 400
47
+ },
48
+ {
49
+ "epoch": 3.3333333333333335,
50
+ "grad_norm": 2.4372506141662598,
51
+ "learning_rate": 0.0004610272871538184,
52
+ "loss": 0.2745,
53
+ "step": 600
54
+ },
55
+ {
56
+ "epoch": 3.3333333333333335,
57
+ "eval_combined_score": 0.9110757735528505,
58
+ "eval_loss": 0.4018155038356781,
59
+ "eval_pearson": 0.909813642501831,
60
+ "eval_runtime": 0.9896,
61
+ "eval_samples_per_second": 1515.766,
62
+ "eval_spearmanr": 0.9123379046038701,
63
+ "eval_steps_per_second": 3.032,
64
+ "step": 600
65
+ },
66
+ {
67
+ "epoch": 4.444444444444445,
68
+ "grad_norm": 1.3469356298446655,
69
+ "learning_rate": 0.00030398717786598174,
70
+ "loss": 0.2145,
71
+ "step": 800
72
+ },
73
+ {
74
+ "epoch": 4.444444444444445,
75
+ "eval_combined_score": 0.9094904305879895,
76
+ "eval_loss": 0.42500725388526917,
77
+ "eval_pearson": 0.90882408618927,
78
+ "eval_runtime": 0.902,
79
+ "eval_samples_per_second": 1663.013,
80
+ "eval_spearmanr": 0.910156774986709,
81
+ "eval_steps_per_second": 3.326,
82
+ "step": 800
83
+ },
84
+ {
85
+ "epoch": 5.555555555555555,
86
+ "grad_norm": 1.4666069746017456,
87
+ "learning_rate": 0.00015629528624926277,
88
+ "loss": 0.1751,
89
+ "step": 1000
90
+ },
91
+ {
92
+ "epoch": 5.555555555555555,
93
+ "eval_combined_score": 0.9102262392356554,
94
+ "eval_loss": 0.4386104643344879,
95
+ "eval_pearson": 0.9095419645309448,
96
+ "eval_runtime": 0.8772,
97
+ "eval_samples_per_second": 1709.904,
98
+ "eval_spearmanr": 0.9109105139403659,
99
+ "eval_steps_per_second": 3.42,
100
+ "step": 1000
101
+ },
102
+ {
103
+ "epoch": 6.666666666666667,
104
+ "grad_norm": 1.1746561527252197,
105
+ "learning_rate": 4.79575082052621e-05,
106
+ "loss": 0.1456,
107
+ "step": 1200
108
+ },
109
+ {
110
+ "epoch": 6.666666666666667,
111
+ "eval_combined_score": 0.9115528104293169,
112
+ "eval_loss": 0.43174925446510315,
113
+ "eval_pearson": 0.9114007949829102,
114
+ "eval_runtime": 1.7155,
115
+ "eval_samples_per_second": 874.37,
116
+ "eval_spearmanr": 0.9117048258757235,
117
+ "eval_steps_per_second": 1.749,
118
+ "step": 1200
119
+ },
120
+ {
121
+ "epoch": 7.777777777777778,
122
+ "grad_norm": 1.453771948814392,
123
+ "learning_rate": 9.843417688702115e-07,
124
+ "loss": 0.1396,
125
+ "step": 1400
126
+ },
127
+ {
128
+ "epoch": 7.777777777777778,
129
+ "eval_combined_score": 0.9113062973566675,
130
+ "eval_loss": 0.42456158995628357,
131
+ "eval_pearson": 0.9110910892486572,
132
+ "eval_runtime": 1.0037,
133
+ "eval_samples_per_second": 1494.43,
134
+ "eval_spearmanr": 0.9115215054646777,
135
+ "eval_steps_per_second": 2.989,
136
+ "step": 1400
137
+ },
138
+ {
139
+ "epoch": 8.0,
140
+ "step": 1440,
141
+ "total_flos": 3072533598695424.0,
142
+ "train_loss": 0.3726578987307019,
143
+ "train_runtime": 419.6789,
144
+ "train_samples_per_second": 109.589,
145
+ "train_steps_per_second": 3.431
146
+ }
147
+ ],
148
+ "logging_steps": 200,
149
+ "max_steps": 1440,
150
+ "num_input_tokens_seen": 0,
151
+ "num_train_epochs": 8,
152
+ "save_steps": 200,
153
+ "stateful_callbacks": {
154
+ "TrainerControl": {
155
+ "args": {
156
+ "should_epoch_stop": false,
157
+ "should_evaluate": false,
158
+ "should_log": false,
159
+ "should_save": true,
160
+ "should_training_stop": true
161
+ },
162
+ "attributes": {}
163
+ }
164
+ },
165
+ "total_flos": 3072533598695424.0,
166
+ "train_batch_size": 32,
167
+ "trial_name": null,
168
+ "trial_params": null
169
+ }
reproduction/glue_expHRA/cola/dr0.0,mlr9e-03,clr9e-03,ep=34.0t=24d12h04m24,sd43/CoLA.tsv ADDED
@@ -0,0 +1,1064 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ index prediction
2
+ 0 1
3
+ 1 1
4
+ 2 1
5
+ 3 1
6
+ 4 1
7
+ 5 1
8
+ 6 1
9
+ 7 1
10
+ 8 1
11
+ 9 1
12
+ 10 1
13
+ 11 1
14
+ 12 1
15
+ 13 1
16
+ 14 1
17
+ 15 1
18
+ 16 1
19
+ 17 1
20
+ 18 0
21
+ 19 1
22
+ 20 1
23
+ 21 1
24
+ 22 0
25
+ 23 1
26
+ 24 1
27
+ 25 1
28
+ 26 1
29
+ 27 1
30
+ 28 1
31
+ 29 1
32
+ 30 1
33
+ 31 0
34
+ 32 1
35
+ 33 1
36
+ 34 1
37
+ 35 1
38
+ 36 1
39
+ 37 1
40
+ 38 1
41
+ 39 1
42
+ 40 0
43
+ 41 1
44
+ 42 1
45
+ 43 1
46
+ 44 0
47
+ 45 1
48
+ 46 0
49
+ 47 1
50
+ 48 1
51
+ 49 1
52
+ 50 1
53
+ 51 1
54
+ 52 1
55
+ 53 1
56
+ 54 1
57
+ 55 1
58
+ 56 1
59
+ 57 1
60
+ 58 1
61
+ 59 1
62
+ 60 1
63
+ 61 1
64
+ 62 1
65
+ 63 1
66
+ 64 1
67
+ 65 1
68
+ 66 1
69
+ 67 1
70
+ 68 1
71
+ 69 1
72
+ 70 1
73
+ 71 1
74
+ 72 1
75
+ 73 1
76
+ 74 1
77
+ 75 1
78
+ 76 1
79
+ 77 1
80
+ 78 1
81
+ 79 1
82
+ 80 1
83
+ 81 0
84
+ 82 0
85
+ 83 1
86
+ 84 1
87
+ 85 1
88
+ 86 1
89
+ 87 1
90
+ 88 1
91
+ 89 1
92
+ 90 1
93
+ 91 1
94
+ 92 1
95
+ 93 1
96
+ 94 1
97
+ 95 1
98
+ 96 1
99
+ 97 1
100
+ 98 1
101
+ 99 1
102
+ 100 1
103
+ 101 1
104
+ 102 1
105
+ 103 1
106
+ 104 1
107
+ 105 1
108
+ 106 1
109
+ 107 1
110
+ 108 1
111
+ 109 1
112
+ 110 1
113
+ 111 1
114
+ 112 1
115
+ 113 1
116
+ 114 1
117
+ 115 0
118
+ 116 1
119
+ 117 1
120
+ 118 1
121
+ 119 1
122
+ 120 1
123
+ 121 1
124
+ 122 1
125
+ 123 1
126
+ 124 1
127
+ 125 1
128
+ 126 1
129
+ 127 1
130
+ 128 1
131
+ 129 1
132
+ 130 1
133
+ 131 1
134
+ 132 1
135
+ 133 1
136
+ 134 1
137
+ 135 1
138
+ 136 1
139
+ 137 1
140
+ 138 0
141
+ 139 0
142
+ 140 1
143
+ 141 1
144
+ 142 1
145
+ 143 1
146
+ 144 1
147
+ 145 1
148
+ 146 1
149
+ 147 1
150
+ 148 1
151
+ 149 1
152
+ 150 1
153
+ 151 1
154
+ 152 1
155
+ 153 1
156
+ 154 1
157
+ 155 1
158
+ 156 0
159
+ 157 1
160
+ 158 0
161
+ 159 1
162
+ 160 1
163
+ 161 1
164
+ 162 1
165
+ 163 1
166
+ 164 1
167
+ 165 0
168
+ 166 1
169
+ 167 1
170
+ 168 1
171
+ 169 1
172
+ 170 1
173
+ 171 1
174
+ 172 1
175
+ 173 1
176
+ 174 1
177
+ 175 1
178
+ 176 1
179
+ 177 1
180
+ 178 1
181
+ 179 1
182
+ 180 1
183
+ 181 1
184
+ 182 1
185
+ 183 1
186
+ 184 1
187
+ 185 1
188
+ 186 1
189
+ 187 1
190
+ 188 1
191
+ 189 1
192
+ 190 1
193
+ 191 1
194
+ 192 0
195
+ 193 1
196
+ 194 1
197
+ 195 1
198
+ 196 1
199
+ 197 1
200
+ 198 1
201
+ 199 0
202
+ 200 1
203
+ 201 0
204
+ 202 1
205
+ 203 1
206
+ 204 1
207
+ 205 1
208
+ 206 1
209
+ 207 1
210
+ 208 1
211
+ 209 1
212
+ 210 1
213
+ 211 1
214
+ 212 1
215
+ 213 1
216
+ 214 1
217
+ 215 1
218
+ 216 1
219
+ 217 1
220
+ 218 1
221
+ 219 1
222
+ 220 1
223
+ 221 1
224
+ 222 1
225
+ 223 1
226
+ 224 1
227
+ 225 1
228
+ 226 1
229
+ 227 1
230
+ 228 1
231
+ 229 1
232
+ 230 1
233
+ 231 1
234
+ 232 1
235
+ 233 1
236
+ 234 1
237
+ 235 1
238
+ 236 1
239
+ 237 1
240
+ 238 1
241
+ 239 1
242
+ 240 1
243
+ 241 1
244
+ 242 1
245
+ 243 1
246
+ 244 1
247
+ 245 1
248
+ 246 1
249
+ 247 1
250
+ 248 1
251
+ 249 1
252
+ 250 1
253
+ 251 1
254
+ 252 1
255
+ 253 1
256
+ 254 1
257
+ 255 1
258
+ 256 1
259
+ 257 1
260
+ 258 1
261
+ 259 1
262
+ 260 1
263
+ 261 1
264
+ 262 1
265
+ 263 1
266
+ 264 1
267
+ 265 1
268
+ 266 1
269
+ 267 1
270
+ 268 1
271
+ 269 1
272
+ 270 1
273
+ 271 1
274
+ 272 1
275
+ 273 1
276
+ 274 1
277
+ 275 1
278
+ 276 1
279
+ 277 1
280
+ 278 1
281
+ 279 0
282
+ 280 1
283
+ 281 0
284
+ 282 1
285
+ 283 1
286
+ 284 1
287
+ 285 0
288
+ 286 1
289
+ 287 1
290
+ 288 1
291
+ 289 1
292
+ 290 1
293
+ 291 1
294
+ 292 1
295
+ 293 1
296
+ 294 1
297
+ 295 1
298
+ 296 1
299
+ 297 1
300
+ 298 1
301
+ 299 1
302
+ 300 1
303
+ 301 0
304
+ 302 1
305
+ 303 1
306
+ 304 1
307
+ 305 1
308
+ 306 0
309
+ 307 1
310
+ 308 1
311
+ 309 1
312
+ 310 1
313
+ 311 1
314
+ 312 1
315
+ 313 1
316
+ 314 0
317
+ 315 0
318
+ 316 1
319
+ 317 1
320
+ 318 1
321
+ 319 1
322
+ 320 1
323
+ 321 1
324
+ 322 1
325
+ 323 1
326
+ 324 1
327
+ 325 1
328
+ 326 1
329
+ 327 0
330
+ 328 1
331
+ 329 1
332
+ 330 1
333
+ 331 1
334
+ 332 1
335
+ 333 1
336
+ 334 1
337
+ 335 1
338
+ 336 1
339
+ 337 1
340
+ 338 1
341
+ 339 1
342
+ 340 1
343
+ 341 1
344
+ 342 1
345
+ 343 1
346
+ 344 1
347
+ 345 1
348
+ 346 1
349
+ 347 1
350
+ 348 1
351
+ 349 1
352
+ 350 1
353
+ 351 1
354
+ 352 1
355
+ 353 1
356
+ 354 1
357
+ 355 1
358
+ 356 1
359
+ 357 1
360
+ 358 1
361
+ 359 1
362
+ 360 1
363
+ 361 1
364
+ 362 1
365
+ 363 1
366
+ 364 1
367
+ 365 1
368
+ 366 1
369
+ 367 1
370
+ 368 1
371
+ 369 1
372
+ 370 1
373
+ 371 1
374
+ 372 1
375
+ 373 1
376
+ 374 1
377
+ 375 1
378
+ 376 1
379
+ 377 1
380
+ 378 1
381
+ 379 1
382
+ 380 1
383
+ 381 1
384
+ 382 1
385
+ 383 1
386
+ 384 1
387
+ 385 1
388
+ 386 1
389
+ 387 1
390
+ 388 1
391
+ 389 0
392
+ 390 1
393
+ 391 1
394
+ 392 0
395
+ 393 1
396
+ 394 1
397
+ 395 1
398
+ 396 1
399
+ 397 1
400
+ 398 1
401
+ 399 1
402
+ 400 1
403
+ 401 1
404
+ 402 1
405
+ 403 1
406
+ 404 1
407
+ 405 1
408
+ 406 1
409
+ 407 1
410
+ 408 1
411
+ 409 1
412
+ 410 1
413
+ 411 1
414
+ 412 1
415
+ 413 1
416
+ 414 1
417
+ 415 1
418
+ 416 1
419
+ 417 1
420
+ 418 0
421
+ 419 1
422
+ 420 1
423
+ 421 1
424
+ 422 1
425
+ 423 1
426
+ 424 1
427
+ 425 1
428
+ 426 1
429
+ 427 1
430
+ 428 1
431
+ 429 1
432
+ 430 1
433
+ 431 1
434
+ 432 1
435
+ 433 1
436
+ 434 1
437
+ 435 1
438
+ 436 1
439
+ 437 1
440
+ 438 1
441
+ 439 1
442
+ 440 1
443
+ 441 1
444
+ 442 1
445
+ 443 1
446
+ 444 1
447
+ 445 1
448
+ 446 1
449
+ 447 1
450
+ 448 1
451
+ 449 1
452
+ 450 1
453
+ 451 1
454
+ 452 1
455
+ 453 1
456
+ 454 1
457
+ 455 1
458
+ 456 1
459
+ 457 1
460
+ 458 1
461
+ 459 0
462
+ 460 1
463
+ 461 1
464
+ 462 1
465
+ 463 1
466
+ 464 1
467
+ 465 1
468
+ 466 1
469
+ 467 1
470
+ 468 1
471
+ 469 1
472
+ 470 1
473
+ 471 1
474
+ 472 1
475
+ 473 0
476
+ 474 1
477
+ 475 0
478
+ 476 0
479
+ 477 1
480
+ 478 1
481
+ 479 1
482
+ 480 1
483
+ 481 1
484
+ 482 1
485
+ 483 1
486
+ 484 1
487
+ 485 1
488
+ 486 1
489
+ 487 1
490
+ 488 0
491
+ 489 0
492
+ 490 1
493
+ 491 1
494
+ 492 1
495
+ 493 1
496
+ 494 1
497
+ 495 1
498
+ 496 0
499
+ 497 1
500
+ 498 1
501
+ 499 1
502
+ 500 1
503
+ 501 1
504
+ 502 1
505
+ 503 1
506
+ 504 0
507
+ 505 1
508
+ 506 1
509
+ 507 1
510
+ 508 0
511
+ 509 1
512
+ 510 1
513
+ 511 0
514
+ 512 1
515
+ 513 1
516
+ 514 1
517
+ 515 1
518
+ 516 1
519
+ 517 1
520
+ 518 1
521
+ 519 1
522
+ 520 1
523
+ 521 0
524
+ 522 1
525
+ 523 1
526
+ 524 0
527
+ 525 1
528
+ 526 1
529
+ 527 1
530
+ 528 1
531
+ 529 1
532
+ 530 1
533
+ 531 1
534
+ 532 1
535
+ 533 1
536
+ 534 1
537
+ 535 1
538
+ 536 1
539
+ 537 1
540
+ 538 1
541
+ 539 1
542
+ 540 1
543
+ 541 1
544
+ 542 1
545
+ 543 1
546
+ 544 1
547
+ 545 1
548
+ 546 1
549
+ 547 1
550
+ 548 1
551
+ 549 1
552
+ 550 1
553
+ 551 1
554
+ 552 1
555
+ 553 1
556
+ 554 1
557
+ 555 1
558
+ 556 1
559
+ 557 1
560
+ 558 1
561
+ 559 1
562
+ 560 1
563
+ 561 1
564
+ 562 1
565
+ 563 1
566
+ 564 1
567
+ 565 1
568
+ 566 1
569
+ 567 1
570
+ 568 1
571
+ 569 1
572
+ 570 1
573
+ 571 1
574
+ 572 1
575
+ 573 1
576
+ 574 1
577
+ 575 1
578
+ 576 1
579
+ 577 1
580
+ 578 1
581
+ 579 1
582
+ 580 1
583
+ 581 1
584
+ 582 1
585
+ 583 1
586
+ 584 1
587
+ 585 1
588
+ 586 1
589
+ 587 1
590
+ 588 1
591
+ 589 1
592
+ 590 1
593
+ 591 1
594
+ 592 1
595
+ 593 1
596
+ 594 1
597
+ 595 1
598
+ 596 1
599
+ 597 1
600
+ 598 1
601
+ 599 1
602
+ 600 1
603
+ 601 1
604
+ 602 1
605
+ 603 1
606
+ 604 0
607
+ 605 1
608
+ 606 1
609
+ 607 1
610
+ 608 1
611
+ 609 1
612
+ 610 1
613
+ 611 1
614
+ 612 1
615
+ 613 1
616
+ 614 1
617
+ 615 1
618
+ 616 1
619
+ 617 1
620
+ 618 1
621
+ 619 1
622
+ 620 1
623
+ 621 1
624
+ 622 1
625
+ 623 1
626
+ 624 1
627
+ 625 1
628
+ 626 1
629
+ 627 1
630
+ 628 1
631
+ 629 1
632
+ 630 1
633
+ 631 1
634
+ 632 1
635
+ 633 1
636
+ 634 1
637
+ 635 0
638
+ 636 0
639
+ 637 1
640
+ 638 1
641
+ 639 1
642
+ 640 0
643
+ 641 1
644
+ 642 0
645
+ 643 1
646
+ 644 1
647
+ 645 1
648
+ 646 1
649
+ 647 1
650
+ 648 1
651
+ 649 1
652
+ 650 1
653
+ 651 1
654
+ 652 1
655
+ 653 1
656
+ 654 1
657
+ 655 1
658
+ 656 1
659
+ 657 1
660
+ 658 1
661
+ 659 1
662
+ 660 1
663
+ 661 1
664
+ 662 1
665
+ 663 1
666
+ 664 1
667
+ 665 1
668
+ 666 1
669
+ 667 1
670
+ 668 1
671
+ 669 1
672
+ 670 1
673
+ 671 1
674
+ 672 1
675
+ 673 1
676
+ 674 1
677
+ 675 1
678
+ 676 1
679
+ 677 1
680
+ 678 1
681
+ 679 1
682
+ 680 1
683
+ 681 1
684
+ 682 1
685
+ 683 1
686
+ 684 1
687
+ 685 1
688
+ 686 1
689
+ 687 1
690
+ 688 1
691
+ 689 1
692
+ 690 1
693
+ 691 1
694
+ 692 1
695
+ 693 1
696
+ 694 0
697
+ 695 0
698
+ 696 1
699
+ 697 1
700
+ 698 1
701
+ 699 1
702
+ 700 1
703
+ 701 0
704
+ 702 0
705
+ 703 1
706
+ 704 1
707
+ 705 1
708
+ 706 1
709
+ 707 1
710
+ 708 1
711
+ 709 1
712
+ 710 1
713
+ 711 1
714
+ 712 1
715
+ 713 1
716
+ 714 1
717
+ 715 0
718
+ 716 0
719
+ 717 0
720
+ 718 0
721
+ 719 1
722
+ 720 1
723
+ 721 1
724
+ 722 1
725
+ 723 1
726
+ 724 1
727
+ 725 1
728
+ 726 1
729
+ 727 1
730
+ 728 1
731
+ 729 1
732
+ 730 1
733
+ 731 1
734
+ 732 1
735
+ 733 1
736
+ 734 1
737
+ 735 1
738
+ 736 1
739
+ 737 1
740
+ 738 1
741
+ 739 1
742
+ 740 1
743
+ 741 1
744
+ 742 1
745
+ 743 1
746
+ 744 1
747
+ 745 1
748
+ 746 1
749
+ 747 1
750
+ 748 1
751
+ 749 1
752
+ 750 1
753
+ 751 1
754
+ 752 1
755
+ 753 1
756
+ 754 1
757
+ 755 0
758
+ 756 1
759
+ 757 1
760
+ 758 1
761
+ 759 1
762
+ 760 1
763
+ 761 1
764
+ 762 1
765
+ 763 1
766
+ 764 1
767
+ 765 1
768
+ 766 1
769
+ 767 1
770
+ 768 1
771
+ 769 1
772
+ 770 1
773
+ 771 1
774
+ 772 1
775
+ 773 1
776
+ 774 1
777
+ 775 0
778
+ 776 1
779
+ 777 1
780
+ 778 1
781
+ 779 1
782
+ 780 1
783
+ 781 1
784
+ 782 1
785
+ 783 1
786
+ 784 1
787
+ 785 1
788
+ 786 1
789
+ 787 1
790
+ 788 1
791
+ 789 1
792
+ 790 1
793
+ 791 1
794
+ 792 1
795
+ 793 1
796
+ 794 1
797
+ 795 1
798
+ 796 1
799
+ 797 1
800
+ 798 0
801
+ 799 0
802
+ 800 1
803
+ 801 1
804
+ 802 1
805
+ 803 1
806
+ 804 1
807
+ 805 1
808
+ 806 1
809
+ 807 1
810
+ 808 1
811
+ 809 1
812
+ 810 1
813
+ 811 1
814
+ 812 1
815
+ 813 1
816
+ 814 1
817
+ 815 0
818
+ 816 1
819
+ 817 1
820
+ 818 1
821
+ 819 1
822
+ 820 1
823
+ 821 1
824
+ 822 0
825
+ 823 0
826
+ 824 1
827
+ 825 1
828
+ 826 0
829
+ 827 1
830
+ 828 1
831
+ 829 1
832
+ 830 0
833
+ 831 0
834
+ 832 0
835
+ 833 1
836
+ 834 1
837
+ 835 1
838
+ 836 1
839
+ 837 1
840
+ 838 1
841
+ 839 1
842
+ 840 1
843
+ 841 1
844
+ 842 1
845
+ 843 1
846
+ 844 1
847
+ 845 1
848
+ 846 1
849
+ 847 1
850
+ 848 1
851
+ 849 1
852
+ 850 1
853
+ 851 1
854
+ 852 1
855
+ 853 1
856
+ 854 1
857
+ 855 1
858
+ 856 1
859
+ 857 1
860
+ 858 0
861
+ 859 1
862
+ 860 1
863
+ 861 1
864
+ 862 1
865
+ 863 1
866
+ 864 1
867
+ 865 1
868
+ 866 1
869
+ 867 1
870
+ 868 1
871
+ 869 1
872
+ 870 1
873
+ 871 1
874
+ 872 1
875
+ 873 1
876
+ 874 1
877
+ 875 1
878
+ 876 1
879
+ 877 1
880
+ 878 1
881
+ 879 1
882
+ 880 1
883
+ 881 1
884
+ 882 1
885
+ 883 1
886
+ 884 1
887
+ 885 0
888
+ 886 0
889
+ 887 0
890
+ 888 0
891
+ 889 0
892
+ 890 0
893
+ 891 1
894
+ 892 0
895
+ 893 1
896
+ 894 1
897
+ 895 1
898
+ 896 1
899
+ 897 0
900
+ 898 0
901
+ 899 1
902
+ 900 1
903
+ 901 1
904
+ 902 1
905
+ 903 1
906
+ 904 0
907
+ 905 1
908
+ 906 1
909
+ 907 1
910
+ 908 1
911
+ 909 1
912
+ 910 1
913
+ 911 1
914
+ 912 1
915
+ 913 1
916
+ 914 1
917
+ 915 1
918
+ 916 1
919
+ 917 1
920
+ 918 1
921
+ 919 1
922
+ 920 1
923
+ 921 1
924
+ 922 1
925
+ 923 1
926
+ 924 1
927
+ 925 1
928
+ 926 1
929
+ 927 1
930
+ 928 1
931
+ 929 1
932
+ 930 0
933
+ 931 0
934
+ 932 1
935
+ 933 1
936
+ 934 1
937
+ 935 1
938
+ 936 1
939
+ 937 0
940
+ 938 0
941
+ 939 1
942
+ 940 1
943
+ 941 1
944
+ 942 1
945
+ 943 1
946
+ 944 1
947
+ 945 1
948
+ 946 1
949
+ 947 1
950
+ 948 0
951
+ 949 1
952
+ 950 1
953
+ 951 1
954
+ 952 1
955
+ 953 1
956
+ 954 1
957
+ 955 0
958
+ 956 1
959
+ 957 1
960
+ 958 1
961
+ 959 1
962
+ 960 1
963
+ 961 1
964
+ 962 1
965
+ 963 1
966
+ 964 1
967
+ 965 1
968
+ 966 1
969
+ 967 1
970
+ 968 1
971
+ 969 1
972
+ 970 1
973
+ 971 1
974
+ 972 1
975
+ 973 1
976
+ 974 1
977
+ 975 1
978
+ 976 1
979
+ 977 1
980
+ 978 1
981
+ 979 1
982
+ 980 1
983
+ 981 1
984
+ 982 0
985
+ 983 1
986
+ 984 1
987
+ 985 1
988
+ 986 1
989
+ 987 1
990
+ 988 1
991
+ 989 1
992
+ 990 1
993
+ 991 1
994
+ 992 1
995
+ 993 1
996
+ 994 1
997
+ 995 1
998
+ 996 1
999
+ 997 1
1000
+ 998 1
1001
+ 999 1
1002
+ 1000 1
1003
+ 1001 1
1004
+ 1002 1
1005
+ 1003 1
1006
+ 1004 0
1007
+ 1005 1
1008
+ 1006 1
1009
+ 1007 0
1010
+ 1008 1
1011
+ 1009 1
1012
+ 1010 1
1013
+ 1011 1
1014
+ 1012 1
1015
+ 1013 1
1016
+ 1014 1
1017
+ 1015 1
1018
+ 1016 1
1019
+ 1017 1
1020
+ 1018 1
1021
+ 1019 1
1022
+ 1020 1
1023
+ 1021 1
1024
+ 1022 1
1025
+ 1023 1
1026
+ 1024 1
1027
+ 1025 1
1028
+ 1026 1
1029
+ 1027 1
1030
+ 1028 1
1031
+ 1029 1
1032
+ 1030 1
1033
+ 1031 1
1034
+ 1032 1
1035
+ 1033 1
1036
+ 1034 1
1037
+ 1035 0
1038
+ 1036 1
1039
+ 1037 1
1040
+ 1038 1
1041
+ 1039 1
1042
+ 1040 1
1043
+ 1041 1
1044
+ 1042 1
1045
+ 1043 1
1046
+ 1044 1
1047
+ 1045 1
1048
+ 1046 1
1049
+ 1047 1
1050
+ 1048 1
1051
+ 1049 0
1052
+ 1050 1
1053
+ 1051 1
1054
+ 1052 1
1055
+ 1053 1
1056
+ 1054 1
1057
+ 1055 1
1058
+ 1056 1
1059
+ 1057 1
1060
+ 1058 1
1061
+ 1059 1
1062
+ 1060 1
1063
+ 1061 1
1064
+ 1062 1
reproduction/glue_expHRA/cola/dr0.0,mlr9e-03,clr9e-03,ep=34.0t=24d12h04m24,sd43/all_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 34.0,
3
+ "eval_loss": 1.082499384880066,
4
+ "eval_matthews_correlation": 0.404833785050489,
5
+ "eval_runtime": 0.476,
6
+ "eval_samples": 1043,
7
+ "eval_samples_per_second": 2191.08,
8
+ "eval_steps_per_second": 6.302
9
+ }
reproduction/glue_expHRA/cola/dr0.0,mlr9e-03,clr9e-03,ep=34.0t=24d12h04m24,sd43/eval_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 34.0,
3
+ "eval_loss": 1.082499384880066,
4
+ "eval_matthews_correlation": 0.404833785050489,
5
+ "eval_runtime": 0.476,
6
+ "eval_samples": 1043,
7
+ "eval_samples_per_second": 2191.08,
8
+ "eval_steps_per_second": 6.302
9
+ }
reproduction/glue_expHRA/cola/dr0.0,mlr9e-03,clr9e-03,ep=34.0t=24d12h04m24,sd43/ft/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[MASK]": 128000
3
+ }
reproduction/glue_expHRA/cola/dr0.0,mlr9e-03,clr9e-03,ep=34.0t=24d12h04m24,sd43/ft/special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": {
9
+ "content": "[UNK]",
10
+ "lstrip": false,
11
+ "normalized": true,
12
+ "rstrip": false,
13
+ "single_word": false
14
+ }
15
+ }
reproduction/glue_expHRA/cola/dr0.0,mlr9e-03,clr9e-03,ep=34.0t=24d12h04m24,sd43/ft/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff