kbberendsen commited on
Commit
6416698
1 Parent(s): f4c38c7

Training in progress, epoch 1

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. model.safetensors +1 -1
  2. run-2/checkpoint-214/added_tokens.json +3 -0
  3. run-2/checkpoint-214/config.json +35 -0
  4. run-2/checkpoint-214/model.safetensors +3 -0
  5. run-2/checkpoint-214/optimizer.pt +3 -0
  6. run-2/checkpoint-214/rng_state.pth +3 -0
  7. run-2/checkpoint-214/scheduler.pt +3 -0
  8. run-2/checkpoint-214/special_tokens_map.json +15 -0
  9. run-2/checkpoint-214/spm.model +3 -0
  10. run-2/checkpoint-214/tokenizer.json +0 -0
  11. run-2/checkpoint-214/tokenizer_config.json +58 -0
  12. run-2/checkpoint-214/trainer_state.json +44 -0
  13. run-2/checkpoint-214/training_args.bin +3 -0
  14. run-2/checkpoint-321/added_tokens.json +3 -0
  15. run-2/checkpoint-321/config.json +35 -0
  16. run-2/checkpoint-321/model.safetensors +3 -0
  17. run-2/checkpoint-321/optimizer.pt +3 -0
  18. run-2/checkpoint-321/rng_state.pth +3 -0
  19. run-2/checkpoint-321/scheduler.pt +3 -0
  20. run-2/checkpoint-321/special_tokens_map.json +15 -0
  21. run-2/checkpoint-321/spm.model +3 -0
  22. run-2/checkpoint-321/tokenizer.json +0 -0
  23. run-2/checkpoint-321/tokenizer_config.json +58 -0
  24. run-2/checkpoint-321/trainer_state.json +53 -0
  25. run-2/checkpoint-321/training_args.bin +3 -0
  26. run-2/checkpoint-428/added_tokens.json +3 -0
  27. run-2/checkpoint-428/config.json +35 -0
  28. run-2/checkpoint-428/model.safetensors +3 -0
  29. run-2/checkpoint-428/optimizer.pt +3 -0
  30. run-2/checkpoint-428/rng_state.pth +3 -0
  31. run-2/checkpoint-428/scheduler.pt +3 -0
  32. run-2/checkpoint-428/special_tokens_map.json +15 -0
  33. run-2/checkpoint-428/spm.model +3 -0
  34. run-2/checkpoint-428/tokenizer.json +0 -0
  35. run-2/checkpoint-428/tokenizer_config.json +58 -0
  36. run-2/checkpoint-428/trainer_state.json +62 -0
  37. run-2/checkpoint-428/training_args.bin +3 -0
  38. run-3/checkpoint-214/added_tokens.json +3 -0
  39. run-3/checkpoint-214/config.json +35 -0
  40. run-3/checkpoint-214/model.safetensors +3 -0
  41. run-3/checkpoint-214/optimizer.pt +3 -0
  42. run-3/checkpoint-214/rng_state.pth +3 -0
  43. run-3/checkpoint-214/scheduler.pt +3 -0
  44. run-3/checkpoint-214/special_tokens_map.json +15 -0
  45. run-3/checkpoint-214/spm.model +3 -0
  46. run-3/checkpoint-214/tokenizer.json +0 -0
  47. run-3/checkpoint-214/tokenizer_config.json +58 -0
  48. run-3/checkpoint-214/trainer_state.json +35 -0
  49. run-3/checkpoint-214/training_args.bin +3 -0
  50. runs/Feb29_09-36-41_12cf901772ff/events.out.tfevents.1709203875.12cf901772ff.264.7 +3 -0
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c791636b6c132216c0275031a093f9924ff93001c41c5f02825e25ebde0b15b8
3
  size 1740304440
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d27952fe13fd3bcf65eddbdd6237e7a718b346a95a0de852ae7d2509925e0aa
3
  size 1740304440
run-2/checkpoint-214/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[MASK]": 128000
3
+ }
run-2/checkpoint-214/config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/deberta-v3-large",
3
+ "architectures": [
4
+ "DebertaV2ForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "hidden_act": "gelu",
8
+ "hidden_dropout_prob": 0.1,
9
+ "hidden_size": 1024,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 4096,
12
+ "layer_norm_eps": 1e-07,
13
+ "max_position_embeddings": 512,
14
+ "max_relative_positions": -1,
15
+ "model_type": "deberta-v2",
16
+ "norm_rel_ebd": "layer_norm",
17
+ "num_attention_heads": 16,
18
+ "num_hidden_layers": 24,
19
+ "pad_token_id": 0,
20
+ "pooler_dropout": 0,
21
+ "pooler_hidden_act": "gelu",
22
+ "pooler_hidden_size": 1024,
23
+ "pos_att_type": [
24
+ "p2c",
25
+ "c2p"
26
+ ],
27
+ "position_biased_input": false,
28
+ "position_buckets": 256,
29
+ "relative_attention": true,
30
+ "share_att_key": true,
31
+ "torch_dtype": "float32",
32
+ "transformers_version": "4.38.1",
33
+ "type_vocab_size": 0,
34
+ "vocab_size": 128100
35
+ }
run-2/checkpoint-214/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53fa5fe62d6138692858e5b51ea10981b82eaaafb44d5581ad9936fb3546e0b3
3
+ size 1740304440
run-2/checkpoint-214/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8d95e4bbe9af7ec874e9588be52f9297731569408c9b84e5e9a2e061c43a1fa
3
+ size 3480840240
run-2/checkpoint-214/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a34aa86e8f0fe31fcba9fa0444447f0da964b0fde3dfd03354626c8761f7e3b5
3
+ size 14244
run-2/checkpoint-214/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd2f3b407f48550af799c5ada8053a601ad4f5c588e8abfcb09a5a35f99790be
3
+ size 1064
run-2/checkpoint-214/special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": {
9
+ "content": "[UNK]",
10
+ "lstrip": false,
11
+ "normalized": true,
12
+ "rstrip": false,
13
+ "single_word": false
14
+ }
15
+ }
run-2/checkpoint-214/spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
3
+ size 2464616
run-2/checkpoint-214/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
run-2/checkpoint-214/tokenizer_config.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[CLS]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[SEP]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[UNK]",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "128000": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "[CLS]",
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "[CLS]",
47
+ "do_lower_case": false,
48
+ "eos_token": "[SEP]",
49
+ "mask_token": "[MASK]",
50
+ "model_max_length": 1000000000000000019884624838656,
51
+ "pad_token": "[PAD]",
52
+ "sep_token": "[SEP]",
53
+ "sp_model_kwargs": {},
54
+ "split_by_punct": false,
55
+ "tokenizer_class": "DebertaV2Tokenizer",
56
+ "unk_token": "[UNK]",
57
+ "vocab_type": "spm"
58
+ }
run-2/checkpoint-214/trainer_state.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.6454833331679224,
3
+ "best_model_checkpoint": "deberta-v3-large-finetuned-cola-midterm/run-2/checkpoint-214",
4
+ "epoch": 2.0,
5
+ "eval_steps": 500,
6
+ "global_step": 214,
7
+ "is_hyper_param_search": true,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "eval_loss": 0.470249742269516,
14
+ "eval_matthews_correlation": 0.37472354614099185,
15
+ "eval_runtime": 8.4859,
16
+ "eval_samples_per_second": 122.91,
17
+ "eval_steps_per_second": 7.778,
18
+ "step": 107
19
+ },
20
+ {
21
+ "epoch": 2.0,
22
+ "eval_loss": 0.37880179286003113,
23
+ "eval_matthews_correlation": 0.6454833331679224,
24
+ "eval_runtime": 8.7227,
25
+ "eval_samples_per_second": 119.574,
26
+ "eval_steps_per_second": 7.566,
27
+ "step": 214
28
+ }
29
+ ],
30
+ "logging_steps": 500,
31
+ "max_steps": 428,
32
+ "num_input_tokens_seen": 0,
33
+ "num_train_epochs": 4,
34
+ "save_steps": 500,
35
+ "total_flos": 0,
36
+ "train_batch_size": 8,
37
+ "trial_name": null,
38
+ "trial_params": {
39
+ "learning_rate": 5.723917767775046e-06,
40
+ "num_train_epochs": 4,
41
+ "per_device_train_batch_size": 8,
42
+ "seed": 4
43
+ }
44
+ }
run-2/checkpoint-214/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d01347cde26c18eec32f6001c04a9e73a3c5a8dee929c3d73a85517dd5970f15
3
+ size 4984
run-2/checkpoint-321/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[MASK]": 128000
3
+ }
run-2/checkpoint-321/config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/deberta-v3-large",
3
+ "architectures": [
4
+ "DebertaV2ForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "hidden_act": "gelu",
8
+ "hidden_dropout_prob": 0.1,
9
+ "hidden_size": 1024,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 4096,
12
+ "layer_norm_eps": 1e-07,
13
+ "max_position_embeddings": 512,
14
+ "max_relative_positions": -1,
15
+ "model_type": "deberta-v2",
16
+ "norm_rel_ebd": "layer_norm",
17
+ "num_attention_heads": 16,
18
+ "num_hidden_layers": 24,
19
+ "pad_token_id": 0,
20
+ "pooler_dropout": 0,
21
+ "pooler_hidden_act": "gelu",
22
+ "pooler_hidden_size": 1024,
23
+ "pos_att_type": [
24
+ "p2c",
25
+ "c2p"
26
+ ],
27
+ "position_biased_input": false,
28
+ "position_buckets": 256,
29
+ "relative_attention": true,
30
+ "share_att_key": true,
31
+ "torch_dtype": "float32",
32
+ "transformers_version": "4.38.1",
33
+ "type_vocab_size": 0,
34
+ "vocab_size": 128100
35
+ }
run-2/checkpoint-321/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e856fc960310e880cf45fe0b91d2fe8413c7a96ee7fef4118ce4ffe8696c7318
3
+ size 1740304440
run-2/checkpoint-321/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57a651e2d86f25b1ee4540390ccddb54214cd64ac37a3444971635c026759eaf
3
+ size 3480840240
run-2/checkpoint-321/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41997ee9eaad95a796779b0306403fa7810426cfabd7e7650c5a8e4432dabae1
3
+ size 14244
run-2/checkpoint-321/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f65d643147a4d53c322ee12a9f11b90cb6add2319700ce1f04b1a7c7d39b123
3
+ size 1064
run-2/checkpoint-321/special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": {
9
+ "content": "[UNK]",
10
+ "lstrip": false,
11
+ "normalized": true,
12
+ "rstrip": false,
13
+ "single_word": false
14
+ }
15
+ }
run-2/checkpoint-321/spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
3
+ size 2464616
run-2/checkpoint-321/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
run-2/checkpoint-321/tokenizer_config.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[CLS]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[SEP]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[UNK]",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "128000": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "[CLS]",
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "[CLS]",
47
+ "do_lower_case": false,
48
+ "eos_token": "[SEP]",
49
+ "mask_token": "[MASK]",
50
+ "model_max_length": 1000000000000000019884624838656,
51
+ "pad_token": "[PAD]",
52
+ "sep_token": "[SEP]",
53
+ "sp_model_kwargs": {},
54
+ "split_by_punct": false,
55
+ "tokenizer_class": "DebertaV2Tokenizer",
56
+ "unk_token": "[UNK]",
57
+ "vocab_type": "spm"
58
+ }
run-2/checkpoint-321/trainer_state.json ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.663247201911855,
3
+ "best_model_checkpoint": "deberta-v3-large-finetuned-cola-midterm/run-2/checkpoint-321",
4
+ "epoch": 3.0,
5
+ "eval_steps": 500,
6
+ "global_step": 321,
7
+ "is_hyper_param_search": true,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "eval_loss": 0.470249742269516,
14
+ "eval_matthews_correlation": 0.37472354614099185,
15
+ "eval_runtime": 8.4859,
16
+ "eval_samples_per_second": 122.91,
17
+ "eval_steps_per_second": 7.778,
18
+ "step": 107
19
+ },
20
+ {
21
+ "epoch": 2.0,
22
+ "eval_loss": 0.37880179286003113,
23
+ "eval_matthews_correlation": 0.6454833331679224,
24
+ "eval_runtime": 8.7227,
25
+ "eval_samples_per_second": 119.574,
26
+ "eval_steps_per_second": 7.566,
27
+ "step": 214
28
+ },
29
+ {
30
+ "epoch": 3.0,
31
+ "eval_loss": 0.4730988144874573,
32
+ "eval_matthews_correlation": 0.663247201911855,
33
+ "eval_runtime": 8.9367,
34
+ "eval_samples_per_second": 116.71,
35
+ "eval_steps_per_second": 7.385,
36
+ "step": 321
37
+ }
38
+ ],
39
+ "logging_steps": 500,
40
+ "max_steps": 428,
41
+ "num_input_tokens_seen": 0,
42
+ "num_train_epochs": 4,
43
+ "save_steps": 500,
44
+ "total_flos": 0,
45
+ "train_batch_size": 8,
46
+ "trial_name": null,
47
+ "trial_params": {
48
+ "learning_rate": 5.723917767775046e-06,
49
+ "num_train_epochs": 4,
50
+ "per_device_train_batch_size": 8,
51
+ "seed": 4
52
+ }
53
+ }
run-2/checkpoint-321/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d01347cde26c18eec32f6001c04a9e73a3c5a8dee929c3d73a85517dd5970f15
3
+ size 4984
run-2/checkpoint-428/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[MASK]": 128000
3
+ }
run-2/checkpoint-428/config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/deberta-v3-large",
3
+ "architectures": [
4
+ "DebertaV2ForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "hidden_act": "gelu",
8
+ "hidden_dropout_prob": 0.1,
9
+ "hidden_size": 1024,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 4096,
12
+ "layer_norm_eps": 1e-07,
13
+ "max_position_embeddings": 512,
14
+ "max_relative_positions": -1,
15
+ "model_type": "deberta-v2",
16
+ "norm_rel_ebd": "layer_norm",
17
+ "num_attention_heads": 16,
18
+ "num_hidden_layers": 24,
19
+ "pad_token_id": 0,
20
+ "pooler_dropout": 0,
21
+ "pooler_hidden_act": "gelu",
22
+ "pooler_hidden_size": 1024,
23
+ "pos_att_type": [
24
+ "p2c",
25
+ "c2p"
26
+ ],
27
+ "position_biased_input": false,
28
+ "position_buckets": 256,
29
+ "relative_attention": true,
30
+ "share_att_key": true,
31
+ "torch_dtype": "float32",
32
+ "transformers_version": "4.38.1",
33
+ "type_vocab_size": 0,
34
+ "vocab_size": 128100
35
+ }
run-2/checkpoint-428/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ebf3f5ef40a128cd0ad2588c8879c6afa84b67b538ce7fa971fc42763cc606d8
3
+ size 1740304440
run-2/checkpoint-428/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d40894c4308a76a76657abd0d3422a04395d24bbc759a86e99b038df67633d27
3
+ size 3480840240
run-2/checkpoint-428/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:853771c55d90e1532788a1a5bb49146cd257e58253f6177a8778555254fc4e85
3
+ size 14244
run-2/checkpoint-428/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a3b58620b6fc29cba5046b26293ceeb026f72b2f17bf1b8be9bb1439c652741
3
+ size 1064
run-2/checkpoint-428/special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": {
9
+ "content": "[UNK]",
10
+ "lstrip": false,
11
+ "normalized": true,
12
+ "rstrip": false,
13
+ "single_word": false
14
+ }
15
+ }
run-2/checkpoint-428/spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
3
+ size 2464616
run-2/checkpoint-428/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
run-2/checkpoint-428/tokenizer_config.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[CLS]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[SEP]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[UNK]",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "128000": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "[CLS]",
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "[CLS]",
47
+ "do_lower_case": false,
48
+ "eos_token": "[SEP]",
49
+ "mask_token": "[MASK]",
50
+ "model_max_length": 1000000000000000019884624838656,
51
+ "pad_token": "[PAD]",
52
+ "sep_token": "[SEP]",
53
+ "sp_model_kwargs": {},
54
+ "split_by_punct": false,
55
+ "tokenizer_class": "DebertaV2Tokenizer",
56
+ "unk_token": "[UNK]",
57
+ "vocab_type": "spm"
58
+ }
run-2/checkpoint-428/trainer_state.json ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.663247201911855,
3
+ "best_model_checkpoint": "deberta-v3-large-finetuned-cola-midterm/run-2/checkpoint-321",
4
+ "epoch": 4.0,
5
+ "eval_steps": 500,
6
+ "global_step": 428,
7
+ "is_hyper_param_search": true,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "eval_loss": 0.470249742269516,
14
+ "eval_matthews_correlation": 0.37472354614099185,
15
+ "eval_runtime": 8.4859,
16
+ "eval_samples_per_second": 122.91,
17
+ "eval_steps_per_second": 7.778,
18
+ "step": 107
19
+ },
20
+ {
21
+ "epoch": 2.0,
22
+ "eval_loss": 0.37880179286003113,
23
+ "eval_matthews_correlation": 0.6454833331679224,
24
+ "eval_runtime": 8.7227,
25
+ "eval_samples_per_second": 119.574,
26
+ "eval_steps_per_second": 7.566,
27
+ "step": 214
28
+ },
29
+ {
30
+ "epoch": 3.0,
31
+ "eval_loss": 0.4730988144874573,
32
+ "eval_matthews_correlation": 0.663247201911855,
33
+ "eval_runtime": 8.9367,
34
+ "eval_samples_per_second": 116.71,
35
+ "eval_steps_per_second": 7.385,
36
+ "step": 321
37
+ },
38
+ {
39
+ "epoch": 4.0,
40
+ "eval_loss": 0.5757762789726257,
41
+ "eval_matthews_correlation": 0.6431581256935667,
42
+ "eval_runtime": 8.8924,
43
+ "eval_samples_per_second": 117.291,
44
+ "eval_steps_per_second": 7.422,
45
+ "step": 428
46
+ }
47
+ ],
48
+ "logging_steps": 500,
49
+ "max_steps": 428,
50
+ "num_input_tokens_seen": 0,
51
+ "num_train_epochs": 4,
52
+ "save_steps": 500,
53
+ "total_flos": 0,
54
+ "train_batch_size": 8,
55
+ "trial_name": null,
56
+ "trial_params": {
57
+ "learning_rate": 5.723917767775046e-06,
58
+ "num_train_epochs": 4,
59
+ "per_device_train_batch_size": 8,
60
+ "seed": 4
61
+ }
62
+ }
run-2/checkpoint-428/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d01347cde26c18eec32f6001c04a9e73a3c5a8dee929c3d73a85517dd5970f15
3
+ size 4984
run-3/checkpoint-214/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[MASK]": 128000
3
+ }
run-3/checkpoint-214/config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/deberta-v3-large",
3
+ "architectures": [
4
+ "DebertaV2ForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "hidden_act": "gelu",
8
+ "hidden_dropout_prob": 0.1,
9
+ "hidden_size": 1024,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 4096,
12
+ "layer_norm_eps": 1e-07,
13
+ "max_position_embeddings": 512,
14
+ "max_relative_positions": -1,
15
+ "model_type": "deberta-v2",
16
+ "norm_rel_ebd": "layer_norm",
17
+ "num_attention_heads": 16,
18
+ "num_hidden_layers": 24,
19
+ "pad_token_id": 0,
20
+ "pooler_dropout": 0,
21
+ "pooler_hidden_act": "gelu",
22
+ "pooler_hidden_size": 1024,
23
+ "pos_att_type": [
24
+ "p2c",
25
+ "c2p"
26
+ ],
27
+ "position_biased_input": false,
28
+ "position_buckets": 256,
29
+ "relative_attention": true,
30
+ "share_att_key": true,
31
+ "torch_dtype": "float32",
32
+ "transformers_version": "4.38.1",
33
+ "type_vocab_size": 0,
34
+ "vocab_size": 128100
35
+ }
run-3/checkpoint-214/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d27952fe13fd3bcf65eddbdd6237e7a718b346a95a0de852ae7d2509925e0aa
3
+ size 1740304440
run-3/checkpoint-214/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f1b392844d34364cf517d304f2f8ede0d41cbfe03e105dd4e7bc22c0a82bb2d
3
+ size 3480840240
run-3/checkpoint-214/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eba8b90ad83c8bffd6ed3662c62cb274d4b1e36137dc5a9829e248e1510ba041
3
+ size 14244
run-3/checkpoint-214/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c09e07b18f4e82605534523f9c123618ead44937f8d50cc756b42bfc43543a48
3
+ size 1064
run-3/checkpoint-214/special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": {
9
+ "content": "[UNK]",
10
+ "lstrip": false,
11
+ "normalized": true,
12
+ "rstrip": false,
13
+ "single_word": false
14
+ }
15
+ }
run-3/checkpoint-214/spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
3
+ size 2464616
run-3/checkpoint-214/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
run-3/checkpoint-214/tokenizer_config.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[CLS]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[SEP]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[UNK]",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "128000": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "[CLS]",
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "[CLS]",
47
+ "do_lower_case": false,
48
+ "eos_token": "[SEP]",
49
+ "mask_token": "[MASK]",
50
+ "model_max_length": 1000000000000000019884624838656,
51
+ "pad_token": "[PAD]",
52
+ "sep_token": "[SEP]",
53
+ "sp_model_kwargs": {},
54
+ "split_by_punct": false,
55
+ "tokenizer_class": "DebertaV2Tokenizer",
56
+ "unk_token": "[UNK]",
57
+ "vocab_type": "spm"
58
+ }
run-3/checkpoint-214/trainer_state.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.0,
3
+ "best_model_checkpoint": "deberta-v3-large-finetuned-cola-midterm/run-3/checkpoint-214",
4
+ "epoch": 1.0,
5
+ "eval_steps": 500,
6
+ "global_step": 214,
7
+ "is_hyper_param_search": true,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "eval_loss": 0.6239587664604187,
14
+ "eval_matthews_correlation": 0.0,
15
+ "eval_runtime": 8.4185,
16
+ "eval_samples_per_second": 123.894,
17
+ "eval_steps_per_second": 7.84,
18
+ "step": 214
19
+ }
20
+ ],
21
+ "logging_steps": 500,
22
+ "max_steps": 428,
23
+ "num_input_tokens_seen": 0,
24
+ "num_train_epochs": 2,
25
+ "save_steps": 500,
26
+ "total_flos": 0,
27
+ "train_batch_size": 4,
28
+ "trial_name": null,
29
+ "trial_params": {
30
+ "learning_rate": 6.93139822379791e-05,
31
+ "num_train_epochs": 2,
32
+ "per_device_train_batch_size": 4,
33
+ "seed": 12
34
+ }
35
+ }
run-3/checkpoint-214/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:365374e711b9f598c24804c684864116b73578314660a35beaf262cdfdbadba4
3
+ size 4984
runs/Feb29_09-36-41_12cf901772ff/events.out.tfevents.1709203875.12cf901772ff.264.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46940e82564c51467b7e3579e2ec0951a5aa802b9bc668155e2af26235dca1a8
3
+ size 5831