jtatman commited on
Commit
2a94c77
·
1 Parent(s): 54bc0b3

End of training

Browse files
README.md CHANGED
@@ -17,7 +17,7 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [vicgalle/gpt2-open-instruct-v1](https://huggingface.co/vicgalle/gpt2-open-instruct-v1) on the gsm8k dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 2.4454
21
 
22
  ## Model description
23
 
@@ -37,23 +37,38 @@ More information needed
37
 
38
  The following hyperparameters were used during training:
39
  - learning_rate: 0.0005
40
- - train_batch_size: 6
41
  - eval_batch_size: 1
42
  - seed: 42
43
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
  - lr_scheduler_type: linear
45
  - lr_scheduler_warmup_steps: 100
46
- - num_epochs: 5
47
 
48
  ### Training results
49
 
50
  | Training Loss | Epoch | Step | Validation Loss |
51
  |:-------------:|:-----:|:----:|:---------------:|
52
- | 2.6943 | 1.0 | 1246 | 2.5155 |
53
- | 2.62 | 2.0 | 2492 | 2.4769 |
54
- | 2.5851 | 3.0 | 3738 | 2.4575 |
55
- | 2.5692 | 4.0 | 4984 | 2.4492 |
56
- | 2.5485 | 5.0 | 6230 | 2.4454 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
57
 
58
 
59
  ### Framework versions
 
17
 
18
  This model is a fine-tuned version of [vicgalle/gpt2-open-instruct-v1](https://huggingface.co/vicgalle/gpt2-open-instruct-v1) on the gsm8k dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 2.3966
21
 
22
  ## Model description
23
 
 
37
 
38
  The following hyperparameters were used during training:
39
  - learning_rate: 0.0005
40
+ - train_batch_size: 16
41
  - eval_batch_size: 1
42
  - seed: 42
43
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
  - lr_scheduler_type: linear
45
  - lr_scheduler_warmup_steps: 100
46
+ - num_epochs: 20
47
 
48
  ### Training results
49
 
50
  | Training Loss | Epoch | Step | Validation Loss |
51
  |:-------------:|:-----:|:----:|:---------------:|
52
+ | No log | 1.0 | 468 | 2.5579 |
53
+ | 2.859 | 2.0 | 936 | 2.5018 |
54
+ | 2.6455 | 3.0 | 1404 | 2.4752 |
55
+ | 2.6025 | 4.0 | 1872 | 2.4590 |
56
+ | 2.5777 | 5.0 | 2340 | 2.4473 |
57
+ | 2.5557 | 6.0 | 2808 | 2.4388 |
58
+ | 2.538 | 7.0 | 3276 | 2.4309 |
59
+ | 2.5246 | 8.0 | 3744 | 2.4236 |
60
+ | 2.514 | 9.0 | 4212 | 2.4186 |
61
+ | 2.5059 | 10.0 | 4680 | 2.4159 |
62
+ | 2.4944 | 11.0 | 5148 | 2.4107 |
63
+ | 2.4874 | 12.0 | 5616 | 2.4078 |
64
+ | 2.4862 | 13.0 | 6084 | 2.4053 |
65
+ | 2.475 | 14.0 | 6552 | 2.4027 |
66
+ | 2.4716 | 15.0 | 7020 | 2.4008 |
67
+ | 2.4716 | 16.0 | 7488 | 2.3995 |
68
+ | 2.4704 | 17.0 | 7956 | 2.3985 |
69
+ | 2.4648 | 18.0 | 8424 | 2.3973 |
70
+ | 2.4634 | 19.0 | 8892 | 2.3968 |
71
+ | 2.459 | 20.0 | 9360 | 2.3966 |
72
 
73
 
74
  ### Framework versions
checkpoint-4984/added_tokens.json DELETED
@@ -1,5 +0,0 @@
1
- {
2
- "### End": 50257,
3
- "### Instruction:": 50258,
4
- "### Response:\n": 50259
5
- }
 
 
 
 
 
 
checkpoint-4984/config.json DELETED
@@ -1,39 +0,0 @@
1
- {
2
- "_name_or_path": "vicgalle/gpt2-open-instruct-v1",
3
- "activation_function": "gelu_new",
4
- "architectures": [
5
- "GPT2LMHeadModel"
6
- ],
7
- "attn_pdrop": 0.1,
8
- "bos_token_id": 50256,
9
- "embd_pdrop": 0.1,
10
- "eos_token_id": 50256,
11
- "initializer_range": 0.02,
12
- "layer_norm_epsilon": 1e-05,
13
- "model_type": "gpt2",
14
- "n_ctx": 1024,
15
- "n_embd": 768,
16
- "n_head": 12,
17
- "n_inner": null,
18
- "n_layer": 12,
19
- "n_positions": 1024,
20
- "reorder_and_upcast_attn": false,
21
- "resid_pdrop": 0.1,
22
- "scale_attn_by_inverse_layer_idx": false,
23
- "scale_attn_weights": true,
24
- "summary_activation": null,
25
- "summary_first_dropout": 0.1,
26
- "summary_proj_to_labels": true,
27
- "summary_type": "cls_index",
28
- "summary_use_proj": true,
29
- "task_specific_params": {
30
- "text-generation": {
31
- "do_sample": true,
32
- "max_length": 50
33
- }
34
- },
35
- "torch_dtype": "float32",
36
- "transformers_version": "4.31.0",
37
- "use_cache": false,
38
- "vocab_size": 50260
39
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-4984/generation_config.json DELETED
@@ -1,7 +0,0 @@
1
- {
2
- "_from_model_config": true,
3
- "bos_token_id": 50256,
4
- "eos_token_id": 50256,
5
- "transformers_version": "4.31.0",
6
- "use_cache": false
7
- }
 
 
 
 
 
 
 
 
checkpoint-4984/merges.txt DELETED
The diff for this file is too large to render. See raw diff
 
checkpoint-4984/pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:7d70adf85a6993424510eeeaf760d3facdb64140786afdab235c554af03bf01e
3
- size 497816413
 
 
 
 
checkpoint-4984/rng_state.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:76feedbbe7df4e3995eb129b79413d6acc81d558e18412dfb3ca00225be0c886
3
- size 14575
 
 
 
 
checkpoint-4984/scheduler.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:4ceb9870ffb08bc8fbbb3b8fb254f8045dec405b8ca6a1772347b4db7641feb0
3
- size 563
 
 
 
 
checkpoint-4984/special_tokens_map.json DELETED
@@ -1,11 +0,0 @@
1
- {
2
- "additional_special_tokens": [
3
- "### End",
4
- "### Instruction:",
5
- "### Response:\n"
6
- ],
7
- "bos_token": "<|endoftext|>",
8
- "eos_token": "<|endoftext|>",
9
- "pad_token": "<|endoftext|>",
10
- "unk_token": "<|endoftext|>"
11
- }
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-4984/tokenizer_config.json DELETED
@@ -1,35 +0,0 @@
1
- {
2
- "add_bos_token": false,
3
- "add_prefix_space": false,
4
- "bos_token": {
5
- "__type": "AddedToken",
6
- "content": "<|endoftext|>",
7
- "lstrip": false,
8
- "normalized": true,
9
- "rstrip": false,
10
- "single_word": false
11
- },
12
- "clean_up_tokenization_spaces": true,
13
- "eos_token": {
14
- "__type": "AddedToken",
15
- "content": "<|endoftext|>",
16
- "lstrip": false,
17
- "normalized": true,
18
- "rstrip": false,
19
- "single_word": false
20
- },
21
- "errors": "replace",
22
- "max_seq_length": 512,
23
- "model_max_length": 1024,
24
- "pad_token": null,
25
- "tokenizer_class": "GPT2Tokenizer",
26
- "unk_token": {
27
- "__type": "AddedToken",
28
- "content": "<|endoftext|>",
29
- "lstrip": false,
30
- "normalized": true,
31
- "rstrip": false,
32
- "single_word": false
33
- },
34
- "use_fast": false
35
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-4984/trainer_state.json DELETED
@@ -1,102 +0,0 @@
1
- {
2
- "best_metric": 2.4491631984710693,
3
- "best_model_checkpoint": "gpt2-open-instruct-v1-gsm8k/checkpoint-4984",
4
- "epoch": 4.0,
5
- "global_step": 4984,
6
- "is_hyper_param_search": false,
7
- "is_local_process_zero": true,
8
- "is_world_process_zero": true,
9
- "log_history": [
10
- {
11
- "epoch": 0.4,
12
- "learning_rate": 0.00934910277324633,
13
- "loss": 2.9068,
14
- "step": 500
15
- },
16
- {
17
- "epoch": 0.8,
18
- "learning_rate": 0.008533442088091354,
19
- "loss": 2.6943,
20
- "step": 1000
21
- },
22
- {
23
- "epoch": 1.0,
24
- "eval_loss": 2.5155069828033447,
25
- "eval_runtime": 22.9749,
26
- "eval_samples_per_second": 57.41,
27
- "eval_steps_per_second": 57.41,
28
- "step": 1246
29
- },
30
- {
31
- "epoch": 1.2,
32
- "learning_rate": 0.007719412724306688,
33
- "loss": 2.6507,
34
- "step": 1500
35
- },
36
- {
37
- "epoch": 1.61,
38
- "learning_rate": 0.006903752039151713,
39
- "loss": 2.62,
40
- "step": 2000
41
- },
42
- {
43
- "epoch": 2.0,
44
- "eval_loss": 2.4768738746643066,
45
- "eval_runtime": 22.6896,
46
- "eval_samples_per_second": 58.132,
47
- "eval_steps_per_second": 58.132,
48
- "step": 2492
49
- },
50
- {
51
- "epoch": 2.01,
52
- "learning_rate": 0.006088091353996738,
53
- "loss": 2.6066,
54
- "step": 2500
55
- },
56
- {
57
- "epoch": 2.41,
58
- "learning_rate": 0.005272430668841762,
59
- "loss": 2.5892,
60
- "step": 3000
61
- },
62
- {
63
- "epoch": 2.81,
64
- "learning_rate": 0.004456769983686787,
65
- "loss": 2.5851,
66
- "step": 3500
67
- },
68
- {
69
- "epoch": 3.0,
70
- "eval_loss": 2.4574520587921143,
71
- "eval_runtime": 22.9332,
72
- "eval_samples_per_second": 57.515,
73
- "eval_steps_per_second": 57.515,
74
- "step": 3738
75
- },
76
- {
77
- "epoch": 3.21,
78
- "learning_rate": 0.0036427406199021207,
79
- "loss": 2.5574,
80
- "step": 4000
81
- },
82
- {
83
- "epoch": 3.61,
84
- "learning_rate": 0.0028270799347471453,
85
- "loss": 2.5692,
86
- "step": 4500
87
- },
88
- {
89
- "epoch": 4.0,
90
- "eval_loss": 2.4491631984710693,
91
- "eval_runtime": 22.7775,
92
- "eval_samples_per_second": 57.908,
93
- "eval_steps_per_second": 57.908,
94
- "step": 4984
95
- }
96
- ],
97
- "max_steps": 6230,
98
- "num_train_epochs": 5,
99
- "total_flos": 3295072161792000.0,
100
- "trial_name": null,
101
- "trial_params": null
102
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-4984/vocab.json DELETED
The diff for this file is too large to render. See raw diff
 
checkpoint-6230/added_tokens.json DELETED
@@ -1,5 +0,0 @@
1
- {
2
- "### End": 50257,
3
- "### Instruction:": 50258,
4
- "### Response:\n": 50259
5
- }
 
 
 
 
 
 
checkpoint-6230/config.json DELETED
@@ -1,39 +0,0 @@
1
- {
2
- "_name_or_path": "vicgalle/gpt2-open-instruct-v1",
3
- "activation_function": "gelu_new",
4
- "architectures": [
5
- "GPT2LMHeadModel"
6
- ],
7
- "attn_pdrop": 0.1,
8
- "bos_token_id": 50256,
9
- "embd_pdrop": 0.1,
10
- "eos_token_id": 50256,
11
- "initializer_range": 0.02,
12
- "layer_norm_epsilon": 1e-05,
13
- "model_type": "gpt2",
14
- "n_ctx": 1024,
15
- "n_embd": 768,
16
- "n_head": 12,
17
- "n_inner": null,
18
- "n_layer": 12,
19
- "n_positions": 1024,
20
- "reorder_and_upcast_attn": false,
21
- "resid_pdrop": 0.1,
22
- "scale_attn_by_inverse_layer_idx": false,
23
- "scale_attn_weights": true,
24
- "summary_activation": null,
25
- "summary_first_dropout": 0.1,
26
- "summary_proj_to_labels": true,
27
- "summary_type": "cls_index",
28
- "summary_use_proj": true,
29
- "task_specific_params": {
30
- "text-generation": {
31
- "do_sample": true,
32
- "max_length": 50
33
- }
34
- },
35
- "torch_dtype": "float32",
36
- "transformers_version": "4.31.0",
37
- "use_cache": false,
38
- "vocab_size": 50260
39
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-6230/generation_config.json DELETED
@@ -1,7 +0,0 @@
1
- {
2
- "_from_model_config": true,
3
- "bos_token_id": 50256,
4
- "eos_token_id": 50256,
5
- "transformers_version": "4.31.0",
6
- "use_cache": false
7
- }
 
 
 
 
 
 
 
 
checkpoint-6230/merges.txt DELETED
The diff for this file is too large to render. See raw diff
 
checkpoint-6230/optimizer.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:f677d070f6bb6cde72da19c977758b81933a889ce85e82afb9f208e22c9b38c9
3
- size 2291
 
 
 
 
checkpoint-6230/pytorch_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:a465e6267a176635c584180cf6fc0d931332074e5f265644e0e21469f228a2dd
3
- size 497816413
 
 
 
 
checkpoint-6230/rng_state.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:a38625868d66fb830b6c44973a821e372af7fedc34777fd7f89f92300a25d9fe
3
- size 14575
 
 
 
 
checkpoint-6230/special_tokens_map.json DELETED
@@ -1,11 +0,0 @@
1
- {
2
- "additional_special_tokens": [
3
- "### End",
4
- "### Instruction:",
5
- "### Response:\n"
6
- ],
7
- "bos_token": "<|endoftext|>",
8
- "eos_token": "<|endoftext|>",
9
- "pad_token": "<|endoftext|>",
10
- "unk_token": "<|endoftext|>"
11
- }
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-6230/tokenizer_config.json DELETED
@@ -1,35 +0,0 @@
1
- {
2
- "add_bos_token": false,
3
- "add_prefix_space": false,
4
- "bos_token": {
5
- "__type": "AddedToken",
6
- "content": "<|endoftext|>",
7
- "lstrip": false,
8
- "normalized": true,
9
- "rstrip": false,
10
- "single_word": false
11
- },
12
- "clean_up_tokenization_spaces": true,
13
- "eos_token": {
14
- "__type": "AddedToken",
15
- "content": "<|endoftext|>",
16
- "lstrip": false,
17
- "normalized": true,
18
- "rstrip": false,
19
- "single_word": false
20
- },
21
- "errors": "replace",
22
- "max_seq_length": 512,
23
- "model_max_length": 1024,
24
- "pad_token": null,
25
- "tokenizer_class": "GPT2Tokenizer",
26
- "unk_token": {
27
- "__type": "AddedToken",
28
- "content": "<|endoftext|>",
29
- "lstrip": false,
30
- "normalized": true,
31
- "rstrip": false,
32
- "single_word": false
33
- },
34
- "use_fast": false
35
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-6230/trainer_state.json DELETED
@@ -1,128 +0,0 @@
1
- {
2
- "best_metric": 2.445409059524536,
3
- "best_model_checkpoint": "gpt2-open-instruct-v1-gsm8k/checkpoint-6230",
4
- "epoch": 5.0,
5
- "global_step": 6230,
6
- "is_hyper_param_search": false,
7
- "is_local_process_zero": true,
8
- "is_world_process_zero": true,
9
- "log_history": [
10
- {
11
- "epoch": 0.4,
12
- "learning_rate": 0.00934910277324633,
13
- "loss": 2.9068,
14
- "step": 500
15
- },
16
- {
17
- "epoch": 0.8,
18
- "learning_rate": 0.008533442088091354,
19
- "loss": 2.6943,
20
- "step": 1000
21
- },
22
- {
23
- "epoch": 1.0,
24
- "eval_loss": 2.5155069828033447,
25
- "eval_runtime": 22.9749,
26
- "eval_samples_per_second": 57.41,
27
- "eval_steps_per_second": 57.41,
28
- "step": 1246
29
- },
30
- {
31
- "epoch": 1.2,
32
- "learning_rate": 0.007719412724306688,
33
- "loss": 2.6507,
34
- "step": 1500
35
- },
36
- {
37
- "epoch": 1.61,
38
- "learning_rate": 0.006903752039151713,
39
- "loss": 2.62,
40
- "step": 2000
41
- },
42
- {
43
- "epoch": 2.0,
44
- "eval_loss": 2.4768738746643066,
45
- "eval_runtime": 22.6896,
46
- "eval_samples_per_second": 58.132,
47
- "eval_steps_per_second": 58.132,
48
- "step": 2492
49
- },
50
- {
51
- "epoch": 2.01,
52
- "learning_rate": 0.006088091353996738,
53
- "loss": 2.6066,
54
- "step": 2500
55
- },
56
- {
57
- "epoch": 2.41,
58
- "learning_rate": 0.005272430668841762,
59
- "loss": 2.5892,
60
- "step": 3000
61
- },
62
- {
63
- "epoch": 2.81,
64
- "learning_rate": 0.004456769983686787,
65
- "loss": 2.5851,
66
- "step": 3500
67
- },
68
- {
69
- "epoch": 3.0,
70
- "eval_loss": 2.4574520587921143,
71
- "eval_runtime": 22.9332,
72
- "eval_samples_per_second": 57.515,
73
- "eval_steps_per_second": 57.515,
74
- "step": 3738
75
- },
76
- {
77
- "epoch": 3.21,
78
- "learning_rate": 0.0036427406199021207,
79
- "loss": 2.5574,
80
- "step": 4000
81
- },
82
- {
83
- "epoch": 3.61,
84
- "learning_rate": 0.0028270799347471453,
85
- "loss": 2.5692,
86
- "step": 4500
87
- },
88
- {
89
- "epoch": 4.0,
90
- "eval_loss": 2.4491631984710693,
91
- "eval_runtime": 22.7775,
92
- "eval_samples_per_second": 57.908,
93
- "eval_steps_per_second": 57.908,
94
- "step": 4984
95
- },
96
- {
97
- "epoch": 4.01,
98
- "learning_rate": 0.0020114192495921694,
99
- "loss": 2.552,
100
- "step": 5000
101
- },
102
- {
103
- "epoch": 4.41,
104
- "learning_rate": 0.0011957585644371941,
105
- "loss": 2.5497,
106
- "step": 5500
107
- },
108
- {
109
- "epoch": 4.82,
110
- "learning_rate": 0.00038009787928221857,
111
- "loss": 2.5485,
112
- "step": 6000
113
- },
114
- {
115
- "epoch": 5.0,
116
- "eval_loss": 2.445409059524536,
117
- "eval_runtime": 23.0868,
118
- "eval_samples_per_second": 57.132,
119
- "eval_steps_per_second": 57.132,
120
- "step": 6230
121
- }
122
- ],
123
- "max_steps": 6230,
124
- "num_train_epochs": 5,
125
- "total_flos": 4118840202240000.0,
126
- "trial_name": null,
127
- "trial_params": null
128
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
checkpoint-6230/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:3bd44df1d17d716db8d85f6c86f061584611315664462a60a46b9ceb8331de89
3
- size 3963
 
 
 
 
checkpoint-6230/vocab.json DELETED
The diff for this file is too large to render. See raw diff
 
checkpoint-4984/optimizer.pt → logs/events.out.tfevents.1689925715.d13dacfe7802.300.0 RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:318971244dfffc95173117a6525c3f6e9f9948288811a51c2788fbb0fc427275
3
- size 2291
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02b87257f276058bac8429e631eff8f3faa19cbdd6b383a1742f0990a6428c0a
3
+ size 4523
checkpoint-4984/training_args.bin → logs/events.out.tfevents.1689925893.d13dacfe7802.300.1 RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3bd44df1d17d716db8d85f6c86f061584611315664462a60a46b9ceb8331de89
3
- size 3963
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c61c3fe17b90c87dd4d63427fdb564133a2f965c2190101789f9d1c6706bcac
3
+ size 12967
checkpoint-6230/scheduler.pt → logs/events.out.tfevents.1689930197.d13dacfe7802.300.2 RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:74ea958f7352f0890043b0bb3bbc372808bc3a972fe88a1fd8c718fd2193e955
3
- size 563
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4febb715e1398d6a487ad5501b31ced295e528995eedf56eb3586da2981db3b
3
+ size 359
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a465e6267a176635c584180cf6fc0d931332074e5f265644e0e21469f228a2dd
3
  size 497816413
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4288585a5ffb6cd395063682bd07810d7dddccb645320b278bafca2cfc1599d0
3
  size 497816413
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3bd44df1d17d716db8d85f6c86f061584611315664462a60a46b9ceb8331de89
3
  size 3963
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8dbf2b8cfdf70234153c42c24487cbca08f6a6285ee3d6fdf1488a7180d01958
3
  size 3963