nateraw commited on
Commit
ccfb510
1 Parent(s): c0dcd38

Upload .outputs/outputs with huggingface_hub

Browse files
.outputs/outputs/README.md ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: apache-2.0
5
+ tags:
6
+ - generated_from_trainer
7
+ datasets:
8
+ - glue
9
+ metrics:
10
+ - accuracy
11
+ - f1
12
+ model-index:
13
+ - name: outputs
14
+ results:
15
+ - task:
16
+ name: Text Classification
17
+ type: text-classification
18
+ dataset:
19
+ name: GLUE MRPC
20
+ type: glue
21
+ config: mrpc
22
+ split: validation
23
+ args: mrpc
24
+ metrics:
25
+ - name: Accuracy
26
+ type: accuracy
27
+ value: 0.8529411764705882
28
+ - name: F1
29
+ type: f1
30
+ value: 0.8969072164948454
31
+ ---
32
+
33
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
34
+ should probably proofread and complete it, then remove this comment. -->
35
+
36
+ # outputs
37
+
38
+ This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the GLUE MRPC dataset.
39
+ It achieves the following results on the evaluation set:
40
+ - Loss: 0.4175
41
+ - Accuracy: 0.8529
42
+ - F1: 0.8969
43
+ - Combined Score: 0.8749
44
+
45
+ ## Model description
46
+
47
+ More information needed
48
+
49
+ ## Intended uses & limitations
50
+
51
+ More information needed
52
+
53
+ ## Training and evaluation data
54
+
55
+ More information needed
56
+
57
+ ## Training procedure
58
+
59
+ ### Training hyperparameters
60
+
61
+ The following hyperparameters were used during training:
62
+ - learning_rate: 2e-05
63
+ - train_batch_size: 32
64
+ - eval_batch_size: 8
65
+ - seed: 42
66
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
67
+ - lr_scheduler_type: linear
68
+ - num_epochs: 3.0
69
+
70
+ ### Training results
71
+
72
+
73
+
74
+ ### Framework versions
75
+
76
+ - Transformers 4.27.0.dev0
77
+ - Pytorch 1.13.1+cu117
78
+ - Datasets 2.9.0
79
+ - Tokenizers 0.13.2
.outputs/outputs/all_results.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "eval_accuracy": 0.8529411764705882,
4
+ "eval_combined_score": 0.8749241964827168,
5
+ "eval_f1": 0.8969072164948454,
6
+ "eval_loss": 0.41747015714645386,
7
+ "eval_runtime": 2.6884,
8
+ "eval_samples": 408,
9
+ "eval_samples_per_second": 151.761,
10
+ "eval_steps_per_second": 18.97,
11
+ "train_loss": 0.42796855594800864,
12
+ "train_runtime": 195.493,
13
+ "train_samples": 3668,
14
+ "train_samples_per_second": 56.288,
15
+ "train_steps_per_second": 1.765
16
+ }
.outputs/outputs/config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-base-cased",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "finetuning_task": "mrpc",
9
+ "gradient_checkpointing": false,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 768,
13
+ "id2label": {
14
+ "0": "not_equivalent",
15
+ "1": "equivalent"
16
+ },
17
+ "initializer_range": 0.02,
18
+ "intermediate_size": 3072,
19
+ "label2id": {
20
+ "equivalent": 1,
21
+ "not_equivalent": 0
22
+ },
23
+ "layer_norm_eps": 1e-12,
24
+ "max_position_embeddings": 512,
25
+ "model_type": "bert",
26
+ "num_attention_heads": 12,
27
+ "num_hidden_layers": 12,
28
+ "pad_token_id": 0,
29
+ "position_embedding_type": "absolute",
30
+ "problem_type": "single_label_classification",
31
+ "torch_dtype": "float32",
32
+ "transformers_version": "4.27.0.dev0",
33
+ "type_vocab_size": 2,
34
+ "use_cache": true,
35
+ "vocab_size": 28996
36
+ }
.outputs/outputs/eval_results.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "eval_accuracy": 0.8529411764705882,
4
+ "eval_combined_score": 0.8749241964827168,
5
+ "eval_f1": 0.8969072164948454,
6
+ "eval_loss": 0.41747015714645386,
7
+ "eval_runtime": 2.6884,
8
+ "eval_samples": 408,
9
+ "eval_samples_per_second": 151.761,
10
+ "eval_steps_per_second": 18.97
11
+ }
.outputs/outputs/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc17c63d4329f110d5ff1c4ab0c1ea919f87748e27867f12b915e9cb3078fcef
3
+ size 433320053
.outputs/outputs/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
.outputs/outputs/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
.outputs/outputs/tokenizer_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_lower_case": false,
4
+ "mask_token": "[MASK]",
5
+ "model_max_length": 512,
6
+ "name_or_path": "bert-base-cased",
7
+ "pad_token": "[PAD]",
8
+ "sep_token": "[SEP]",
9
+ "special_tokens_map_file": null,
10
+ "strip_accents": null,
11
+ "tokenize_chinese_chars": true,
12
+ "tokenizer_class": "BertTokenizer",
13
+ "unk_token": "[UNK]"
14
+ }
.outputs/outputs/train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "train_loss": 0.42796855594800864,
4
+ "train_runtime": 195.493,
5
+ "train_samples": 3668,
6
+ "train_samples_per_second": 56.288,
7
+ "train_steps_per_second": 1.765
8
+ }
.outputs/outputs/trainer_state.json ADDED
@@ -0,0 +1,127 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 3.0,
5
+ "global_step": 345,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.17,
12
+ "learning_rate": 1.8840579710144928e-05,
13
+ "loss": 0.8005,
14
+ "step": 20
15
+ },
16
+ {
17
+ "epoch": 0.35,
18
+ "learning_rate": 1.7681159420289858e-05,
19
+ "loss": 0.5749,
20
+ "step": 40
21
+ },
22
+ {
23
+ "epoch": 0.52,
24
+ "learning_rate": 1.6521739130434785e-05,
25
+ "loss": 0.567,
26
+ "step": 60
27
+ },
28
+ {
29
+ "epoch": 0.7,
30
+ "learning_rate": 1.536231884057971e-05,
31
+ "loss": 0.5386,
32
+ "step": 80
33
+ },
34
+ {
35
+ "epoch": 0.87,
36
+ "learning_rate": 1.420289855072464e-05,
37
+ "loss": 0.5704,
38
+ "step": 100
39
+ },
40
+ {
41
+ "epoch": 1.04,
42
+ "learning_rate": 1.3043478260869566e-05,
43
+ "loss": 0.4953,
44
+ "step": 120
45
+ },
46
+ {
47
+ "epoch": 1.22,
48
+ "learning_rate": 1.1884057971014494e-05,
49
+ "loss": 0.4468,
50
+ "step": 140
51
+ },
52
+ {
53
+ "epoch": 1.39,
54
+ "learning_rate": 1.0724637681159422e-05,
55
+ "loss": 0.4556,
56
+ "step": 160
57
+ },
58
+ {
59
+ "epoch": 1.57,
60
+ "learning_rate": 9.565217391304349e-06,
61
+ "loss": 0.4017,
62
+ "step": 180
63
+ },
64
+ {
65
+ "epoch": 1.74,
66
+ "learning_rate": 8.405797101449275e-06,
67
+ "loss": 0.3741,
68
+ "step": 200
69
+ },
70
+ {
71
+ "epoch": 1.91,
72
+ "learning_rate": 7.246376811594203e-06,
73
+ "loss": 0.3916,
74
+ "step": 220
75
+ },
76
+ {
77
+ "epoch": 2.09,
78
+ "learning_rate": 6.086956521739132e-06,
79
+ "loss": 0.3098,
80
+ "step": 240
81
+ },
82
+ {
83
+ "epoch": 2.26,
84
+ "learning_rate": 4.927536231884059e-06,
85
+ "loss": 0.29,
86
+ "step": 260
87
+ },
88
+ {
89
+ "epoch": 2.43,
90
+ "learning_rate": 3.768115942028986e-06,
91
+ "loss": 0.3175,
92
+ "step": 280
93
+ },
94
+ {
95
+ "epoch": 2.61,
96
+ "learning_rate": 2.6086956521739132e-06,
97
+ "loss": 0.2574,
98
+ "step": 300
99
+ },
100
+ {
101
+ "epoch": 2.78,
102
+ "learning_rate": 1.4492753623188408e-06,
103
+ "loss": 0.2649,
104
+ "step": 320
105
+ },
106
+ {
107
+ "epoch": 2.96,
108
+ "learning_rate": 2.8985507246376816e-07,
109
+ "loss": 0.2768,
110
+ "step": 340
111
+ },
112
+ {
113
+ "epoch": 3.0,
114
+ "step": 345,
115
+ "total_flos": 723818513295360.0,
116
+ "train_loss": 0.42796855594800864,
117
+ "train_runtime": 195.493,
118
+ "train_samples_per_second": 56.288,
119
+ "train_steps_per_second": 1.765
120
+ }
121
+ ],
122
+ "max_steps": 345,
123
+ "num_train_epochs": 3,
124
+ "total_flos": 723818513295360.0,
125
+ "trial_name": null,
126
+ "trial_params": null
127
+ }
.outputs/outputs/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f135431444b98bb9aa8d8971add320af2a016e334d846844040e9a80244dcff3
3
+ size 3451
.outputs/outputs/vocab.txt ADDED
The diff for this file is too large to render. See raw diff