LogischeIP commited on
Commit
344d352
1 Parent(s): 0dd8dff

End of training

Browse files
Files changed (26) hide show
  1. README.md +69 -0
  2. config.json +40 -0
  3. merges.txt +0 -0
  4. model.safetensors +3 -0
  5. runs/Jan24_14-37-05_579ed4a51ca4/events.out.tfevents.1706107026.579ed4a51ca4.457.0 +3 -0
  6. runs/Jan24_14-37-05_579ed4a51ca4/events.out.tfevents.1706107294.579ed4a51ca4.457.1 +3 -0
  7. runs/Jan24_17-57-00_7d2b1e2d4e1a/events.out.tfevents.1706119022.7d2b1e2d4e1a.2199.0 +3 -0
  8. runs/Jan24_17-57-30_7d2b1e2d4e1a/events.out.tfevents.1706119051.7d2b1e2d4e1a.2199.1 +3 -0
  9. runs/Jan24_17-57-30_7d2b1e2d4e1a/events.out.tfevents.1706119493.7d2b1e2d4e1a.2199.3 +3 -0
  10. runs/Jan24_17-59-31_7d2b1e2d4e1a/events.out.tfevents.1706119172.7d2b1e2d4e1a.2199.2 +3 -0
  11. runs/Jan24_17-59-31_7d2b1e2d4e1a/events.out.tfevents.1706119503.7d2b1e2d4e1a.2199.4 +3 -0
  12. runs/Jan24_18-39-04_7d2b1e2d4e1a/events.out.tfevents.1706121545.7d2b1e2d4e1a.2199.7 +3 -0
  13. runs/Jan24_19-20-11_7d2b1e2d4e1a/events.out.tfevents.1706124012.7d2b1e2d4e1a.2199.8 +3 -0
  14. runs/Jan24_19-23-12_7d2b1e2d4e1a/events.out.tfevents.1706124192.7d2b1e2d4e1a.2199.9 +3 -0
  15. runs/Jan24_19-23-37_7d2b1e2d4e1a/events.out.tfevents.1706124217.7d2b1e2d4e1a.2199.10 +3 -0
  16. runs/Jan24_19-31-14_7d2b1e2d4e1a/events.out.tfevents.1706124675.7d2b1e2d4e1a.26662.0 +3 -0
  17. runs/Jan24_19-32-18_7d2b1e2d4e1a/events.out.tfevents.1706124738.7d2b1e2d4e1a.26662.1 +3 -0
  18. runs/Jan24_19-33-23_7d2b1e2d4e1a/events.out.tfevents.1706124804.7d2b1e2d4e1a.26662.2 +3 -0
  19. runs/Jan24_19-33-53_7d2b1e2d4e1a/events.out.tfevents.1706124834.7d2b1e2d4e1a.26662.3 +3 -0
  20. runs/Jan24_19-39-23_7d2b1e2d4e1a/events.out.tfevents.1706125164.7d2b1e2d4e1a.26662.4 +3 -0
  21. special_tokens_map.json +6 -0
  22. tokenizer.json +0 -0
  23. tokenizer_config.json +20 -0
  24. training_args.bin +3 -0
  25. vocab.json +0 -0
  26. vocab.txt +0 -0
README.md ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ base_model: gpt2
4
+ tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - accuracy
8
+ - f1
9
+ model-index:
10
+ - name: SentimentT2_GPT2
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # SentimentT2_GPT2
18
+
19
+ This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on the None dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 0.7249
22
+ - Accuracy: 0.8669
23
+ - F1: 0.8749
24
+ - Auc Roc: 0.9317
25
+ - Log Loss: 0.7249
26
+
27
+ ## Model description
28
+
29
+ More information needed
30
+
31
+ ## Intended uses & limitations
32
+
33
+ More information needed
34
+
35
+ ## Training and evaluation data
36
+
37
+ More information needed
38
+
39
+ ## Training procedure
40
+
41
+ ### Training hyperparameters
42
+
43
+ The following hyperparameters were used during training:
44
+ - learning_rate: 1e-05
45
+ - train_batch_size: 1
46
+ - eval_batch_size: 1
47
+ - seed: 42
48
+ - gradient_accumulation_steps: 2
49
+ - total_train_batch_size: 2
50
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
+ - lr_scheduler_type: cosine
52
+ - lr_scheduler_warmup_steps: 500
53
+ - num_epochs: 2
54
+ - mixed_precision_training: Native AMP
55
+
56
+ ### Training results
57
+
58
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Auc Roc | Log Loss |
59
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:-------:|:--------:|
60
+ | 1.1736 | 1.0 | 1618 | 0.6146 | 0.8507 | 0.8592 | 0.9256 | 0.6146 |
61
+ | 0.6452 | 2.0 | 3236 | 0.7249 | 0.8669 | 0.8749 | 0.9317 | 0.7249 |
62
+
63
+
64
+ ### Framework versions
65
+
66
+ - Transformers 4.35.2
67
+ - Pytorch 2.1.0+cu121
68
+ - Datasets 2.16.1
69
+ - Tokenizers 0.15.1
config.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "gpt2",
3
+ "activation_function": "gelu_new",
4
+ "architectures": [
5
+ "GPT2ForSequenceClassification"
6
+ ],
7
+ "attn_pdrop": 0.1,
8
+ "bos_token_id": 50256,
9
+ "embd_pdrop": 0.1,
10
+ "eos_token_id": 50256,
11
+ "initializer_range": 0.02,
12
+ "layer_norm_epsilon": 1e-05,
13
+ "model_type": "gpt2",
14
+ "n_ctx": 1024,
15
+ "n_embd": 768,
16
+ "n_head": 12,
17
+ "n_inner": null,
18
+ "n_layer": 12,
19
+ "n_positions": 1024,
20
+ "problem_type": "single_label_classification",
21
+ "reorder_and_upcast_attn": false,
22
+ "resid_pdrop": 0.1,
23
+ "scale_attn_by_inverse_layer_idx": false,
24
+ "scale_attn_weights": true,
25
+ "summary_activation": null,
26
+ "summary_first_dropout": 0.1,
27
+ "summary_proj_to_labels": true,
28
+ "summary_type": "cls_index",
29
+ "summary_use_proj": true,
30
+ "task_specific_params": {
31
+ "text-generation": {
32
+ "do_sample": true,
33
+ "max_length": 50
34
+ }
35
+ },
36
+ "torch_dtype": "float32",
37
+ "transformers_version": "4.35.2",
38
+ "use_cache": true,
39
+ "vocab_size": 50257
40
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02f0b5d0a85a3b873b4a183a19c22b0acf256a4c08e0f9b1355b1c34211241e5
3
+ size 497780432
runs/Jan24_14-37-05_579ed4a51ca4/events.out.tfevents.1706107026.579ed4a51ca4.457.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9793fe0c6556e907b1c614be0ace204cb95291a8b77b8a77fb2966750c729a64
3
+ size 7042
runs/Jan24_14-37-05_579ed4a51ca4/events.out.tfevents.1706107294.579ed4a51ca4.457.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a839574f814954dbb18f4b789b2bf531497e649d8830ae90a2484a723adc8d1e
3
+ size 560
runs/Jan24_17-57-00_7d2b1e2d4e1a/events.out.tfevents.1706119022.7d2b1e2d4e1a.2199.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd2d584b67487707bc9efea3ec34eb13466a73d8f15f61b3954cc69b9c0715d8
3
+ size 4184
runs/Jan24_17-57-30_7d2b1e2d4e1a/events.out.tfevents.1706119051.7d2b1e2d4e1a.2199.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ade5df7e518384ed3538208ee6c86b7c11214a6eceaa75b0e2a65effdbec6729
3
+ size 5784
runs/Jan24_17-57-30_7d2b1e2d4e1a/events.out.tfevents.1706119493.7d2b1e2d4e1a.2199.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b3214ab547043011326ad1369c0d723f88faf092bab310c987101761c07b531
3
+ size 560
runs/Jan24_17-59-31_7d2b1e2d4e1a/events.out.tfevents.1706119172.7d2b1e2d4e1a.2199.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a61710090d075321d12b7f8cacff814f10e9b8b64a06b7373eadc10dd1dee00
3
+ size 5905
runs/Jan24_17-59-31_7d2b1e2d4e1a/events.out.tfevents.1706119503.7d2b1e2d4e1a.2199.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ebeb647374b9b6ed5833b093f646e9c089c5a8ac9c0be5fd010d01d22832d1ca
3
+ size 560
runs/Jan24_18-39-04_7d2b1e2d4e1a/events.out.tfevents.1706121545.7d2b1e2d4e1a.2199.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2aa18069c9bab305333570793f1aaab2ebe626ed027930529345da1d81bd1924
3
+ size 10938
runs/Jan24_19-20-11_7d2b1e2d4e1a/events.out.tfevents.1706124012.7d2b1e2d4e1a.2199.8 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1ff5cea8170bbdd6e49c9181719bfa37e826252a7912b96c5bfd3379d9ac8e9
3
+ size 4544
runs/Jan24_19-23-12_7d2b1e2d4e1a/events.out.tfevents.1706124192.7d2b1e2d4e1a.2199.9 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d48965daa51cd8c292e9726fa9ce868c27b7b7a0b26953610a1284b0a3244db0
3
+ size 4544
runs/Jan24_19-23-37_7d2b1e2d4e1a/events.out.tfevents.1706124217.7d2b1e2d4e1a.2199.10 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92ce18dee6663fdd82856765199eafdf48f24704820cb05dea5b7f9a1d3e3b4d
3
+ size 4543
runs/Jan24_19-31-14_7d2b1e2d4e1a/events.out.tfevents.1706124675.7d2b1e2d4e1a.26662.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f052f9d74b7d2a9b38e8f79422b89a997f4df9ad34e859927deefd6aeb99e59c
3
+ size 4543
runs/Jan24_19-32-18_7d2b1e2d4e1a/events.out.tfevents.1706124738.7d2b1e2d4e1a.26662.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:983fea0d1e6ed2f5a2d7d497e5a5e25f161ea62d41b9150f99dd72d0af6a2211
3
+ size 4543
runs/Jan24_19-33-23_7d2b1e2d4e1a/events.out.tfevents.1706124804.7d2b1e2d4e1a.26662.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d826890967001774b0ee65bd6c10a7a5ac575902f6bff7de0c7a97a409809e58
3
+ size 4543
runs/Jan24_19-33-53_7d2b1e2d4e1a/events.out.tfevents.1706124834.7d2b1e2d4e1a.26662.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53026ad3317456be6c8ffc66d5cc581d9e09640d1bfb7ae3af5c1f947ca13845
3
+ size 4700
runs/Jan24_19-39-23_7d2b1e2d4e1a/events.out.tfevents.1706125164.7d2b1e2d4e1a.26662.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f0106f7873903e963089ccc4d13d050b3100e0bccfd2c4fb500c1e0835913c4
3
+ size 6154
special_tokens_map.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<|endoftext|>",
3
+ "eos_token": "<|endoftext|>",
4
+ "pad_token": "<|endoftext|>",
5
+ "unk_token": "<|endoftext|>"
6
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "50256": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ }
12
+ },
13
+ "bos_token": "<|endoftext|>",
14
+ "clean_up_tokenization_spaces": true,
15
+ "eos_token": "<|endoftext|>",
16
+ "model_max_length": 1024,
17
+ "pad_token": "<|endoftext|>",
18
+ "tokenizer_class": "GPT2Tokenizer",
19
+ "unk_token": "<|endoftext|>"
20
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3de59719afd225fb332ed6d1b42b6d228a99601e6a3b5db147d421b6288ddab6
3
+ size 4664
vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
vocab.txt ADDED
The diff for this file is too large to render. See raw diff