Eliac11 commited on
Commit
548a7c0
1 Parent(s): 164cef9

Training in progress epoch 0

Browse files
README.md CHANGED
@@ -14,9 +14,9 @@ probably proofread and complete it, then remove this comment. -->
14
 
15
  This model is a fine-tuned version of [Eliac11/tinkNLP](https://huggingface.co/Eliac11/tinkNLP) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
- - Train Loss: 0.5899
18
- - Validation Loss: 2.1619
19
- - Epoch: 23
20
 
21
  ## Model description
22
 
@@ -42,35 +42,12 @@ The following hyperparameters were used during training:
42
 
43
  | Train Loss | Validation Loss | Epoch |
44
  |:----------:|:---------------:|:-----:|
45
- | 2.7611 | 3.2917 | 0 |
46
- | 2.5272 | 3.1344 | 1 |
47
- | 2.3172 | 2.9982 | 2 |
48
- | 2.1298 | 2.8936 | 3 |
49
- | 1.9620 | 2.7805 | 4 |
50
- | 1.8145 | 2.6929 | 5 |
51
- | 1.6762 | 2.6139 | 6 |
52
- | 1.5531 | 2.5450 | 7 |
53
- | 1.4394 | 2.4895 | 8 |
54
- | 1.3463 | 2.4272 | 9 |
55
- | 1.2504 | 2.3718 | 10 |
56
- | 1.1740 | 2.3381 | 11 |
57
- | 1.0953 | 2.3067 | 12 |
58
- | 1.0245 | 2.2730 | 13 |
59
- | 0.9623 | 2.2482 | 14 |
60
- | 0.9074 | 2.2176 | 15 |
61
- | 0.8601 | 2.2021 | 16 |
62
- | 0.8079 | 2.1901 | 17 |
63
- | 0.7651 | 2.1776 | 18 |
64
- | 0.7242 | 2.1688 | 19 |
65
- | 0.6864 | 2.1613 | 20 |
66
- | 0.6494 | 2.1500 | 21 |
67
- | 0.6192 | 2.1510 | 22 |
68
- | 0.5899 | 2.1619 | 23 |
69
 
70
 
71
  ### Framework versions
72
 
73
- - Transformers 4.33.1
74
- - TensorFlow 2.12.0
75
- - Datasets 2.14.5
76
- - Tokenizers 0.13.3
 
14
 
15
  This model is a fine-tuned version of [Eliac11/tinkNLP](https://huggingface.co/Eliac11/tinkNLP) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
+ - Train Loss: 5.5514
18
+ - Validation Loss: 4.8451
19
+ - Epoch: 0
20
 
21
  ## Model description
22
 
 
42
 
43
  | Train Loss | Validation Loss | Epoch |
44
  |:----------:|:---------------:|:-----:|
45
+ | 5.5514 | 4.8451 | 0 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
46
 
47
 
48
  ### Framework versions
49
 
50
+ - Transformers 4.38.2
51
+ - TensorFlow 2.15.0
52
+ - Datasets 2.18.0
53
+ - Tokenizers 0.15.2
config.json CHANGED
@@ -28,7 +28,7 @@
28
  "summary_type": "cls_index",
29
  "summary_use_proj": true,
30
  "torch_dtype": "float32",
31
- "transformers_version": "4.33.1",
32
  "use_cache": true,
33
  "vocab_size": 50261
34
  }
 
28
  "summary_type": "cls_index",
29
  "summary_use_proj": true,
30
  "torch_dtype": "float32",
31
+ "transformers_version": "4.38.2",
32
  "use_cache": true,
33
  "vocab_size": 50261
34
  }
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 50256,
4
  "eos_token_id": 50256,
5
- "transformers_version": "4.33.1"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 50256,
4
  "eos_token_id": 50256,
5
+ "transformers_version": "4.38.2"
6
  }
special_tokens_map.json CHANGED
@@ -13,7 +13,13 @@
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
- "pad_token": "<pad>",
 
 
 
 
 
 
17
  "unk_token": {
18
  "content": "<|endoftext|>",
19
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "pad_token": {
17
+ "content": "<pad>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
  "unk_token": {
24
  "content": "<|endoftext|>",
25
  "lstrip": false,
tf_model.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a08f76e2b8be1216348de2098c78f2645e955866e342de5cd0fa3b48c56702ed
3
  size 501093456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3a2028ccd2ea3105af8c5b15022591793468213467cd12bfbb5e5aba742beaf
3
  size 501093456
tokenizer_config.json CHANGED
@@ -1,34 +1,63 @@
1
  {
2
  "add_bos_token": false,
3
  "add_prefix_space": false,
4
- "bos_token": {
5
- "__type": "AddedToken",
6
- "content": "<|endoftext|>",
7
- "lstrip": false,
8
- "normalized": true,
9
- "rstrip": false,
10
- "single_word": false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  },
 
12
  "clean_up_tokenization_spaces": true,
13
- "eos_token": {
14
- "__type": "AddedToken",
15
- "content": "<|endoftext|>",
16
- "lstrip": false,
17
- "normalized": true,
18
- "rstrip": false,
19
- "single_word": false
20
- },
21
  "errors": "replace",
22
  "model_max_length": 1000000000000000019884624838656,
23
- "pad_token": null,
24
  "tokenizer_class": "GPT2Tokenizer",
25
- "unk_token": {
26
- "__type": "AddedToken",
27
- "content": "<|endoftext|>",
28
- "lstrip": false,
29
- "normalized": true,
30
- "rstrip": false,
31
- "single_word": false
32
- },
33
  "use_fast": true
34
  }
 
1
  {
2
  "add_bos_token": false,
3
  "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<pad>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "50257": {
14
+ "content": "@@ПЕРВЫЙ@@",
15
+ "lstrip": false,
16
+ "normalized": true,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": false
20
+ },
21
+ "50258": {
22
+ "content": "@@ВТОРОЙ@@",
23
+ "lstrip": false,
24
+ "normalized": true,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": false
28
+ },
29
+ "50259": {
30
+ "content": "<FIRST_SPEAKER>",
31
+ "lstrip": false,
32
+ "normalized": true,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": false
36
+ },
37
+ "50260": {
38
+ "content": "<SECOND_SPEAKER>",
39
+ "lstrip": false,
40
+ "normalized": true,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": false
44
+ },
45
+ "50261": {
46
+ "content": "<|endoftext|>",
47
+ "lstrip": false,
48
+ "normalized": true,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ }
53
  },
54
+ "bos_token": "<|endoftext|>",
55
  "clean_up_tokenization_spaces": true,
56
+ "eos_token": "<|endoftext|>",
 
 
 
 
 
 
 
57
  "errors": "replace",
58
  "model_max_length": 1000000000000000019884624838656,
59
+ "pad_token": "<pad>",
60
  "tokenizer_class": "GPT2Tokenizer",
61
+ "unk_token": "<|endoftext|>",
 
 
 
 
 
 
 
62
  "use_fast": true
63
  }