Leo97 commited on
Commit
3c243cd
1 Parent(s): 2c0d93c

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -10,72 +10,72 @@
10
  "hidden_dropout_prob": 0.1,
11
  "hidden_size": 256,
12
  "id2label": {
13
- "0": "O",
14
- "1": "B-PS",
15
- "2": "I-PS",
16
- "3": "B-FD",
17
- "4": "I-FD",
18
- "5": "B-TR",
19
- "6": "I-TR",
20
- "7": "B-AF",
21
- "8": "I-AF",
22
- "9": "B-OG",
23
- "10": "I-OG",
24
- "11": "B-LC",
25
- "12": "I-LC",
26
- "13": "B-CV",
27
- "14": "I-CV",
28
- "15": "B-DT",
29
- "16": "I-DT",
30
- "17": "B-TI",
31
- "18": "I-TI",
32
- "19": "B-QT",
33
- "20": "I-QT",
34
- "21": "B-EV",
35
- "22": "I-EV",
36
- "23": "B-AM",
37
- "24": "I-AM",
38
- "25": "B-PT",
39
- "26": "I-PT",
40
- "27": "B-MT",
41
- "28": "I-MT",
42
- "29": "B-TM",
43
- "30": "I-TM"
44
  },
45
  "initializer_range": 0.02,
46
  "intermediate_size": 1024,
47
  "label2id": {
48
- "O": 0,
49
- "B-PS": 1,
50
- "I-PS": 2,
51
- "B-FD": 3,
52
- "I-FD": 4,
53
- "B-TR": 5,
54
- "I-TR": 6,
55
- "B-AF": 7,
56
- "I-AF": 8,
57
- "B-OG": 9,
58
- "I-OG": 10,
59
- "B-LC": 11,
60
- "I-LC": 12,
61
- "B-CV": 13,
62
- "I-CV": 14,
63
- "B-DT": 15,
64
- "I-DT": 16,
65
- "B-TI": 17,
66
- "I-TI": 18,
67
- "B-QT": 19,
68
- "I-QT": 20,
69
- "B-EV": 21,
70
- "I-EV": 22,
71
- "B-AM": 23,
72
- "I-AM": 24,
73
- "B-PT": 25,
74
- "I-PT": 26,
75
- "B-MT": 27,
76
- "I-MT": 28,
77
- "B-TM": 29,
78
- "I-TM": 30
79
  },
80
  "layer_norm_eps": 1e-12,
81
  "max_position_embeddings": 512,
@@ -89,7 +89,7 @@
89
  "summary_type": "first",
90
  "summary_use_proj": true,
91
  "torch_dtype": "float32",
92
- "transformers_version": "4.27.3",
93
  "type_vocab_size": 2,
94
  "use_cache": true,
95
  "vocab_size": 35000
 
10
  "hidden_dropout_prob": 0.1,
11
  "hidden_size": 256,
12
  "id2label": {
13
+ "0": "LABEL_0",
14
+ "1": "LABEL_1",
15
+ "2": "LABEL_2",
16
+ "3": "LABEL_3",
17
+ "4": "LABEL_4",
18
+ "5": "LABEL_5",
19
+ "6": "LABEL_6",
20
+ "7": "LABEL_7",
21
+ "8": "LABEL_8",
22
+ "9": "LABEL_9",
23
+ "10": "LABEL_10",
24
+ "11": "LABEL_11",
25
+ "12": "LABEL_12",
26
+ "13": "LABEL_13",
27
+ "14": "LABEL_14",
28
+ "15": "LABEL_15",
29
+ "16": "LABEL_16",
30
+ "17": "LABEL_17",
31
+ "18": "LABEL_18",
32
+ "19": "LABEL_19",
33
+ "20": "LABEL_20",
34
+ "21": "LABEL_21",
35
+ "22": "LABEL_22",
36
+ "23": "LABEL_23",
37
+ "24": "LABEL_24",
38
+ "25": "LABEL_25",
39
+ "26": "LABEL_26",
40
+ "27": "LABEL_27",
41
+ "28": "LABEL_28",
42
+ "29": "LABEL_29",
43
+ "30": "LABEL_30"
44
  },
45
  "initializer_range": 0.02,
46
  "intermediate_size": 1024,
47
  "label2id": {
48
+ "LABEL_0": 0,
49
+ "LABEL_1": 1,
50
+ "LABEL_10": 10,
51
+ "LABEL_11": 11,
52
+ "LABEL_12": 12,
53
+ "LABEL_13": 13,
54
+ "LABEL_14": 14,
55
+ "LABEL_15": 15,
56
+ "LABEL_16": 16,
57
+ "LABEL_17": 17,
58
+ "LABEL_18": 18,
59
+ "LABEL_19": 19,
60
+ "LABEL_2": 2,
61
+ "LABEL_20": 20,
62
+ "LABEL_21": 21,
63
+ "LABEL_22": 22,
64
+ "LABEL_23": 23,
65
+ "LABEL_24": 24,
66
+ "LABEL_25": 25,
67
+ "LABEL_26": 26,
68
+ "LABEL_27": 27,
69
+ "LABEL_28": 28,
70
+ "LABEL_29": 29,
71
+ "LABEL_3": 3,
72
+ "LABEL_30": 30,
73
+ "LABEL_4": 4,
74
+ "LABEL_5": 5,
75
+ "LABEL_6": 6,
76
+ "LABEL_7": 7,
77
+ "LABEL_8": 8,
78
+ "LABEL_9": 9
79
  },
80
  "layer_norm_eps": 1e-12,
81
  "max_position_embeddings": 512,
 
89
  "summary_type": "first",
90
  "summary_use_proj": true,
91
  "torch_dtype": "float32",
92
+ "transformers_version": "4.27.4",
93
  "type_vocab_size": 2,
94
  "use_cache": true,
95
  "vocab_size": 35000
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:196f63b6da105f5adb41ef9974f065c2b0d5af345569c9e8b8e593b2d678c2e4
3
  size 56330037
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aac05d5a809cbaa447c9c93b4ee5da22a69ad16276eb0da17b649af069d13462
3
  size 56330037
runs/Mar31_08-59-31_efb60e2b93ff/1680253186.4787786/events.out.tfevents.1680253186.efb60e2b93ff.189.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd4562c4558b0410fb23d07cf94fe541330cf935e78d8d011f29ee1290734465
3
+ size 6116
runs/Mar31_08-59-31_efb60e2b93ff/events.out.tfevents.1680253186.efb60e2b93ff.189.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db19cd103b257a69153cad0532d59486e78231f3a63499410571ef2e047fa34a
3
+ size 6052
runs/Mar31_09-24-05_efb60e2b93ff/1680254654.5058205/events.out.tfevents.1680254654.efb60e2b93ff.189.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed63abc0500474d8e6b1acd3f5a65eedd56123312ad4e72cd34f5b39153ca569
3
+ size 5872
runs/Mar31_09-24-05_efb60e2b93ff/events.out.tfevents.1680254654.efb60e2b93ff.189.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eec3d271965d9733856e89f91aff83d6d7c8cbbbd149f96946523783e44fbcca
3
+ size 6121
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_basic_tokenize": true,
4
+ "do_lower_case": false,
5
+ "mask_token": "[MASK]",
6
+ "model_max_length": 512,
7
+ "never_split": null,
8
+ "pad_token": "[PAD]",
9
+ "sep_token": "[SEP]",
10
+ "special_tokens_map_file": null,
11
+ "strip_accents": null,
12
+ "tokenize_chinese_chars": true,
13
+ "tokenizer_class": "ElectraTokenizer",
14
+ "unk_token": "[UNK]"
15
+ }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:81bdda8a167727a8787c89122a38f661d610804181d4798306b156b04be4f10c
3
  size 3643
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e82d2ef065ad6f052607d6ca772d4d1433fafd5d4664dae841beb0ea9ad1f52d
3
  size 3643
vocab.txt ADDED
The diff for this file is too large to render. See raw diff