Training in progress, epoch 1
Browse files- .gitignore +1 -0
- config.json +37 -0
- pytorch_model.bin +3 -0
- runs/Apr02_08-52-49_f6b96d503954/1680425836.3309686/events.out.tfevents.1680425836.f6b96d503954.137.1 +3 -0
- runs/Apr02_08-52-49_f6b96d503954/events.out.tfevents.1680425836.f6b96d503954.137.0 +3 -0
- runs/Apr02_09-08-05_f6b96d503954/1680426494.160478/events.out.tfevents.1680426494.f6b96d503954.137.3 +3 -0
- runs/Apr02_09-08-05_f6b96d503954/events.out.tfevents.1680426494.f6b96d503954.137.2 +3 -0
- runs/Apr02_09-58-03_f6b96d503954/1680429490.1736627/events.out.tfevents.1680429490.f6b96d503954.137.7 +3 -0
- runs/Apr02_09-58-03_f6b96d503954/events.out.tfevents.1680429490.f6b96d503954.137.6 +3 -0
- runs/Apr02_10-01-15_f6b96d503954/1680429682.2514462/events.out.tfevents.1680429682.f6b96d503954.137.9 +3 -0
- runs/Apr02_10-01-15_f6b96d503954/events.out.tfevents.1680429682.f6b96d503954.137.8 +3 -0
- special_tokens_map.json +7 -0
- tokenizer.json +0 -0
- tokenizer_config.json +13 -0
- training_args.bin +3 -0
- vocab.txt +0 -0
.gitignore
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
checkpoint-*/
|
config.json
ADDED
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "bert-base-uncased",
|
3 |
+
"architectures": [
|
4 |
+
"BertForSequenceClassification"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0.1,
|
7 |
+
"classifier_dropout": null,
|
8 |
+
"gradient_checkpointing": false,
|
9 |
+
"hidden_act": "gelu",
|
10 |
+
"hidden_dropout_prob": 0.1,
|
11 |
+
"hidden_size": 768,
|
12 |
+
"id2label": {
|
13 |
+
"0": "LABEL_0",
|
14 |
+
"1": "LABEL_1",
|
15 |
+
"2": "LABEL_2"
|
16 |
+
},
|
17 |
+
"initializer_range": 0.02,
|
18 |
+
"intermediate_size": 3072,
|
19 |
+
"label2id": {
|
20 |
+
"LABEL_0": 0,
|
21 |
+
"LABEL_1": 1,
|
22 |
+
"LABEL_2": 2
|
23 |
+
},
|
24 |
+
"layer_norm_eps": 1e-12,
|
25 |
+
"max_position_embeddings": 512,
|
26 |
+
"model_type": "bert",
|
27 |
+
"num_attention_heads": 12,
|
28 |
+
"num_hidden_layers": 12,
|
29 |
+
"pad_token_id": 0,
|
30 |
+
"position_embedding_type": "absolute",
|
31 |
+
"problem_type": "single_label_classification",
|
32 |
+
"torch_dtype": "float32",
|
33 |
+
"transformers_version": "4.27.4",
|
34 |
+
"type_vocab_size": 2,
|
35 |
+
"use_cache": true,
|
36 |
+
"vocab_size": 30522
|
37 |
+
}
|
pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2dbd79438c8cdda722ab3b49edbabbf1d077aa39317ad2fbafa712ab54344e60
|
3 |
+
size 438010997
|
runs/Apr02_08-52-49_f6b96d503954/1680425836.3309686/events.out.tfevents.1680425836.f6b96d503954.137.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7621c316a23243390bce826763f7f07c2a7def795393f7efce40c24c01daca1f
|
3 |
+
size 5799
|
runs/Apr02_08-52-49_f6b96d503954/events.out.tfevents.1680425836.f6b96d503954.137.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:33fdf11e303e7ba5b5221a94ca21a221fb16acc5567a1dd9791a5889e7595378
|
3 |
+
size 3937
|
runs/Apr02_09-08-05_f6b96d503954/1680426494.160478/events.out.tfevents.1680426494.f6b96d503954.137.3
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1a19046ade7a18f41ed9713bc07e3142d040f331a6b1e94930f2f2ebdfb5cc0d
|
3 |
+
size 5799
|
runs/Apr02_09-08-05_f6b96d503954/events.out.tfevents.1680426494.f6b96d503954.137.2
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1943768fdb119669a5ae4fc9832a9635b3bd76203aecec14608b81e9ef32456a
|
3 |
+
size 4091
|
runs/Apr02_09-58-03_f6b96d503954/1680429490.1736627/events.out.tfevents.1680429490.f6b96d503954.137.7
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:326b1b01cce6e4e9f1fc73dfa0867ee5a6e86aed5a685dc2a5b5b660a79c9afc
|
3 |
+
size 5799
|
runs/Apr02_09-58-03_f6b96d503954/events.out.tfevents.1680429490.f6b96d503954.137.6
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:226d78d275a50382bfd90f9dd4a5371bb12f331b28e32d21a63ff058da92dc8b
|
3 |
+
size 4094
|
runs/Apr02_10-01-15_f6b96d503954/1680429682.2514462/events.out.tfevents.1680429682.f6b96d503954.137.9
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:37543753a9590acc6bbb96840f4f3c2c8f17d69834f905cff10d3e04cf41b350
|
3 |
+
size 5799
|
runs/Apr02_10-01-15_f6b96d503954/events.out.tfevents.1680429682.f6b96d503954.137.8
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f41382e2f6384b75d7fb5d767417a0bd04e0532c0373380fca1931b8e47a2cd5
|
3 |
+
size 4405
|
special_tokens_map.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"cls_token": "[CLS]",
|
3 |
+
"mask_token": "[MASK]",
|
4 |
+
"pad_token": "[PAD]",
|
5 |
+
"sep_token": "[SEP]",
|
6 |
+
"unk_token": "[UNK]"
|
7 |
+
}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"cls_token": "[CLS]",
|
3 |
+
"do_lower_case": true,
|
4 |
+
"mask_token": "[MASK]",
|
5 |
+
"model_max_length": 512,
|
6 |
+
"pad_token": "[PAD]",
|
7 |
+
"sep_token": "[SEP]",
|
8 |
+
"special_tokens_map_file": null,
|
9 |
+
"strip_accents": null,
|
10 |
+
"tokenize_chinese_chars": true,
|
11 |
+
"tokenizer_class": "BertTokenizer",
|
12 |
+
"unk_token": "[UNK]"
|
13 |
+
}
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:63488ebfee2f6b32b1d79ca8a1eec071025dd127bbac2872fdee419fcf674a8a
|
3 |
+
size 3579
|
vocab.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|