oldhu commited on
Commit
86148c8
1 Parent(s): 18b2f73

Upload 7 files

Browse files
config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "./models/models--hfl--chinese-lert-base/snapshots/d579de41695f52b4451ebe0e087a3cfa2b374e7e",
3
+ "architectures": [
4
+ "BertModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "directionality": "bidi",
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "pooler_fc_size": 768,
21
+ "pooler_num_attention_heads": 12,
22
+ "pooler_num_fc_layers": 3,
23
+ "pooler_size_per_head": 128,
24
+ "pooler_type": "first_token_transform",
25
+ "position_embedding_type": "absolute",
26
+ "torch_dtype": "float32",
27
+ "transformers_version": "4.29.1",
28
+ "type_vocab_size": 2,
29
+ "use_cache": true,
30
+ "vocab_size": 21128
31
+ }
eval_results.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ eval_pearson = 0.5167570048030995
2
+ eval_spearman = 0.5799801031488225
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "clean_up_tokenization_spaces": true,
3
+ "cls_token": "[CLS]",
4
+ "do_lower_case": true,
5
+ "mask_token": "[MASK]",
6
+ "model_max_length": 1000000000000000019884624838656,
7
+ "pad_token": "[PAD]",
8
+ "sep_token": "[SEP]",
9
+ "strip_accents": null,
10
+ "tokenize_chinese_chars": true,
11
+ "tokenizer_class": "BertTokenizer",
12
+ "unk_token": "[UNK]"
13
+ }
training_progress_scores.csv ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ global_step,train_loss,eval_spearman,eval_pearson
2
+ 14247,0.0,0.44640911117513316,0.37204173786267986
3
+ 28494,0.0,0.5368602869667283,0.46418897883549226
4
+ 42741,0.0,0.5363432540335564,0.4693935695598132
5
+ 56988,0.0,0.5524778906429818,0.48711551165474787
6
+ 71235,0.0,0.5601294876480567,0.49374641851065676
7
+ 85482,0.0,0.5642261647874868,0.4953772027897005
8
+ 99729,0.0,0.5664328609772676,0.5015888844280011
9
+ 113976,0.0,0.5683871881340451,0.5057241284359141
10
+ 128223,0.0,0.5732890915719637,0.5113278130411432
11
+ 142470,0.0,0.5799801031488225,0.5167570048030995
vocab.txt ADDED
The diff for this file is too large to render. See raw diff