jnferfer commited on
Commit
8d7e518
1 Parent(s): 35f6981

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -1,11 +1,11 @@
1
  {
2
- "_name_or_path": "bert-base-uncased",
3
  "architectures": [
4
  "BertForTokenClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "classifier_dropout": null,
8
- "gradient_checkpointing": false,
9
  "hidden_act": "gelu",
10
  "hidden_dropout_prob": 0.1,
11
  "hidden_size": 768,
@@ -27,10 +27,15 @@
27
  "num_attention_heads": 12,
28
  "num_hidden_layers": 12,
29
  "pad_token_id": 0,
 
 
 
 
 
30
  "position_embedding_type": "absolute",
31
  "torch_dtype": "float32",
32
  "transformers_version": "4.41.2",
33
  "type_vocab_size": 2,
34
  "use_cache": true,
35
- "vocab_size": 30522
36
  }
 
1
  {
2
+ "_name_or_path": "google-bert/bert-base-multilingual-cased",
3
  "architectures": [
4
  "BertForTokenClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "classifier_dropout": null,
8
+ "directionality": "bidi",
9
  "hidden_act": "gelu",
10
  "hidden_dropout_prob": 0.1,
11
  "hidden_size": 768,
 
27
  "num_attention_heads": 12,
28
  "num_hidden_layers": 12,
29
  "pad_token_id": 0,
30
+ "pooler_fc_size": 768,
31
+ "pooler_num_attention_heads": 12,
32
+ "pooler_num_fc_layers": 3,
33
+ "pooler_size_per_head": 128,
34
+ "pooler_type": "first_token_transform",
35
  "position_embedding_type": "absolute",
36
  "torch_dtype": "float32",
37
  "transformers_version": "4.41.2",
38
  "type_vocab_size": 2,
39
  "use_cache": true,
40
+ "vocab_size": 119547
41
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4879168973532b5e903cf9355e2a40ce3594b540e420e15361ae1ee14e27d94b
3
- size 435599164
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:044dfd32818cddb73403f4ab283af1a494029e8bf4d4edd7a5d35149938f8f69
3
+ size 709083980
runs/Jun20_15-32-23_1d781fe7715a/events.out.tfevents.1718897544.1d781fe7715a.1519.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54c46fb69b5e25d5be55ea18bfb0deb0592fa73f8ed928181e21c31477cc70ab
3
+ size 5604
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -43,7 +43,7 @@
43
  },
44
  "clean_up_tokenization_spaces": true,
45
  "cls_token": "[CLS]",
46
- "do_lower_case": true,
47
  "mask_token": "[MASK]",
48
  "model_max_length": 512,
49
  "pad_token": "[PAD]",
 
43
  },
44
  "clean_up_tokenization_spaces": true,
45
  "cls_token": "[CLS]",
46
+ "do_lower_case": false,
47
  "mask_token": "[MASK]",
48
  "model_max_length": 512,
49
  "pad_token": "[PAD]",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3e654d508b821089e26b47094cc0d751e59bf572712b3884103c34a8f0e94744
3
  size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1cb377bf37f9f5c2fc32872f354f8168c65fb486366d07b22bc0030e13a97e34
3
  size 5112
vocab.txt CHANGED
The diff for this file is too large to render. See raw diff