Akari commited on
Commit
61d2c1b
1 Parent(s): c7de2c4

End of training

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "albert-base-v2",
3
  "architectures": [
4
  "AlbertForQuestionAnswering"
5
  ],
 
1
  {
2
+ "_name_or_path": "test-squad-trained",
3
  "architectures": [
4
  "AlbertForQuestionAnswering"
5
  ],
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0551528a2e8b2b422515bf3a179d4d28f1e0d0046d64ad16c2cdd8344eba3332
3
- size 44393497
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b34e009b733ec0ea576dcc2b14144f065b05d3a6f152aa5ee88be97cb92afe4
3
+ size 44388794
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"do_lower_case": true, "remove_space": true, "keep_accents": false, "bos_token": "[CLS]", "eos_token": "[SEP]", "unk_token": "<unk>", "sep_token": "[SEP]", "pad_token": "<pad>", "cls_token": "[CLS]", "mask_token": {"content": "[MASK]", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "albert-base-v2", "tokenizer_class": "AlbertTokenizer"}
 
1
+ {"do_lower_case": true, "remove_space": true, "keep_accents": false, "bos_token": "[CLS]", "eos_token": "[SEP]", "unk_token": "<unk>", "sep_token": "[SEP]", "pad_token": "<pad>", "cls_token": "[CLS]", "mask_token": {"content": "[MASK]", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "test-squad-trained", "tokenizer_class": "AlbertTokenizer"}
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3fbed5388d2f392294008d29098f2dfe2d037853efa0be720ee8f93c32ee4627
3
- size 2863
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f1f1eacf88a9b520172906c67b4af13d0b4cfcde63d455bcee070b7ba9ca146
3
+ size 2593