pardeepSF commited on
Commit
cec27ac
1 Parent(s): 0e02ac6

previous model

Browse files
config.json CHANGED
@@ -4,27 +4,24 @@
4
  "LayoutLMForQuestionAnswering"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
- "bos_token_id": 0,
8
  "classifier_dropout": null,
9
- "eos_token_id": 2,
10
- "gradient_checkpointing": false,
11
  "hidden_act": "gelu",
12
  "hidden_dropout_prob": 0.1,
13
  "hidden_size": 1024,
14
  "initializer_range": 0.02,
15
  "intermediate_size": 4096,
16
- "layer_norm_eps": 1e-05,
17
  "max_2d_position_embeddings": 1024,
18
- "max_position_embeddings": 514,
19
- "model_type": "roberta",
20
  "num_attention_heads": 16,
21
  "num_hidden_layers": 24,
22
- "pad_token_id": 1,
 
23
  "position_embedding_type": "absolute",
24
- "tokenizer_class": "RobertaTokenizer",
25
  "torch_dtype": "float32",
26
  "transformers_version": "4.23.1",
27
  "type_vocab_size": 2,
28
  "use_cache": true,
29
- "vocab_size": 50265
30
  }
 
4
  "LayoutLMForQuestionAnswering"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
 
7
  "classifier_dropout": null,
 
 
8
  "hidden_act": "gelu",
9
  "hidden_dropout_prob": 0.1,
10
  "hidden_size": 1024,
11
  "initializer_range": 0.02,
12
  "intermediate_size": 4096,
13
+ "layer_norm_eps": 1e-12,
14
  "max_2d_position_embeddings": 1024,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
  "num_attention_heads": 16,
18
  "num_hidden_layers": 24,
19
+ "output_past": true,
20
+ "pad_token_id": 0,
21
  "position_embedding_type": "absolute",
 
22
  "torch_dtype": "float32",
23
  "transformers_version": "4.23.1",
24
  "type_vocab_size": 2,
25
  "use_cache": true,
26
+ "vocab_size": 30522
27
  }
merges.txt DELETED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9a112fb3f782ff14965f87e1916c2557de365b2827fd2fb3c7ebfe26773c3631
3
- size 1438367461
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfc74f8966780254bf142a3eb13696b432812aa8fffbaa21340460090b8c51c2
3
+ size 1357491877
special_tokens_map.json CHANGED
@@ -1,15 +1,7 @@
1
  {
2
- "bos_token": "<s>",
3
- "cls_token": "<s>",
4
- "eos_token": "</s>",
5
- "mask_token": {
6
- "content": "<mask>",
7
- "lstrip": true,
8
- "normalized": false,
9
- "rstrip": false,
10
- "single_word": false
11
- },
12
- "pad_token": "<pad>",
13
- "sep_token": "</s>",
14
- "unk_token": "<unk>"
15
  }
 
1
  {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
 
 
 
 
 
 
 
 
7
  }
tokenizer_config.json CHANGED
@@ -1,64 +1,16 @@
1
  {
2
- "add_prefix_space": true,
3
- "bos_token": {
4
- "__type": "AddedToken",
5
- "content": "<s>",
6
- "lstrip": false,
7
- "normalized": true,
8
- "rstrip": false,
9
- "single_word": false
10
- },
11
- "cls_token": {
12
- "__type": "AddedToken",
13
- "content": "<s>",
14
- "lstrip": false,
15
- "normalized": true,
16
- "rstrip": false,
17
- "single_word": false
18
- },
19
- "eos_token": {
20
- "__type": "AddedToken",
21
- "content": "</s>",
22
- "lstrip": false,
23
- "normalized": true,
24
- "rstrip": false,
25
- "single_word": false
26
- },
27
- "errors": "replace",
28
- "mask_token": {
29
- "__type": "AddedToken",
30
- "content": "<mask>",
31
- "lstrip": true,
32
- "normalized": true,
33
- "rstrip": false,
34
- "single_word": false
35
- },
36
- "model_max_length": 512,
37
- "name_or_path": "/content/docvqa/data/models",
38
- "pad_token": {
39
- "__type": "AddedToken",
40
- "content": "<pad>",
41
- "lstrip": false,
42
- "normalized": true,
43
- "rstrip": false,
44
- "single_word": false
45
- },
46
- "sep_token": {
47
- "__type": "AddedToken",
48
- "content": "</s>",
49
- "lstrip": false,
50
- "normalized": true,
51
- "rstrip": false,
52
- "single_word": false
53
- },
54
- "special_tokens_map_file": null,
55
- "tokenizer_class": "RobertaTokenizer",
56
- "unk_token": {
57
- "__type": "AddedToken",
58
- "content": "<unk>",
59
- "lstrip": false,
60
- "normalized": true,
61
- "rstrip": false,
62
- "single_word": false
63
- }
64
  }
 
1
  {
2
+ "cls_token": "[CLS]",
3
+ "do_basic_tokenize": true,
4
+ "do_lower_case": true,
5
+ "mask_token": "[MASK]",
6
+ "max_len": 512,
7
+ "name_or_path": "microsoft/layoutlm-large-uncased",
8
+ "never_split": null,
9
+ "pad_token": "[PAD]",
10
+ "sep_token": "[SEP]",
11
+ "special_tokens_map_file": "data/models/models--microsoft--layoutlm-large-uncased/snapshots/1e7d50dced3cdfea3a3d63c610e2aab36933dbef/special_tokens_map.json",
12
+ "strip_accents": null,
13
+ "tokenize_chinese_chars": true,
14
+ "tokenizer_class": "BertTokenizer",
15
+ "unk_token": "[UNK]"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b6c73682deb5b30de71e4fd9245521002664b7f5f8d8809380a0ac11e52e5da7
3
  size 1583
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9518cb20ba30d0bc101127dbdfff2574f450d60c3031967b86272a24d6ee0123
3
  size 1583
vocab.json DELETED
The diff for this file is too large to render. See raw diff
 
vocab.txt ADDED
The diff for this file is too large to render. See raw diff