adityaprakhar commited on
Commit
17f69e2
·
verified ·
1 Parent(s): b741397

Training in progress, step 500

Browse files
config.json CHANGED
@@ -1,15 +1,9 @@
1
  {
2
  "_name_or_path": "microsoft/layoutlm-base-uncased",
3
  "architectures": [
4
- "LayoutLMv3ForTokenClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
- "bos_token_id": 0,
8
- "classifier_dropout": null,
9
- "coordinate_size": 128,
10
- "eos_token_id": 2,
11
- "has_relative_attention_bias": true,
12
- "has_spatial_attention_bias": true,
13
  "hidden_act": "gelu",
14
  "hidden_dropout_prob": 0.1,
15
  "hidden_size": 768,
@@ -32,7 +26,6 @@
32
  "15": "address"
33
  },
34
  "initializer_range": 0.02,
35
- "input_size": 224,
36
  "intermediate_size": 3072,
37
  "label2id": {
38
  "SBN": 1,
@@ -55,24 +48,15 @@
55
  "layer_norm_eps": 1e-12,
56
  "max_2d_position_embeddings": 1024,
57
  "max_position_embeddings": 512,
58
- "max_rel_2d_pos": 256,
59
- "max_rel_pos": 128,
60
- "model_type": "layoutlmv3",
61
  "num_attention_heads": 12,
62
- "num_channels": 3,
63
  "num_hidden_layers": 12,
64
  "output_past": true,
65
  "pad_token_id": 0,
66
- "patch_size": 16,
67
  "position_embedding_type": "absolute",
68
- "rel_2d_pos_bins": 64,
69
- "rel_pos_bins": 32,
70
- "shape_size": 128,
71
- "text_embed": true,
72
  "torch_dtype": "float32",
73
  "transformers_version": "4.39.0.dev0",
74
  "type_vocab_size": 2,
75
  "use_cache": true,
76
- "visual_embed": true,
77
  "vocab_size": 30522
78
  }
 
1
  {
2
  "_name_or_path": "microsoft/layoutlm-base-uncased",
3
  "architectures": [
4
+ "LayoutLMForTokenClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
 
 
 
 
 
 
7
  "hidden_act": "gelu",
8
  "hidden_dropout_prob": 0.1,
9
  "hidden_size": 768,
 
26
  "15": "address"
27
  },
28
  "initializer_range": 0.02,
 
29
  "intermediate_size": 3072,
30
  "label2id": {
31
  "SBN": 1,
 
48
  "layer_norm_eps": 1e-12,
49
  "max_2d_position_embeddings": 1024,
50
  "max_position_embeddings": 512,
51
+ "model_type": "layoutlm",
 
 
52
  "num_attention_heads": 12,
 
53
  "num_hidden_layers": 12,
54
  "output_past": true,
55
  "pad_token_id": 0,
 
56
  "position_embedding_type": "absolute",
 
 
 
 
57
  "torch_dtype": "float32",
58
  "transformers_version": "4.39.0.dev0",
59
  "type_vocab_size": 2,
60
  "use_cache": true,
 
61
  "vocab_size": 30522
62
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:81a497fbae4db76bf633d0c6179841f4951588dfdb5cd15f671fe6e34121d89c
3
- size 443092232
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1911c5c593ef49f3e167d257fd2dba516a7e388a14aae8fd26211be5d53e347
3
+ size 450585896
runs/Mar13_06-11-13_bfebc32db449/events.out.tfevents.1710310540.bfebc32db449.363.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fea88a01ea4126fb29e123dfb70aedd595b4726d6c2a33925af2b85c3a211e77
3
+ size 560
runs/Mar13_06-31-41_bfebc32db449/events.out.tfevents.1710311502.bfebc32db449.363.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5dc67db4dd53581a5ef7dbd91339b06b9c72ea625464cfb69c5784b3d843a93d
3
+ size 5435
runs/Mar13_06-32-27_bfebc32db449/events.out.tfevents.1710311547.bfebc32db449.363.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db4de0dd3fb4a77ca13973d265d6aeea38eabb3128a96d28cd657e96bd55aaac
3
+ size 10702
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f6f0b84f08d7b979070db356c4d7052a12fc971f626f5fce7f6d3826c779b461
3
  size 4856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b55c693582b0897226132509bd587c6fb64a2195b8d4ede340833ede316d59b1
3
  size 4856