Benedict-L commited on
Commit
9a9e438
1 Parent(s): bea9526

End of training

Browse files
README.md CHANGED
@@ -4,7 +4,7 @@ base_model: microsoft/layoutlm-base-uncased
4
  tags:
5
  - generated_from_trainer
6
  datasets:
7
- - funsd
8
  model-index:
9
  - name: layoutlm-funsd1
10
  results: []
@@ -15,16 +15,16 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # layoutlm-funsd1
17
 
18
- This model is a fine-tuned version of [microsoft/layoutlm-base-uncased](https://huggingface.co/microsoft/layoutlm-base-uncased) on the funsd dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.6667
21
- - Answer: {'precision': 0.6578947368421053, 'recall': 0.7725587144622992, 'f1': 0.7106310403638432, 'number': 809}
22
- - Header: {'precision': 0.19658119658119658, 'recall': 0.19327731092436976, 'f1': 0.19491525423728814, 'number': 119}
23
- - Question: {'precision': 0.7215958369470945, 'recall': 0.7812206572769953, 'f1': 0.7502254283137962, 'number': 1065}
24
- - Overall Precision: 0.6667
25
- - Overall Recall: 0.7426
26
- - Overall F1: 0.7026
27
- - Overall Accuracy: 0.7964
28
 
29
  ## Model description
30
 
@@ -54,18 +54,18 @@ The following hyperparameters were used during training:
54
 
55
  ### Training results
56
 
57
- | Training Loss | Epoch | Step | Validation Loss | Answer | Header | Question | Overall Precision | Overall Recall | Overall F1 | Overall Accuracy |
58
- |:-------------:|:-----:|:----:|:---------------:|:-----------------------------------------------------------------------------------------------------------:|:------------------------------------------------------------------------------------------------------------:|:------------------------------------------------------------------------------------------------------------:|:-----------------:|:--------------:|:----------:|:----------------:|
59
- | 1.7693 | 1.0 | 10 | 1.5725 | {'precision': 0.03488372093023256, 'recall': 0.0407911001236094, 'f1': 0.037606837606837605, 'number': 809} | {'precision': 0.0, 'recall': 0.0, 'f1': 0.0, 'number': 119} | {'precision': 0.20363636363636364, 'recall': 0.21032863849765257, 'f1': 0.20692840646651267, 'number': 1065} | 0.1256 | 0.1290 | 0.1273 | 0.3991 |
60
- | 1.425 | 2.0 | 20 | 1.2448 | {'precision': 0.12746386333771353, 'recall': 0.11990111248454882, 'f1': 0.1235668789808917, 'number': 809} | {'precision': 0.0, 'recall': 0.0, 'f1': 0.0, 'number': 119} | {'precision': 0.44836716681376876, 'recall': 0.47699530516431926, 'f1': 0.46223839854413107, 'number': 1065} | 0.3194 | 0.3036 | 0.3113 | 0.5601 |
61
- | 1.1125 | 3.0 | 30 | 0.9760 | {'precision': 0.43318485523385303, 'recall': 0.48084054388133496, 'f1': 0.45577035735207966, 'number': 809} | {'precision': 0.06060606060606061, 'recall': 0.01680672268907563, 'f1': 0.02631578947368421, 'number': 119} | {'precision': 0.6073674752920036, 'recall': 0.6347417840375587, 'f1': 0.620752984389348, 'number': 1065} | 0.5220 | 0.5354 | 0.5286 | 0.6992 |
62
- | 0.8731 | 4.0 | 40 | 0.7844 | {'precision': 0.5927835051546392, 'recall': 0.7107540173053152, 'f1': 0.6464305789769533, 'number': 809} | {'precision': 0.12280701754385964, 'recall': 0.058823529411764705, 'f1': 0.07954545454545454, 'number': 119} | {'precision': 0.6381909547738693, 'recall': 0.7154929577464789, 'f1': 0.6746347941567065, 'number': 1065} | 0.6051 | 0.6744 | 0.6379 | 0.7573 |
63
- | 0.6964 | 5.0 | 50 | 0.7420 | {'precision': 0.6131868131868132, 'recall': 0.6897404202719407, 'f1': 0.6492146596858639, 'number': 809} | {'precision': 0.17857142857142858, 'recall': 0.12605042016806722, 'f1': 0.14778325123152708, 'number': 119} | {'precision': 0.6419951729686243, 'recall': 0.7492957746478873, 'f1': 0.6915077989601386, 'number': 1065} | 0.6129 | 0.6879 | 0.6482 | 0.7719 |
64
- | 0.6156 | 6.0 | 60 | 0.7064 | {'precision': 0.6271008403361344, 'recall': 0.7379480840543882, 'f1': 0.678023850085179, 'number': 809} | {'precision': 0.24, 'recall': 0.15126050420168066, 'f1': 0.18556701030927833, 'number': 119} | {'precision': 0.6932409012131716, 'recall': 0.7511737089201878, 'f1': 0.7210455159981973, 'number': 1065} | 0.6488 | 0.7100 | 0.6780 | 0.7780 |
65
- | 0.5557 | 7.0 | 70 | 0.6802 | {'precision': 0.6476793248945147, 'recall': 0.7589616810877626, 'f1': 0.6989186112692088, 'number': 809} | {'precision': 0.22105263157894736, 'recall': 0.17647058823529413, 'f1': 0.19626168224299065, 'number': 119} | {'precision': 0.7050298380221653, 'recall': 0.7765258215962442, 'f1': 0.7390527256479, 'number': 1065} | 0.6597 | 0.7336 | 0.6947 | 0.7915 |
66
- | 0.5151 | 8.0 | 80 | 0.6709 | {'precision': 0.6634920634920635, 'recall': 0.7750309023485785, 'f1': 0.7149372862029646, 'number': 809} | {'precision': 0.2072072072072072, 'recall': 0.19327731092436976, 'f1': 0.2, 'number': 119} | {'precision': 0.7220756376429199, 'recall': 0.7708920187793428, 'f1': 0.7456857402361489, 'number': 1065} | 0.6708 | 0.7381 | 0.7028 | 0.7936 |
67
- | 0.4746 | 9.0 | 90 | 0.6726 | {'precision': 0.6552462526766595, 'recall': 0.7564894932014833, 'f1': 0.7022375215146299, 'number': 809} | {'precision': 0.21621621621621623, 'recall': 0.20168067226890757, 'f1': 0.20869565217391306, 'number': 119} | {'precision': 0.7148900169204738, 'recall': 0.7934272300469484, 'f1': 0.7521139296840232, 'number': 1065} | 0.6650 | 0.7431 | 0.7019 | 0.7949 |
68
- | 0.4849 | 10.0 | 100 | 0.6667 | {'precision': 0.6578947368421053, 'recall': 0.7725587144622992, 'f1': 0.7106310403638432, 'number': 809} | {'precision': 0.19658119658119658, 'recall': 0.19327731092436976, 'f1': 0.19491525423728814, 'number': 119} | {'precision': 0.7215958369470945, 'recall': 0.7812206572769953, 'f1': 0.7502254283137962, 'number': 1065} | 0.6667 | 0.7426 | 0.7026 | 0.7964 |
69
 
70
 
71
  ### Framework versions
 
4
  tags:
5
  - generated_from_trainer
6
  datasets:
7
+ - my_funsd
8
  model-index:
9
  - name: layoutlm-funsd1
10
  results: []
 
15
 
16
  # layoutlm-funsd1
17
 
18
+ This model is a fine-tuned version of [microsoft/layoutlm-base-uncased](https://huggingface.co/microsoft/layoutlm-base-uncased) on the my_funsd dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 1.0974
21
+ - Answer: {'precision': 0.5384615384615384, 'recall': 0.4666666666666667, 'f1': 0.5, 'number': 15}
22
+ - Header: {'precision': 0.0, 'recall': 0.0, 'f1': 0.0, 'number': 1}
23
+ - Question: {'precision': 0.3333333333333333, 'recall': 0.13333333333333333, 'f1': 0.19047619047619044, 'number': 15}
24
+ - Overall Precision: 0.4737
25
+ - Overall Recall: 0.2903
26
+ - Overall F1: 0.36
27
+ - Overall Accuracy: 0.6937
28
 
29
  ## Model description
30
 
 
54
 
55
  ### Training results
56
 
57
+ | Training Loss | Epoch | Step | Validation Loss | Answer | Header | Question | Overall Precision | Overall Recall | Overall F1 | Overall Accuracy |
58
+ |:-------------:|:-----:|:----:|:---------------:|:--------------------------------------------------------------------------------------------------------:|:-----------------------------------------------------------------------------------------:|:----------------------------------------------------------------------------------------------------------:|:-----------------:|:--------------:|:----------:|:----------------:|
59
+ | 1.9267 | 1.0 | 1 | 1.9079 | {'precision': 0.3125, 'recall': 0.3333333333333333, 'f1': 0.3225806451612903, 'number': 15} | {'precision': 0.07692307692307693, 'recall': 1.0, 'f1': 0.14285714285714288, 'number': 1} | {'precision': 0.21052631578947367, 'recall': 0.26666666666666666, 'f1': 0.23529411764705882, 'number': 15} | 0.2083 | 0.3226 | 0.2532 | 0.2523 |
60
+ | 1.9277 | 2.0 | 2 | 1.7357 | {'precision': 0.42857142857142855, 'recall': 0.4, 'f1': 0.4137931034482759, 'number': 15} | {'precision': 0.0, 'recall': 0.0, 'f1': 0.0, 'number': 1} | {'precision': 0.2222222222222222, 'recall': 0.13333333333333333, 'f1': 0.16666666666666669, 'number': 15} | 0.3333 | 0.2581 | 0.2909 | 0.4144 |
61
+ | 1.7536 | 3.0 | 3 | 1.5947 | {'precision': 0.3333333333333333, 'recall': 0.26666666666666666, 'f1': 0.2962962962962963, 'number': 15} | {'precision': 0.0, 'recall': 0.0, 'f1': 0.0, 'number': 1} | {'precision': 0.0, 'recall': 0.0, 'f1': 0.0, 'number': 15} | 0.2353 | 0.1290 | 0.1667 | 0.4955 |
62
+ | 1.6284 | 4.0 | 4 | 1.4767 | {'precision': 0.3333333333333333, 'recall': 0.26666666666666666, 'f1': 0.2962962962962963, 'number': 15} | {'precision': 0.0, 'recall': 0.0, 'f1': 0.0, 'number': 1} | {'precision': 0.0, 'recall': 0.0, 'f1': 0.0, 'number': 15} | 0.2667 | 0.1290 | 0.1739 | 0.5586 |
63
+ | 1.5334 | 5.0 | 5 | 1.3768 | {'precision': 0.4166666666666667, 'recall': 0.3333333333333333, 'f1': 0.3703703703703704, 'number': 15} | {'precision': 0.0, 'recall': 0.0, 'f1': 0.0, 'number': 1} | {'precision': 0.0, 'recall': 0.0, 'f1': 0.0, 'number': 15} | 0.3125 | 0.1613 | 0.2128 | 0.6036 |
64
+ | 1.4284 | 6.0 | 6 | 1.2919 | {'precision': 0.5454545454545454, 'recall': 0.4, 'f1': 0.4615384615384615, 'number': 15} | {'precision': 0.0, 'recall': 0.0, 'f1': 0.0, 'number': 1} | {'precision': 0.0, 'recall': 0.0, 'f1': 0.0, 'number': 15} | 0.375 | 0.1935 | 0.2553 | 0.6216 |
65
+ | 1.3646 | 7.0 | 7 | 1.2220 | {'precision': 0.5454545454545454, 'recall': 0.4, 'f1': 0.4615384615384615, 'number': 15} | {'precision': 0.0, 'recall': 0.0, 'f1': 0.0, 'number': 1} | {'precision': 0.0, 'recall': 0.0, 'f1': 0.0, 'number': 15} | 0.375 | 0.1935 | 0.2553 | 0.6577 |
66
+ | 1.3005 | 8.0 | 8 | 1.1665 | {'precision': 0.5384615384615384, 'recall': 0.4666666666666667, 'f1': 0.5, 'number': 15} | {'precision': 0.0, 'recall': 0.0, 'f1': 0.0, 'number': 1} | {'precision': 0.0, 'recall': 0.0, 'f1': 0.0, 'number': 15} | 0.3889 | 0.2258 | 0.2857 | 0.6667 |
67
+ | 1.2501 | 9.0 | 9 | 1.1250 | {'precision': 0.5384615384615384, 'recall': 0.4666666666666667, 'f1': 0.5, 'number': 15} | {'precision': 0.0, 'recall': 0.0, 'f1': 0.0, 'number': 1} | {'precision': 0.16666666666666666, 'recall': 0.06666666666666667, 'f1': 0.09523809523809522, 'number': 15} | 0.4211 | 0.2581 | 0.3200 | 0.6847 |
68
+ | 1.1952 | 10.0 | 10 | 1.0974 | {'precision': 0.5384615384615384, 'recall': 0.4666666666666667, 'f1': 0.5, 'number': 15} | {'precision': 0.0, 'recall': 0.0, 'f1': 0.0, 'number': 1} | {'precision': 0.3333333333333333, 'recall': 0.13333333333333333, 'f1': 0.19047619047619044, 'number': 15} | 0.4737 | 0.2903 | 0.36 | 0.6937 |
69
 
70
 
71
  ### Framework versions
logs/events.out.tfevents.1719559201.HCIDC-SV-DMZ-ORC-NODE02.3119508.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:265d131baa66b074bfbb3ca5c1aa37cbfdf53d6d6f03a1a78de6cc735effc760
3
+ size 5065
logs/events.out.tfevents.1719559498.HCIDC-SV-DMZ-ORC-NODE02.3131463.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:422020639838df5a79d70de94304aca3a7001067e51eb6ec021dec1a288d9c30
3
+ size 12433
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:96a7188ae1eb809a579162896aea6cb330761dfb502201e012b3de35ca9963d1
3
  size 450558212
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a8a4c95d9978d5a27493698dd2d0f04095e9531b12b460823aebf0f3e217c5e
3
  size 450558212