clfegg commited on
Commit
ccb118a
1 Parent(s): ee091c9

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +6 -0
  2. hub/.locks/models--vikp--surya_det3/17579df25d3b063dedb036aaca5b495efe5088b8.lock +0 -0
  3. hub/.locks/models--vikp--surya_det3/5a2a74e413345541b7ca0db0cb1d41785649eb99fe6a1b5166aa8bd7c0a8881d.lock +0 -0
  4. hub/.locks/models--vikp--surya_det3/9888778190fd6ecff72bde2ecab5b24cda345851.lock +0 -0
  5. hub/.locks/models--vikp--surya_layout3/2cd406fa7808c997daa69ef095b1992f243dadd7.lock +0 -0
  6. hub/.locks/models--vikp--surya_layout3/64a74e7f804289d76c13f7aab35ad32c6071ccb41db04d8626909c84063971b4.lock +0 -0
  7. hub/.locks/models--vikp--surya_layout3/afaa751c9f40e9df6796a7c2f3d7a9a2159ccae0.lock +0 -0
  8. hub/.locks/models--vikp--surya_order/4566f6f099f2c8ff7426500ff4a7d84a7a8514a0.lock +0 -0
  9. hub/.locks/models--vikp--surya_order/bb3f109ce8d6999ebfda8fbe7940ce7e68ea7e60.lock +0 -0
  10. hub/.locks/models--vikp--surya_order/cda4db4b7f886de50399b061100af28767cf2b71.lock +0 -0
  11. hub/.locks/models--vikp--surya_order/d6688651c37649cc60492f18b5e6bf085f852e0f05e0543b6ca57f25fa884b4a.lock +0 -0
  12. hub/.locks/models--vikp--surya_rec2/5497e8690cfe93cbedec7efaf91f6ac734496ac8.lock +0 -0
  13. hub/.locks/models--vikp--surya_rec2/9a75b64cbeaed06820559bcda4e12c1235de62b5bce787d57cf56a9c3a7123d1.lock +0 -0
  14. hub/.locks/models--vikp--surya_rec2/dd34282c30833587a799d334d44a637694d41c8e.lock +0 -0
  15. hub/.locks/models--vikp--surya_rec2/e237701f4293e736f74d2c968582935590107034.lock +0 -0
  16. hub/.locks/models--vikp--surya_tablerec/59a0e54bb3f1ccc8313ed1a75035435e477773ecc62cc1bb0cee0e5dc58889c8.lock +0 -0
  17. hub/.locks/models--vikp--surya_tablerec/5f777357fc63326274fb93cbedf4948d61a89439.lock +0 -0
  18. hub/.locks/models--vikp--texify/4c48133aabc13f4e5f04badd8214b294a033f85b.lock +0 -0
  19. hub/.locks/models--vikp--texify/614a553895a796faf884b140cae23a04c10eda6b.lock +0 -0
  20. hub/.locks/models--vikp--texify/75dad5a63711d242ad16d0e2a11e194fa073fcce.lock +0 -0
  21. hub/.locks/models--vikp--texify/a782b2f1cdab4d0bacb2dc0f85d02c4b1e31f0bd.lock +0 -0
  22. hub/.locks/models--vikp--texify/aa6fa677ace60c9b55199d5db22a1dce5198c5ac.lock +0 -0
  23. hub/.locks/models--vikp--texify/c4eed27b2bbeb493d7fbba31feda6af5c7527246a35a96d49eaa1010f9c7e9af.lock +0 -0
  24. hub/.locks/models--vikp--texify/cb0af56b5c3710c1f721270799366b1ac33ea76a.lock +0 -0
  25. hub/.locks/models--vikp--texify/ea462a33e5f84b37a56e99a304aec22a89d53670.lock +0 -0
  26. hub/models--vikp--surya_det3/.no_exist/467ee9ec33e6e6c5f73e57dbc1415b14032f5b95/adapter_config.json +0 -0
  27. hub/models--vikp--surya_det3/blobs/17579df25d3b063dedb036aaca5b495efe5088b8 +39 -0
  28. hub/models--vikp--surya_det3/blobs/5a2a74e413345541b7ca0db0cb1d41785649eb99fe6a1b5166aa8bd7c0a8881d +3 -0
  29. hub/models--vikp--surya_det3/blobs/9888778190fd6ecff72bde2ecab5b24cda345851 +52 -0
  30. hub/models--vikp--surya_det3/refs/main +1 -0
  31. hub/models--vikp--surya_det3/snapshots/467ee9ec33e6e6c5f73e57dbc1415b14032f5b95/config.json +52 -0
  32. hub/models--vikp--surya_det3/snapshots/467ee9ec33e6e6c5f73e57dbc1415b14032f5b95/model.safetensors +3 -0
  33. hub/models--vikp--surya_det3/snapshots/467ee9ec33e6e6c5f73e57dbc1415b14032f5b95/preprocessor_config.json +39 -0
  34. hub/models--vikp--surya_layout3/.no_exist/6f4c319b213ddbe736c9fe6de87ad7dc3bbb0a1e/adapter_config.json +0 -0
  35. hub/models--vikp--surya_layout3/blobs/2cd406fa7808c997daa69ef095b1992f243dadd7 +23 -0
  36. hub/models--vikp--surya_layout3/blobs/64a74e7f804289d76c13f7aab35ad32c6071ccb41db04d8626909c84063971b4 +3 -0
  37. hub/models--vikp--surya_layout3/blobs/afaa751c9f40e9df6796a7c2f3d7a9a2159ccae0 +80 -0
  38. hub/models--vikp--surya_layout3/refs/main +1 -0
  39. hub/models--vikp--surya_layout3/snapshots/6f4c319b213ddbe736c9fe6de87ad7dc3bbb0a1e/config.json +80 -0
  40. hub/models--vikp--surya_layout3/snapshots/6f4c319b213ddbe736c9fe6de87ad7dc3bbb0a1e/model.safetensors +3 -0
  41. hub/models--vikp--surya_layout3/snapshots/6f4c319b213ddbe736c9fe6de87ad7dc3bbb0a1e/preprocessor_config.json +23 -0
  42. hub/models--vikp--surya_order/.no_exist/7b727d0a2c942cdc8596d186115f65c12c812bd8/adapter_config.json +0 -0
  43. hub/models--vikp--surya_order/blobs/4566f6f099f2c8ff7426500ff4a7d84a7a8514a0 +199 -0
  44. hub/models--vikp--surya_order/blobs/bb3f109ce8d6999ebfda8fbe7940ce7e68ea7e60 +39 -0
  45. hub/models--vikp--surya_order/blobs/cda4db4b7f886de50399b061100af28767cf2b71 +8 -0
  46. hub/models--vikp--surya_order/blobs/d6688651c37649cc60492f18b5e6bf085f852e0f05e0543b6ca57f25fa884b4a +3 -0
  47. hub/models--vikp--surya_order/refs/main +1 -0
  48. hub/models--vikp--surya_order/snapshots/7b727d0a2c942cdc8596d186115f65c12c812bd8/config.json +199 -0
  49. hub/models--vikp--surya_order/snapshots/7b727d0a2c942cdc8596d186115f65c12c812bd8/generation_config.json +8 -0
  50. hub/models--vikp--surya_order/snapshots/7b727d0a2c942cdc8596d186115f65c12c812bd8/model.safetensors +3 -0
.gitattributes CHANGED
@@ -33,3 +33,9 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ hub/models--vikp--surya_det3/blobs/5a2a74e413345541b7ca0db0cb1d41785649eb99fe6a1b5166aa8bd7c0a8881d filter=lfs diff=lfs merge=lfs -text
37
+ hub/models--vikp--surya_layout3/blobs/64a74e7f804289d76c13f7aab35ad32c6071ccb41db04d8626909c84063971b4 filter=lfs diff=lfs merge=lfs -text
38
+ hub/models--vikp--surya_order/blobs/d6688651c37649cc60492f18b5e6bf085f852e0f05e0543b6ca57f25fa884b4a filter=lfs diff=lfs merge=lfs -text
39
+ hub/models--vikp--surya_rec2/blobs/9a75b64cbeaed06820559bcda4e12c1235de62b5bce787d57cf56a9c3a7123d1 filter=lfs diff=lfs merge=lfs -text
40
+ hub/models--vikp--surya_tablerec/blobs/59a0e54bb3f1ccc8313ed1a75035435e477773ecc62cc1bb0cee0e5dc58889c8 filter=lfs diff=lfs merge=lfs -text
41
+ hub/models--vikp--texify/blobs/c4eed27b2bbeb493d7fbba31feda6af5c7527246a35a96d49eaa1010f9c7e9af filter=lfs diff=lfs merge=lfs -text
hub/.locks/models--vikp--surya_det3/17579df25d3b063dedb036aaca5b495efe5088b8.lock ADDED
File without changes
hub/.locks/models--vikp--surya_det3/5a2a74e413345541b7ca0db0cb1d41785649eb99fe6a1b5166aa8bd7c0a8881d.lock ADDED
File without changes
hub/.locks/models--vikp--surya_det3/9888778190fd6ecff72bde2ecab5b24cda345851.lock ADDED
File without changes
hub/.locks/models--vikp--surya_layout3/2cd406fa7808c997daa69ef095b1992f243dadd7.lock ADDED
File without changes
hub/.locks/models--vikp--surya_layout3/64a74e7f804289d76c13f7aab35ad32c6071ccb41db04d8626909c84063971b4.lock ADDED
File without changes
hub/.locks/models--vikp--surya_layout3/afaa751c9f40e9df6796a7c2f3d7a9a2159ccae0.lock ADDED
File without changes
hub/.locks/models--vikp--surya_order/4566f6f099f2c8ff7426500ff4a7d84a7a8514a0.lock ADDED
File without changes
hub/.locks/models--vikp--surya_order/bb3f109ce8d6999ebfda8fbe7940ce7e68ea7e60.lock ADDED
File without changes
hub/.locks/models--vikp--surya_order/cda4db4b7f886de50399b061100af28767cf2b71.lock ADDED
File without changes
hub/.locks/models--vikp--surya_order/d6688651c37649cc60492f18b5e6bf085f852e0f05e0543b6ca57f25fa884b4a.lock ADDED
File without changes
hub/.locks/models--vikp--surya_rec2/5497e8690cfe93cbedec7efaf91f6ac734496ac8.lock ADDED
File without changes
hub/.locks/models--vikp--surya_rec2/9a75b64cbeaed06820559bcda4e12c1235de62b5bce787d57cf56a9c3a7123d1.lock ADDED
File without changes
hub/.locks/models--vikp--surya_rec2/dd34282c30833587a799d334d44a637694d41c8e.lock ADDED
File without changes
hub/.locks/models--vikp--surya_rec2/e237701f4293e736f74d2c968582935590107034.lock ADDED
File without changes
hub/.locks/models--vikp--surya_tablerec/59a0e54bb3f1ccc8313ed1a75035435e477773ecc62cc1bb0cee0e5dc58889c8.lock ADDED
File without changes
hub/.locks/models--vikp--surya_tablerec/5f777357fc63326274fb93cbedf4948d61a89439.lock ADDED
File without changes
hub/.locks/models--vikp--texify/4c48133aabc13f4e5f04badd8214b294a033f85b.lock ADDED
File without changes
hub/.locks/models--vikp--texify/614a553895a796faf884b140cae23a04c10eda6b.lock ADDED
File without changes
hub/.locks/models--vikp--texify/75dad5a63711d242ad16d0e2a11e194fa073fcce.lock ADDED
File without changes
hub/.locks/models--vikp--texify/a782b2f1cdab4d0bacb2dc0f85d02c4b1e31f0bd.lock ADDED
File without changes
hub/.locks/models--vikp--texify/aa6fa677ace60c9b55199d5db22a1dce5198c5ac.lock ADDED
File without changes
hub/.locks/models--vikp--texify/c4eed27b2bbeb493d7fbba31feda6af5c7527246a35a96d49eaa1010f9c7e9af.lock ADDED
File without changes
hub/.locks/models--vikp--texify/cb0af56b5c3710c1f721270799366b1ac33ea76a.lock ADDED
File without changes
hub/.locks/models--vikp--texify/ea462a33e5f84b37a56e99a304aec22a89d53670.lock ADDED
File without changes
hub/models--vikp--surya_det3/.no_exist/467ee9ec33e6e6c5f73e57dbc1415b14032f5b95/adapter_config.json ADDED
File without changes
hub/models--vikp--surya_det3/blobs/17579df25d3b063dedb036aaca5b495efe5088b8 ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_valid_processor_keys": [
3
+ "images",
4
+ "segmentation_maps",
5
+ "do_resize",
6
+ "size",
7
+ "resample",
8
+ "do_rescale",
9
+ "rescale_factor",
10
+ "do_normalize",
11
+ "image_mean",
12
+ "image_std",
13
+ "do_reduce_labels",
14
+ "return_tensors",
15
+ "data_format",
16
+ "input_data_format"
17
+ ],
18
+ "do_normalize": true,
19
+ "do_reduce_labels": true,
20
+ "do_rescale": true,
21
+ "do_resize": true,
22
+ "image_mean": [
23
+ 0.485,
24
+ 0.456,
25
+ 0.406
26
+ ],
27
+ "image_processor_type": "SegformerImageProcessor",
28
+ "image_std": [
29
+ 0.229,
30
+ 0.224,
31
+ 0.225
32
+ ],
33
+ "resample": 2,
34
+ "rescale_factor": 0.00392156862745098,
35
+ "size": {
36
+ "height": 1200,
37
+ "width": 1200
38
+ }
39
+ }
hub/models--vikp--surya_det3/blobs/5a2a74e413345541b7ca0db0cb1d41785649eb99fe6a1b5166aa8bd7c0a8881d ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a2a74e413345541b7ca0db0cb1d41785649eb99fe6a1b5166aa8bd7c0a8881d
3
+ size 153827656
hub/models--vikp--surya_det3/blobs/9888778190fd6ecff72bde2ecab5b24cda345851 ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "vikp/line_detector_3",
3
+ "architectures": [
4
+ "EfficientViTForSemanticSegmentation"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.0,
7
+ "classifier_dropout_prob": 0.0,
8
+ "decoder_hidden_size": 512,
9
+ "decoder_layer_hidden_size": 128,
10
+ "depths": [
11
+ 1,
12
+ 1,
13
+ 1,
14
+ 6,
15
+ 6
16
+ ],
17
+ "head_dim": 32,
18
+ "hidden_dropout_prob": 0.0,
19
+ "hidden_sizes": [
20
+ 32,
21
+ 64,
22
+ 160,
23
+ 256
24
+ ],
25
+ "initializer_range": 0.02,
26
+ "layer_norm_eps": 1e-06,
27
+ "model_type": "efficientvit",
28
+ "num_channels": 3,
29
+ "num_classes": 2,
30
+ "num_stages": 4,
31
+ "patch_size": [
32
+ 7,
33
+ 7
34
+ ],
35
+ "semantic_loss_ignore_index": -1,
36
+ "strides": [
37
+ 2,
38
+ 2,
39
+ 2,
40
+ 2,
41
+ 2
42
+ ],
43
+ "torch_dtype": "float32",
44
+ "transformers_version": "4.42.3",
45
+ "widths": [
46
+ 32,
47
+ 64,
48
+ 128,
49
+ 256,
50
+ 512
51
+ ]
52
+ }
hub/models--vikp--surya_det3/refs/main ADDED
@@ -0,0 +1 @@
 
 
1
+ 467ee9ec33e6e6c5f73e57dbc1415b14032f5b95
hub/models--vikp--surya_det3/snapshots/467ee9ec33e6e6c5f73e57dbc1415b14032f5b95/config.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "vikp/line_detector_3",
3
+ "architectures": [
4
+ "EfficientViTForSemanticSegmentation"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.0,
7
+ "classifier_dropout_prob": 0.0,
8
+ "decoder_hidden_size": 512,
9
+ "decoder_layer_hidden_size": 128,
10
+ "depths": [
11
+ 1,
12
+ 1,
13
+ 1,
14
+ 6,
15
+ 6
16
+ ],
17
+ "head_dim": 32,
18
+ "hidden_dropout_prob": 0.0,
19
+ "hidden_sizes": [
20
+ 32,
21
+ 64,
22
+ 160,
23
+ 256
24
+ ],
25
+ "initializer_range": 0.02,
26
+ "layer_norm_eps": 1e-06,
27
+ "model_type": "efficientvit",
28
+ "num_channels": 3,
29
+ "num_classes": 2,
30
+ "num_stages": 4,
31
+ "patch_size": [
32
+ 7,
33
+ 7
34
+ ],
35
+ "semantic_loss_ignore_index": -1,
36
+ "strides": [
37
+ 2,
38
+ 2,
39
+ 2,
40
+ 2,
41
+ 2
42
+ ],
43
+ "torch_dtype": "float32",
44
+ "transformers_version": "4.42.3",
45
+ "widths": [
46
+ 32,
47
+ 64,
48
+ 128,
49
+ 256,
50
+ 512
51
+ ]
52
+ }
hub/models--vikp--surya_det3/snapshots/467ee9ec33e6e6c5f73e57dbc1415b14032f5b95/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a2a74e413345541b7ca0db0cb1d41785649eb99fe6a1b5166aa8bd7c0a8881d
3
+ size 153827656
hub/models--vikp--surya_det3/snapshots/467ee9ec33e6e6c5f73e57dbc1415b14032f5b95/preprocessor_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_valid_processor_keys": [
3
+ "images",
4
+ "segmentation_maps",
5
+ "do_resize",
6
+ "size",
7
+ "resample",
8
+ "do_rescale",
9
+ "rescale_factor",
10
+ "do_normalize",
11
+ "image_mean",
12
+ "image_std",
13
+ "do_reduce_labels",
14
+ "return_tensors",
15
+ "data_format",
16
+ "input_data_format"
17
+ ],
18
+ "do_normalize": true,
19
+ "do_reduce_labels": true,
20
+ "do_rescale": true,
21
+ "do_resize": true,
22
+ "image_mean": [
23
+ 0.485,
24
+ 0.456,
25
+ 0.406
26
+ ],
27
+ "image_processor_type": "SegformerImageProcessor",
28
+ "image_std": [
29
+ 0.229,
30
+ 0.224,
31
+ 0.225
32
+ ],
33
+ "resample": 2,
34
+ "rescale_factor": 0.00392156862745098,
35
+ "size": {
36
+ "height": 1200,
37
+ "width": 1200
38
+ }
39
+ }
hub/models--vikp--surya_layout3/.no_exist/6f4c319b213ddbe736c9fe6de87ad7dc3bbb0a1e/adapter_config.json ADDED
File without changes
hub/models--vikp--surya_layout3/blobs/2cd406fa7808c997daa69ef095b1992f243dadd7 ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_normalize": true,
3
+ "do_reduce_labels": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.485,
8
+ 0.456,
9
+ 0.406
10
+ ],
11
+ "image_processor_type": "SegformerImageProcessor",
12
+ "image_std": [
13
+ 0.229,
14
+ 0.224,
15
+ 0.225
16
+ ],
17
+ "resample": 2,
18
+ "rescale_factor": 0.00392156862745098,
19
+ "size": {
20
+ "height": 1024,
21
+ "width": 1024
22
+ }
23
+ }
hub/models--vikp--surya_layout3/blobs/64a74e7f804289d76c13f7aab35ad32c6071ccb41db04d8626909c84063971b4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64a74e7f804289d76c13f7aab35ad32c6071ccb41db04d8626909c84063971b4
3
+ size 153848176
hub/models--vikp--surya_layout3/blobs/afaa751c9f40e9df6796a7c2f3d7a9a2159ccae0 ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "vikp/layout4",
3
+ "architectures": [
4
+ "EfficientViTForSemanticSegmentation"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.0,
7
+ "classifier_dropout_prob": 0.0,
8
+ "decoder_hidden_size": 512,
9
+ "decoder_layer_hidden_size": 128,
10
+ "depths": [
11
+ 1,
12
+ 1,
13
+ 1,
14
+ 6,
15
+ 6
16
+ ],
17
+ "head_dim": 32,
18
+ "hidden_dropout_prob": 0.0,
19
+ "hidden_sizes": [
20
+ 32,
21
+ 64,
22
+ 160,
23
+ 256
24
+ ],
25
+ "id2label": {
26
+ "0": "Blank",
27
+ "1": "Caption",
28
+ "2": "Footnote",
29
+ "3": "Formula",
30
+ "4": "List-item",
31
+ "5": "Page-footer",
32
+ "6": "Page-header",
33
+ "7": "Picture",
34
+ "8": "Section-header",
35
+ "9": "Table",
36
+ "10": "Text",
37
+ "11": "Title"
38
+ },
39
+ "initializer_range": 0.02,
40
+ "label2id": {
41
+ "Blank": 0,
42
+ "Caption": 1,
43
+ "Footnote": 2,
44
+ "Formula": 3,
45
+ "List-item": 4,
46
+ "Page-footer": 5,
47
+ "Page-header": 6,
48
+ "Picture": 7,
49
+ "Section-header": 8,
50
+ "Table": 9,
51
+ "Text": 10,
52
+ "Title": 11
53
+ },
54
+ "layer_norm_eps": 1e-06,
55
+ "model_type": "efficientvit",
56
+ "num_channels": 3,
57
+ "num_classes": 2,
58
+ "num_stages": 4,
59
+ "patch_size": [
60
+ 7,
61
+ 7
62
+ ],
63
+ "semantic_loss_ignore_index": -1,
64
+ "strides": [
65
+ 2,
66
+ 2,
67
+ 2,
68
+ 2,
69
+ 2
70
+ ],
71
+ "torch_dtype": "float32",
72
+ "transformers_version": "4.42.4",
73
+ "widths": [
74
+ 32,
75
+ 64,
76
+ 128,
77
+ 256,
78
+ 512
79
+ ]
80
+ }
hub/models--vikp--surya_layout3/refs/main ADDED
@@ -0,0 +1 @@
 
 
1
+ 6f4c319b213ddbe736c9fe6de87ad7dc3bbb0a1e
hub/models--vikp--surya_layout3/snapshots/6f4c319b213ddbe736c9fe6de87ad7dc3bbb0a1e/config.json ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "vikp/layout4",
3
+ "architectures": [
4
+ "EfficientViTForSemanticSegmentation"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.0,
7
+ "classifier_dropout_prob": 0.0,
8
+ "decoder_hidden_size": 512,
9
+ "decoder_layer_hidden_size": 128,
10
+ "depths": [
11
+ 1,
12
+ 1,
13
+ 1,
14
+ 6,
15
+ 6
16
+ ],
17
+ "head_dim": 32,
18
+ "hidden_dropout_prob": 0.0,
19
+ "hidden_sizes": [
20
+ 32,
21
+ 64,
22
+ 160,
23
+ 256
24
+ ],
25
+ "id2label": {
26
+ "0": "Blank",
27
+ "1": "Caption",
28
+ "2": "Footnote",
29
+ "3": "Formula",
30
+ "4": "List-item",
31
+ "5": "Page-footer",
32
+ "6": "Page-header",
33
+ "7": "Picture",
34
+ "8": "Section-header",
35
+ "9": "Table",
36
+ "10": "Text",
37
+ "11": "Title"
38
+ },
39
+ "initializer_range": 0.02,
40
+ "label2id": {
41
+ "Blank": 0,
42
+ "Caption": 1,
43
+ "Footnote": 2,
44
+ "Formula": 3,
45
+ "List-item": 4,
46
+ "Page-footer": 5,
47
+ "Page-header": 6,
48
+ "Picture": 7,
49
+ "Section-header": 8,
50
+ "Table": 9,
51
+ "Text": 10,
52
+ "Title": 11
53
+ },
54
+ "layer_norm_eps": 1e-06,
55
+ "model_type": "efficientvit",
56
+ "num_channels": 3,
57
+ "num_classes": 2,
58
+ "num_stages": 4,
59
+ "patch_size": [
60
+ 7,
61
+ 7
62
+ ],
63
+ "semantic_loss_ignore_index": -1,
64
+ "strides": [
65
+ 2,
66
+ 2,
67
+ 2,
68
+ 2,
69
+ 2
70
+ ],
71
+ "torch_dtype": "float32",
72
+ "transformers_version": "4.42.4",
73
+ "widths": [
74
+ 32,
75
+ 64,
76
+ 128,
77
+ 256,
78
+ 512
79
+ ]
80
+ }
hub/models--vikp--surya_layout3/snapshots/6f4c319b213ddbe736c9fe6de87ad7dc3bbb0a1e/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64a74e7f804289d76c13f7aab35ad32c6071ccb41db04d8626909c84063971b4
3
+ size 153848176
hub/models--vikp--surya_layout3/snapshots/6f4c319b213ddbe736c9fe6de87ad7dc3bbb0a1e/preprocessor_config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_normalize": true,
3
+ "do_reduce_labels": true,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.485,
8
+ 0.456,
9
+ 0.406
10
+ ],
11
+ "image_processor_type": "SegformerImageProcessor",
12
+ "image_std": [
13
+ 0.229,
14
+ 0.224,
15
+ 0.225
16
+ ],
17
+ "resample": 2,
18
+ "rescale_factor": 0.00392156862745098,
19
+ "size": {
20
+ "height": 1024,
21
+ "width": 1024
22
+ }
23
+ }
hub/models--vikp--surya_order/.no_exist/7b727d0a2c942cdc8596d186115f65c12c812bd8/adapter_config.json ADDED
File without changes
hub/models--vikp--surya_order/blobs/4566f6f099f2c8ff7426500ff4a7d84a7a8514a0 ADDED
@@ -0,0 +1,199 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "vikp/order_hr",
3
+ "architectures": [
4
+ "OrderVisionEncoderDecoderModel"
5
+ ],
6
+ "decoder": {
7
+ "_name_or_path": "",
8
+ "activation_dropout": 0.0,
9
+ "activation_function": "gelu",
10
+ "add_cross_attention": true,
11
+ "add_final_layer_norm": true,
12
+ "architectures": [
13
+ "MBartForCausalLM"
14
+ ],
15
+ "attention_dropout": 0.0,
16
+ "bad_words_ids": null,
17
+ "begin_suppress_tokens": null,
18
+ "bos_token_id": 0,
19
+ "chunk_size_feed_forward": 0,
20
+ "classifier_dropout": 0.0,
21
+ "cross_attention_hidden_size": null,
22
+ "d_model": 1024,
23
+ "decoder_attention_heads": 16,
24
+ "decoder_ffn_dim": 4096,
25
+ "decoder_layerdrop": 0.0,
26
+ "decoder_layers": 3,
27
+ "decoder_start_token_id": null,
28
+ "diversity_penalty": 0.0,
29
+ "do_sample": false,
30
+ "dropout": 0.1,
31
+ "early_stopping": false,
32
+ "encoder_attention_heads": 16,
33
+ "encoder_ffn_dim": 4096,
34
+ "encoder_layerdrop": 0.0,
35
+ "encoder_layers": 12,
36
+ "encoder_no_repeat_ngram_size": 0,
37
+ "eos_token_id": 2,
38
+ "exponential_decay_length_penalty": null,
39
+ "finetuning_task": null,
40
+ "forced_bos_token_id": null,
41
+ "forced_eos_token_id": 2,
42
+ "id2label": {
43
+ "0": "LABEL_0",
44
+ "1": "LABEL_1"
45
+ },
46
+ "init_std": 0.02,
47
+ "is_decoder": true,
48
+ "is_encoder_decoder": false,
49
+ "kv_heads": 4,
50
+ "label2id": {
51
+ "LABEL_0": 0,
52
+ "LABEL_1": 1
53
+ },
54
+ "length_penalty": 1.0,
55
+ "max_height": 1285,
56
+ "max_length": 20,
57
+ "max_position_embeddings": 1536,
58
+ "max_target_positions": 517,
59
+ "max_width": 1285,
60
+ "min_length": 0,
61
+ "model_type": "mbart",
62
+ "no_repeat_ngram_size": 0,
63
+ "num_beam_groups": 1,
64
+ "num_beams": 1,
65
+ "num_hidden_layers": 12,
66
+ "num_return_sequences": 1,
67
+ "output_attentions": false,
68
+ "output_hidden_states": false,
69
+ "output_scores": false,
70
+ "pad_token_id": 1,
71
+ "prefix": null,
72
+ "problem_type": null,
73
+ "pruned_heads": {},
74
+ "remove_invalid_values": false,
75
+ "repetition_penalty": 1.0,
76
+ "return_dict": true,
77
+ "return_dict_in_generate": false,
78
+ "scale_embedding": true,
79
+ "sep_token_id": null,
80
+ "suppress_tokens": null,
81
+ "task_specific_params": null,
82
+ "temperature": 1.0,
83
+ "tf_legacy_loss": false,
84
+ "tie_encoder_decoder": false,
85
+ "tie_word_embeddings": false,
86
+ "tokenizer_class": null,
87
+ "top_k": 50,
88
+ "top_p": 1.0,
89
+ "torch_dtype": "float32",
90
+ "torchscript": false,
91
+ "typical_p": 1.0,
92
+ "use_bfloat16": false,
93
+ "use_cache": true,
94
+ "vocab_size": 261
95
+ },
96
+ "encoder": {
97
+ "_name_or_path": "vikp/donut-encoder",
98
+ "add_cross_attention": false,
99
+ "architectures": [
100
+ "DonutSwinModel"
101
+ ],
102
+ "attention_probs_dropout_prob": 0.0,
103
+ "bad_words_ids": null,
104
+ "begin_suppress_tokens": null,
105
+ "bos_token_id": null,
106
+ "chunk_size_feed_forward": 0,
107
+ "cross_attention_hidden_size": null,
108
+ "decoder_start_token_id": null,
109
+ "depths": [
110
+ 2,
111
+ 2,
112
+ 14,
113
+ 2
114
+ ],
115
+ "diversity_penalty": 0.0,
116
+ "do_sample": false,
117
+ "drop_path_rate": 0.1,
118
+ "early_stopping": false,
119
+ "embed_dim": 128,
120
+ "encoder_no_repeat_ngram_size": 0,
121
+ "eos_token_id": null,
122
+ "exponential_decay_length_penalty": null,
123
+ "finetuning_task": null,
124
+ "forced_bos_token_id": null,
125
+ "forced_eos_token_id": null,
126
+ "hidden_act": "gelu",
127
+ "hidden_dropout_prob": 0.0,
128
+ "hidden_size": 1024,
129
+ "id2label": {
130
+ "0": "LABEL_0",
131
+ "1": "LABEL_1"
132
+ },
133
+ "image_size": [
134
+ 1024,
135
+ 1024
136
+ ],
137
+ "initializer_range": 0.02,
138
+ "is_decoder": false,
139
+ "is_encoder_decoder": false,
140
+ "label2id": {
141
+ "LABEL_0": 0,
142
+ "LABEL_1": 1
143
+ },
144
+ "layer_norm_eps": 1e-05,
145
+ "length_penalty": 1.0,
146
+ "max_length": 20,
147
+ "min_length": 0,
148
+ "mlp_ratio": 4.0,
149
+ "model_type": "donut-swin",
150
+ "no_repeat_ngram_size": 0,
151
+ "num_beam_groups": 1,
152
+ "num_beams": 1,
153
+ "num_channels": 3,
154
+ "num_heads": [
155
+ 4,
156
+ 8,
157
+ 16,
158
+ 32
159
+ ],
160
+ "num_layers": 4,
161
+ "num_return_sequences": 1,
162
+ "output_attentions": false,
163
+ "output_hidden_states": false,
164
+ "output_scores": false,
165
+ "pad_token_id": null,
166
+ "patch_size": 4,
167
+ "path_norm": true,
168
+ "prefix": null,
169
+ "problem_type": null,
170
+ "pruned_heads": {},
171
+ "qkv_bias": true,
172
+ "remove_invalid_values": false,
173
+ "repetition_penalty": 1.0,
174
+ "return_dict": true,
175
+ "return_dict_in_generate": false,
176
+ "sep_token_id": null,
177
+ "suppress_tokens": null,
178
+ "task_specific_params": null,
179
+ "temperature": 1.0,
180
+ "tf_legacy_loss": false,
181
+ "tie_encoder_decoder": false,
182
+ "tie_word_embeddings": true,
183
+ "tokenizer_class": null,
184
+ "top_k": 50,
185
+ "top_p": 1.0,
186
+ "torch_dtype": "float32",
187
+ "torchscript": false,
188
+ "typical_p": 1.0,
189
+ "use_2d_embeddings": false,
190
+ "use_absolute_embeddings": true,
191
+ "use_bfloat16": false,
192
+ "window_size": 7
193
+ },
194
+ "is_encoder_decoder": true,
195
+ "model_type": "vision-encoder-decoder",
196
+ "tie_word_embeddings": false,
197
+ "torch_dtype": "float32",
198
+ "transformers_version": "4.36.2"
199
+ }
hub/models--vikp--surya_order/blobs/bb3f109ce8d6999ebfda8fbe7940ce7e68ea7e60 ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "box_pad_id": 1001,
3
+ "box_size": {
4
+ "height": 1024,
5
+ "width": 1024
6
+ },
7
+ "do_align_long_axis": false,
8
+ "do_normalize": true,
9
+ "do_pad": true,
10
+ "do_reduce_labels": true,
11
+ "do_rescale": true,
12
+ "do_resize": true,
13
+ "do_thumbnail": true,
14
+ "feature_extractor_type": "SegformerFeatureExtractor",
15
+ "image_mean": [
16
+ 0.485,
17
+ 0.456,
18
+ 0.406
19
+ ],
20
+ "image_processor_type": "OrderImageProcessor",
21
+ "image_std": [
22
+ 0.229,
23
+ 0.224,
24
+ 0.225
25
+ ],
26
+ "max_boxes": 255,
27
+ "patch_size": [
28
+ 4,
29
+ 4
30
+ ],
31
+ "resample": 2,
32
+ "rescale_factor": 0.00392156862745098,
33
+ "size": {
34
+ "height": 1024,
35
+ "width": 1024
36
+ },
37
+ "token_pad_id": 1282,
38
+ "token_sep_id": 1281
39
+ }
hub/models--vikp--surya_order/blobs/cda4db4b7f886de50399b061100af28767cf2b71 ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 0,
4
+ "eos_token_id": 2,
5
+ "forced_eos_token_id": 2,
6
+ "pad_token_id": 1,
7
+ "transformers_version": "4.36.2"
8
+ }
hub/models--vikp--surya_order/blobs/d6688651c37649cc60492f18b5e6bf085f852e0f05e0543b6ca57f25fa884b4a ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6688651c37649cc60492f18b5e6bf085f852e0f05e0543b6ca57f25fa884b4a
3
+ size 550040752
hub/models--vikp--surya_order/refs/main ADDED
@@ -0,0 +1 @@
 
 
1
+ 7b727d0a2c942cdc8596d186115f65c12c812bd8
hub/models--vikp--surya_order/snapshots/7b727d0a2c942cdc8596d186115f65c12c812bd8/config.json ADDED
@@ -0,0 +1,199 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "vikp/order_hr",
3
+ "architectures": [
4
+ "OrderVisionEncoderDecoderModel"
5
+ ],
6
+ "decoder": {
7
+ "_name_or_path": "",
8
+ "activation_dropout": 0.0,
9
+ "activation_function": "gelu",
10
+ "add_cross_attention": true,
11
+ "add_final_layer_norm": true,
12
+ "architectures": [
13
+ "MBartForCausalLM"
14
+ ],
15
+ "attention_dropout": 0.0,
16
+ "bad_words_ids": null,
17
+ "begin_suppress_tokens": null,
18
+ "bos_token_id": 0,
19
+ "chunk_size_feed_forward": 0,
20
+ "classifier_dropout": 0.0,
21
+ "cross_attention_hidden_size": null,
22
+ "d_model": 1024,
23
+ "decoder_attention_heads": 16,
24
+ "decoder_ffn_dim": 4096,
25
+ "decoder_layerdrop": 0.0,
26
+ "decoder_layers": 3,
27
+ "decoder_start_token_id": null,
28
+ "diversity_penalty": 0.0,
29
+ "do_sample": false,
30
+ "dropout": 0.1,
31
+ "early_stopping": false,
32
+ "encoder_attention_heads": 16,
33
+ "encoder_ffn_dim": 4096,
34
+ "encoder_layerdrop": 0.0,
35
+ "encoder_layers": 12,
36
+ "encoder_no_repeat_ngram_size": 0,
37
+ "eos_token_id": 2,
38
+ "exponential_decay_length_penalty": null,
39
+ "finetuning_task": null,
40
+ "forced_bos_token_id": null,
41
+ "forced_eos_token_id": 2,
42
+ "id2label": {
43
+ "0": "LABEL_0",
44
+ "1": "LABEL_1"
45
+ },
46
+ "init_std": 0.02,
47
+ "is_decoder": true,
48
+ "is_encoder_decoder": false,
49
+ "kv_heads": 4,
50
+ "label2id": {
51
+ "LABEL_0": 0,
52
+ "LABEL_1": 1
53
+ },
54
+ "length_penalty": 1.0,
55
+ "max_height": 1285,
56
+ "max_length": 20,
57
+ "max_position_embeddings": 1536,
58
+ "max_target_positions": 517,
59
+ "max_width": 1285,
60
+ "min_length": 0,
61
+ "model_type": "mbart",
62
+ "no_repeat_ngram_size": 0,
63
+ "num_beam_groups": 1,
64
+ "num_beams": 1,
65
+ "num_hidden_layers": 12,
66
+ "num_return_sequences": 1,
67
+ "output_attentions": false,
68
+ "output_hidden_states": false,
69
+ "output_scores": false,
70
+ "pad_token_id": 1,
71
+ "prefix": null,
72
+ "problem_type": null,
73
+ "pruned_heads": {},
74
+ "remove_invalid_values": false,
75
+ "repetition_penalty": 1.0,
76
+ "return_dict": true,
77
+ "return_dict_in_generate": false,
78
+ "scale_embedding": true,
79
+ "sep_token_id": null,
80
+ "suppress_tokens": null,
81
+ "task_specific_params": null,
82
+ "temperature": 1.0,
83
+ "tf_legacy_loss": false,
84
+ "tie_encoder_decoder": false,
85
+ "tie_word_embeddings": false,
86
+ "tokenizer_class": null,
87
+ "top_k": 50,
88
+ "top_p": 1.0,
89
+ "torch_dtype": "float32",
90
+ "torchscript": false,
91
+ "typical_p": 1.0,
92
+ "use_bfloat16": false,
93
+ "use_cache": true,
94
+ "vocab_size": 261
95
+ },
96
+ "encoder": {
97
+ "_name_or_path": "vikp/donut-encoder",
98
+ "add_cross_attention": false,
99
+ "architectures": [
100
+ "DonutSwinModel"
101
+ ],
102
+ "attention_probs_dropout_prob": 0.0,
103
+ "bad_words_ids": null,
104
+ "begin_suppress_tokens": null,
105
+ "bos_token_id": null,
106
+ "chunk_size_feed_forward": 0,
107
+ "cross_attention_hidden_size": null,
108
+ "decoder_start_token_id": null,
109
+ "depths": [
110
+ 2,
111
+ 2,
112
+ 14,
113
+ 2
114
+ ],
115
+ "diversity_penalty": 0.0,
116
+ "do_sample": false,
117
+ "drop_path_rate": 0.1,
118
+ "early_stopping": false,
119
+ "embed_dim": 128,
120
+ "encoder_no_repeat_ngram_size": 0,
121
+ "eos_token_id": null,
122
+ "exponential_decay_length_penalty": null,
123
+ "finetuning_task": null,
124
+ "forced_bos_token_id": null,
125
+ "forced_eos_token_id": null,
126
+ "hidden_act": "gelu",
127
+ "hidden_dropout_prob": 0.0,
128
+ "hidden_size": 1024,
129
+ "id2label": {
130
+ "0": "LABEL_0",
131
+ "1": "LABEL_1"
132
+ },
133
+ "image_size": [
134
+ 1024,
135
+ 1024
136
+ ],
137
+ "initializer_range": 0.02,
138
+ "is_decoder": false,
139
+ "is_encoder_decoder": false,
140
+ "label2id": {
141
+ "LABEL_0": 0,
142
+ "LABEL_1": 1
143
+ },
144
+ "layer_norm_eps": 1e-05,
145
+ "length_penalty": 1.0,
146
+ "max_length": 20,
147
+ "min_length": 0,
148
+ "mlp_ratio": 4.0,
149
+ "model_type": "donut-swin",
150
+ "no_repeat_ngram_size": 0,
151
+ "num_beam_groups": 1,
152
+ "num_beams": 1,
153
+ "num_channels": 3,
154
+ "num_heads": [
155
+ 4,
156
+ 8,
157
+ 16,
158
+ 32
159
+ ],
160
+ "num_layers": 4,
161
+ "num_return_sequences": 1,
162
+ "output_attentions": false,
163
+ "output_hidden_states": false,
164
+ "output_scores": false,
165
+ "pad_token_id": null,
166
+ "patch_size": 4,
167
+ "path_norm": true,
168
+ "prefix": null,
169
+ "problem_type": null,
170
+ "pruned_heads": {},
171
+ "qkv_bias": true,
172
+ "remove_invalid_values": false,
173
+ "repetition_penalty": 1.0,
174
+ "return_dict": true,
175
+ "return_dict_in_generate": false,
176
+ "sep_token_id": null,
177
+ "suppress_tokens": null,
178
+ "task_specific_params": null,
179
+ "temperature": 1.0,
180
+ "tf_legacy_loss": false,
181
+ "tie_encoder_decoder": false,
182
+ "tie_word_embeddings": true,
183
+ "tokenizer_class": null,
184
+ "top_k": 50,
185
+ "top_p": 1.0,
186
+ "torch_dtype": "float32",
187
+ "torchscript": false,
188
+ "typical_p": 1.0,
189
+ "use_2d_embeddings": false,
190
+ "use_absolute_embeddings": true,
191
+ "use_bfloat16": false,
192
+ "window_size": 7
193
+ },
194
+ "is_encoder_decoder": true,
195
+ "model_type": "vision-encoder-decoder",
196
+ "tie_word_embeddings": false,
197
+ "torch_dtype": "float32",
198
+ "transformers_version": "4.36.2"
199
+ }
hub/models--vikp--surya_order/snapshots/7b727d0a2c942cdc8596d186115f65c12c812bd8/generation_config.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 0,
4
+ "eos_token_id": 2,
5
+ "forced_eos_token_id": 2,
6
+ "pad_token_id": 1,
7
+ "transformers_version": "4.36.2"
8
+ }
hub/models--vikp--surya_order/snapshots/7b727d0a2c942cdc8596d186115f65c12c812bd8/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6688651c37649cc60492f18b5e6bf085f852e0f05e0543b6ca57f25fa884b4a
3
+ size 550040752