nielsr HF staff commited on
Commit
5c1ca07
β€’
1 Parent(s): 7ae75ea

Add link to docs

Browse files
Files changed (1) hide show
  1. README.md +2 -0
README.md CHANGED
@@ -7,6 +7,8 @@ license: cc-by-nc-sa-4.0
7
  # LayoutLMv2
8
  **Multimodal (text + layout/format + image) pre-training for document AI**
9
 
 
 
10
  [Microsoft Document AI](https://www.microsoft.com/en-us/research/project/document-ai/) | [GitHub](https://github.com/microsoft/unilm/tree/master/layoutlmv2)
11
  ## Introduction
12
  LayoutLMv2 is an improved version of LayoutLM with new pre-training tasks to model the interaction among text, layout, and image in a single multi-modal framework. It outperforms strong baselines and achieves new state-of-the-art results on a wide variety of downstream visually-rich document understanding tasks, including , including FUNSD (0.7895 β†’ 0.8420), CORD (0.9493 β†’ 0.9601), SROIE (0.9524 β†’ 0.9781), Kleister-NDA (0.834 β†’ 0.852), RVL-CDIP (0.9443 β†’ 0.9564), and DocVQA (0.7295 β†’ 0.8672).
 
7
  # LayoutLMv2
8
  **Multimodal (text + layout/format + image) pre-training for document AI**
9
 
10
+ The documentation of this model in the Transformers library can be found [here](https://huggingface.co/docs/transformers/model_doc/layoutlmv2).
11
+
12
  [Microsoft Document AI](https://www.microsoft.com/en-us/research/project/document-ai/) | [GitHub](https://github.com/microsoft/unilm/tree/master/layoutlmv2)
13
  ## Introduction
14
  LayoutLMv2 is an improved version of LayoutLM with new pre-training tasks to model the interaction among text, layout, and image in a single multi-modal framework. It outperforms strong baselines and achieves new state-of-the-art results on a wide variety of downstream visually-rich document understanding tasks, including , including FUNSD (0.7895 β†’ 0.8420), CORD (0.9493 β†’ 0.9601), SROIE (0.9524 β†’ 0.9781), Kleister-NDA (0.834 β†’ 0.852), RVL-CDIP (0.9443 β†’ 0.9564), and DocVQA (0.7295 β†’ 0.8672).