--- thumbnail: https://huggingface.co/front/thumbnails/lladoc.png language: - en license: cc-by-4.0 tags: - transformers datasets: - idocvqa metrics: - accuracy --- ## LLaDoc (Large Language and Document) model This is a fine-tuned model of LLaVA1.5 (7B) on the iDocVQA dataset. It is intended to be used as a multimodal system. The dataset it's trained on is limited in scope, as it covers only certain domains. The accuracy achieved on the validation set is 29.58%. Please find the information about preprocessing, training and full details of the LLaVA model in the [original link](https://llava-vl.github.io/) The paper for this work is available on arXiv: [https://arxiv.org/abs/2402.00453](https://arxiv.org/abs/2402.00453)