{"paper_url": "https://huggingface.co/papers/2309.09506", "comment": "This is an automated message from the [Librarian Bot](https://huggingface.co/librarian-bots). I found the following papers similar to this paper. \n\nThe following papers were recommended by the Semantic Scholar API \n\n* [DocLLM: A layout-aware generative language model for multimodal document understanding](https://huggingface.co/papers/2401.00908) (2023)\n* [InstructDoc: A Dataset for Zero-Shot Generalization of Visual Document Understanding with Instructions](https://huggingface.co/papers/2401.13313) (2024)\n* [Large Language Models for Generative Information Extraction: A Survey](https://huggingface.co/papers/2312.17617) (2023)\n* [Designing with Language: Wireframing UI Design Intent with Generative Large Language Models](https://huggingface.co/papers/2312.07755) (2023)\n* [UMIE: Unified Multimodal Information Extraction with Instruction Tuning](https://huggingface.co/papers/2401.03082) (2024)\n\n\n Please give a thumbs up to this comment if you found it helpful!\n\n If you want recommendations for any Paper on Hugging Face checkout [this](https://huggingface.co/spaces/librarian-bots/recommend_similar_papers) Space"}