--- library_name: transformers language: en license: mit --- # BART-base-ocr This model is released as part of the paper [Leveraging LLMs for Post-OCR Correction of Historical Newspapers](https://aclanthology.org/2024.lt4hala-1.14/) and designed to correct OCR text. [BART-base](https://huggingface.co/facebook/bart-base) is fine-tuned for post-OCR correction of historical English, using [BLN600](https://aclanthology.org/2024.lrec-main.219/), a parallel corpus of 19th century newspaper machine/human transcription. ## Usage ```python from transformers import AutoTokenizer, AutoModelForSeq2SeqLM, pipeline model = AutoModelForSeq2SeqLM.from_pretrained('pykale/bart-base-ocr') tokenizer = AutoTokenizer.from_pretrained('pykale/bart-base-ocr') generator = pipeline('text2text-generation', model=model.to('cuda'), tokenizer=tokenizer, device='cuda', max_length=1024) ocr = "The defendant wits'fined �5 and costs." pred = generator(ocr)[0]['generated_text'] print(pred) ``` ## Citation ``` @inproceedings{thomas-etal-2024-leveraging, title = "Leveraging {LLM}s for Post-{OCR} Correction of Historical Newspapers", author = "Thomas, Alan and Gaizauskas, Robert and Lu, Haiping", editor = "Sprugnoli, Rachele and Passarotti, Marco", booktitle = "Proceedings of the Third Workshop on Language Technologies for Historical and Ancient Languages (LT4HALA) @ LREC-COLING-2024", month = "may", year = "2024", address = "Torino, Italia", publisher = "ELRA and ICCL", url = "https://aclanthology.org/2024.lt4hala-1.14", pages = "116--121", } ```