--- license: mit --- # Model description LegalBert is a BERT-base-cased model fine-tuned on a subset of the `case.law` corpus. Further details can be found in this paper: [A Dataset for Statutory Reasoning in Tax Law Entailment and Question Answering](http://ceur-ws.org/Vol-2645/paper5.pdf) Nils Holzenberger, Andrew Blair-Stanek and Benjamin Van Durme *Proceedings of the 2020 Natural Legal Language Processing (NLLP) Workshop, 24 August 2020* # Usage ``` from transformers import AutoModel, AutoTokenizer model = AutoModel.from_pretrained("jhu-clsp/LegalBert") tokenizer = AutoTokenizer.from_pretrained("jhu-clsp/LegalBert") ``` # Citation ``` @inproceedings{holzenberger20dataset, author = {Nils Holzenberger and Andrew Blair{-}Stanek and Benjamin Van Durme}, title = {A Dataset for Statutory Reasoning in Tax Law Entailment and Question Answering}, booktitle = {Proceedings of the Natural Legal Language Processing Workshop 2020 co-located with the 26th {ACM} {SIGKDD} International Conference on Knowledge Discovery {\&} Data Mining {(KDD} 2020), Virtual Workshop, August 24, 2020}, series = {{CEUR} Workshop Proceedings}, volume = {2645}, pages = {31--38}, publisher = {CEUR-WS.org}, year = {2020}, url = {http://ceur-ws.org/Vol-2645/paper5.pdf}, } ```