Edit model card
YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

This is a gpt2-large model, finetuned on the Wikitext-103 dataset.

It achieves a perplexity of 10.56 using a "sliding window" context, using the run_clm.py script at https://github.com/neulab/knn-transformers.

Base LM: distilgpt2 gpt2
base perplexity 18.25 14.84
+ kNN-LM 15.03 12.57
+ RetoMaton 14.70 12.46

This model was released as part of the paper "Neuro-Symbolic Language Modeling with Automaton-augmented Retrieval" (ICML'2022).

For more information, see: https://github.com/neulab/knn-transformers

If you use this model, please cite:

@inproceedings{alon2022neuro,
  title={Neuro-Symbolic Language Modeling with Automaton-augmented Retrieval},
  author={Alon, Uri and Xu, Frank and He, Junxian and Sengupta, Sudipta and Roth, Dan and Neubig, Graham},
  booktitle={International Conference on Machine Learning},
  pages={468--485},
  year={2022},
  organization={PMLR}
}
Downloads last month
7