gpt-neo-2.7B-titles / README.md
csinva's picture
update readme
eb636b6
|
raw
history blame
1.35 kB
metadata
license: apache-2.0
widget:
  - text: |+
      2021

Full code and details at https://github.com/csinva/gpt-paper-title-generator

Model

from transformers import AutoModelForCausalLM, pipeline, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("csinva/gpt-neo-2.7B-titles")
tokenizer = AutoTokenizer.from_pretrained("EleutherAI/gpt-neo-2.7B")
pipe = pipeline('text-generation', model=model, tokenizer=tokenizer)
pipe('2022\n\n')

Data

  • all papers on arXiv in the categories cs.AI, cs.LG, stat.ML
    • date cutoff: only finetuned on papers with dat on or before Apr 1, 2022
    • random 5% of papers also excluded
    • this results in 98,388 papers for finetuning
  • during finetuning each paper title was given starting with the prompt <year>\n\n <title>\n (e.g. 2022\n\n Emb-GAM: an Interpretable and Efficient Predictor using Pre-trained Language Models\n)