--- license: apache-2.0 widget: - text: "2021\n\n" --- Full code and details at https://github.com/csinva/gpt-paper-title-generator **Model** - finetunes starting from the[gpt-neo-2.7B checkpoint](https://huggingface.co/EleutherAI/gpt-neo-2.7B) - for training details see [the training script](https://github.com/csinva/gpt-paper-title-generator/blob/0157f26be9b0763b4ea6480e5b149fdb8dff4626/gptneo/02_finetune_hf.py) - inference - should prepend with a year and two newlines before querying for a title, e.g. `2022\n\n` ```python from transformers import AutoModelForCausalLM, pipeline, AutoTokenizer model = AutoModelForCausalLM.from_pretrained("csinva/gpt-neo-2.7B-titles") tokenizer = AutoTokenizer.from_pretrained("EleutherAI/gpt-neo-2.7B") pipe = pipeline('text-generation', model=model, tokenizer=tokenizer) pipe('2022\n\n') ``` **Data** - all [papers on arXiv](https://www.kaggle.com/datasets/Cornell-University/arxiv) in the categories cs.AI, cs.LG, stat.ML - date cutoff: only finetuned on papers with dat on or before Apr 1, 2022 - random 5% of papers also excluded - this results in 98,388 papers for finetuning - during finetuning each paper title was given starting with the prompt `\n\n \n` (e.g. `2022\n\n Emb-GAM: an Interpretable and Efficient Predictor using Pre-trained Language Models\n`)