Edit model card

OpenAI GPT-2 Samsum

Model description

This model has been trained with the SAMSum dataset. The SAMSum dataset contains approximately 16,000 conversational dialogues accompanied by summaries. These conversations were created and written by linguists proficient in fluent English. Linguists were instructed to create conversations that reflect the ratio of topics found in real-life journalistic conversations similar to their daily written conversations. The style and tone vary; conversations can be informal, semi-formal, or formal, and may include slang terms, expressions, and spelling errors. Subsequently, the conversations were annotated with summaries. The summaries are expected to be concise summaries of what people were talking about during the conversation, written in the third person. The SAMSum dataset was prepared by the Samsung Research Institute Poland and is distributed for research purposes.

Training

This GPT-2 model is rated for an average of 1 hour with an L4 GPU.

Training Results

examples Authors

Downloads last month
58
Safetensors
Model size
124M params
Tensor type
F32
·

Dataset used to train anezatra/gpt2-samsum-124M