File size: 3,800 Bytes
852d65e
 
 
07bb7da
d34e107
 
 
 
852d65e
 
 
07bb7da
852d65e
d34e107
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
852d65e
 
 
d34e107
27a1f25
7d4d432
 
90a6525
 
98a6434
90a6525
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2f1cbbb
 
 
90a6525
 
7d4d432
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
---
language:
- en
- multilingual
- de
- it
- es
- fr
tags:
- instruction-tuning
- text-generation-inference
- text2text-generation
widget:
- text: Write an essay about meditation.
  example_title: Essay Generation
- text: Give me 5 steps to clean my room.
  example_title: How-to Instructions
- text: How are the continents formed?
  example_title: Question-Answering
- text: >-
    Prompt: A man draws a gun in a dark alley and asks for your wallet. You
    begrudgingly obey. He throws it on the ground, shoots it till it screeches,
    and turns to you; 'you are safe now'. Write a story about given prompt.
  example_title: Story Generation
- text: >-
    Write directions of a cooking recipe with these ingredients: chicken breast,
    carrots, green peas, celery, butter, onion, flour, salt, black pepper,
    celery seed, chicken broth, milk, unbaked pie crusts
  example_title: Recipe Generation
- text: Schreiben Sie einen Blogbeitrag über die Vorteile des Lesens von Büchern.
  example_title: German Essay Generation
inference:
  parameters:
    top_p: 0.9
    do_sample: true
    max_length: 75
datasets:
- akoksal/LongForm
---

## LongForm-T5-XL
The LongForm dataset is created by leveraging English corpus examples with augmented instructions. We select a diverse set of human-written documents from existing corpora such as C4 and Wikipedia and generate instructions for the given documents via LLMs. Then, we extend these examples with structured corpora examples such as Stack Exchange and WikiHow and task examples such as question answering, email writing, grammar error correction, story/poem generation, and text summarization.


Github Repo: https://github.com/akoksal/LongForm

LongForm-**OPT-2.7B**: https://huggingface.co/akoksal/LongForm-OPT-2.7B

LongForm-**OPT-6.7B**: https://huggingface.co/akoksal/LongForm-OPT-6.7B


## Evaluation
We provide in-depth evaluation of LongForm models and baselines in the paper. We present the METEOR scores of models in out-of-domain datasets. In all tasks, Recipe Generation (RGen), long-form question answering (ELI5), short story generation (WritingPrompts/WP), LongForm models outperform prior instruction-tuned models.
|          | **All** | **Recipe Generation**             | **ELI5** | **Writing Prompts** |
|-----------------------|---------|-----------------------------------|----------|---------------------|
| **T0++**              | 10.9    | 18.7                              | 3.8      | 10.2                |
| **Tk-Instruct**       | 6.3     | 12.9* | 3.6      | 2.4                 |
| **Flan-T5**           | 10.6    | 20.9* | 3.5      | 7.4                 |
| **Alpaca-LLaMA-7B**   | 14.6    | 19.5                              | 12.5     | 11.8                |
| **OPT-30B**           | 11.1    | 18.6                              | 12.2     | 2.6                 |
| **[LongForm-T5-XL](https://huggingface.co/akoksal/LongForm-T5-XL)**    | 16.3    | 20.2                              | 18.3     | 10.6                |
| **[LongForm-OPT-2.7B](https://huggingface.co/akoksal/LongForm-OPT-2.7B)**   | 17.8    | 15.5                              | 17.9     | **19.9**                |
| **[LongForm-OPT-6.7B](https://huggingface.co/akoksal/LongForm-OPT-6.7B)** | 17.7    | 16.9                              | 17.2     | 19.0                |
| **LongForm-LLaMA-7B**‡ | **19.7**    | **21.7**                              | **18.6**     | 18.9                |

‡: We cannot release LongForm-LLaMA-7B publicly due to restrictions of LLaMA models.

## Limitations
The LongForm dataset and models mainly focus on long text generation and have limitations regarding structured prediction tasks in NLP. Additionally, we observe that LongForm models may present hallucination problems similar to those found in LLMs.