Skip to content

Instruction Tuning Dataset and Models for Long Text Generation with Corpus Extraction

Notifications You must be signed in to change notification settings

ecent2000/LongForm

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

1 Commit
 
 
 
 
 
 

Repository files navigation

LongForm: Optimizing Instruction Tuning for Long Text Generation with Corpus Extraction

The LongForm dataset is created by leveraging English corpus examples with augmented instructions. We select a diverse set of human-written documents from existing corpora such as C4 and Wikipedia and generate instructions for the given documents via LLMs. Then, we extend these examples with structured corpora examples such as Stack Exchange and WikiHow and task examples such as question answering, email writing, grammar error correction, story/poem generation, and text summarization.

The LongForm dataset

The Dataset and Models

We release the LongForm dataset on Github and on HuggingFace. Check out the paper or these links for more detail.

The LongForm Models: We release the models on the model hub of HuggingFace. We cannot release LongForm-LLaMA-7B publicly due to restrictions of LLaMA models.

LongForm-T5-XL: https://huggingface.co/akoksal/LongForm-T5-XL

LongForm-OPT-2.7B: https://huggingface.co/akoksal/LongForm-OPT-2.7B

LongForm-OPT-6.7B: https://huggingface.co/akoksal/LongForm-OPT-6.7B

Evaluation

We provide in-depth evaluation of LongForm models and baselines in the paper. We present the METEOR scores of models in out-of-domain datasets. In all tasks, Recipe Generation (RGen), long-form question answering (ELI5), short story generation (WritingPrompts/WP), LongForm models outperform prior instruction-tuned models.

All Recipe Generation ELI5 Writing Prompts
T0++ 10.9 18.7 3.8 10.2
Tk-Instruct 6.3 12.9* 3.6 2.4
Flan-T5 10.6 20.9* 3.5 7.4
Alpaca-LLaMA-7B 14.6 19.5 12.5 11.8
OPT-30B 11.1 18.6 12.2 2.6
LongForm-T5-XL 16.3 20.2 18.3 10.6
LongForm-OPT-2.7B 17.8 15.5 17.9 19.9
LongForm-OPT-6.7B 17.7 16.9 17.2 19.0
LongForm-LLaMA-7B 19.7 21.7 18.6 18.9

Limitations

The LongForm dataset and models mainly focus on long text generation and have limitations regarding structured prediction tasks in NLP. Additionally, we observe that LongForm models may present hallucination problems similar to those found in LLMs.

About

Instruction Tuning Dataset and Models for Long Text Generation with Corpus Extraction

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published