Skip to yearly menu bar Skip to main content


Poster

Exploring the Benefits of Training Expert Language Models over Instruction Tuning

Joel Jang · Seungone Kim · Seonghyeon Ye · Doyoung Kim · Lajanugen Logeswaran · Moontae Lee · Kyungjae Lee · Minjoon Seo

Exhibit Hall 1 #400
[ ]
[ PDF [ Poster

Abstract:

Recently, Language Models (LMs) instruction-tuned on multiple tasks, also known as multitask-prompted fine-tuning (MT), have shown capabilities to generalize to unseen tasks. Previous work has shown that scaling the number of finetuning datasets and instructions is the key component in making stronger MT LMs. In this work, we report surprising findings that show an expert LM trained on just a single task can outperform an MT LM trained with 300+ different tasks on 11 different unseen datasets and on 13 datasets of the BIG-bench benchmark by an average of 3.20% and 1.29%, respectively. This finding casts doubt on the previously held belief that simply scaling the number of tasks makes stronger MT LMs. Leveraging this finding, we further show that this distributed approach of training multiple expert LMs instead of a single MT LM for zero-shot inference possesses many benefits including (1) avoiding negative task transfer that often occurs during instruction tuning, (2) being able to continually learn new tasks without having to re-train on previous tasks to avoid catastrophic forgetting, and (3) showing compositional capabilities when merging individual experts together.

Chat is not available.