Timezone: »

What Language Model Architecture and Pretraining Objective Works Best for Zero-Shot Generalization?
Thomas Wang · Adam Roberts · Daniel Hesslow · Teven Le Scao · Hyung Won Chung · Iz Beltagy · Julien Launay · Colin Raffel

Tue Jul 19 08:45 AM -- 08:50 AM (PDT) @ Room 301 - 303

Large pretrained Transformer language models have been shown to exhibit zero-shot generalization, i.e. they can perform a wide variety of tasks that they were not explicitly trained on. However, the architectures and pretraining objectives used across state-of-the-art models differ significantly, and there has been limited systematic comparison of these factors. In this work, we present a large-scale evaluation of modeling choices and their impact on zero-shot generalization. In particular, we focus on text-to-text models and experiment with three model architectures (causal/non-causal decoder-only and encoder-decoder), trained with two different pretraining objectives (autoregressive and masked language modeling), and evaluated with and without multitask prompted finetuning. We train models with over 5 billion parameters for more than 168 billion tokens, thereby increasing the likelihood that our conclusions will transfer to even larger scales. Our experiments show that causal decoder-only models trained on an autoregressive language modeling objective exhibit the strongest zero-shot generalization after purely self-supervised pretraining. However, models with non-causal visibility on their input trained with a masked language modeling objective followed by multitask finetuning perform the best among our experiments. We therefore consider the adaptation of pretrained models across architectures and objectives. Code and checkpoints are available at https://github.com/bigscience- workshop/architecture-objective.

Author Information

Thomas Wang (Hugging Face)
Adam Roberts (Google Brain)
Daniel Hesslow (Lighton)
Teven Le Scao (Hugging Face)
Hyung Won Chung (Google)
Iz Beltagy (Allen Institute for AI (AI2))
Julien Launay (École Normale Supérieure)
Colin Raffel (Google Brain)

Related Events (a corresponding poster, oral, or spotlight)

More from the Same Authors