Skip to yearly menu bar Skip to main content


Poster
in
Workshop: Workshop on Theoretical Foundations of Foundation Models (TF2M)

Unveiling Induction Heads: Provable Training Dynamics and Feature Learning in Transformers

Siyu Chen · Heejune Sheen · Tianhao Wang · Zhuoran Yang


Abstract: In-context learning (ICL) is a cornerstone of large language model functionality, yet its theoretical foundations remain elusive due to the complexity of transformer architectures. In particular, most existing work only theoretically explains how the attention mechanism facilitates ICL under certain data models. It remains unclear how the other building blocks of the transformer contribute to ICL. To address this question, we study how a two-attention-layer transformer is trained to perform ICL on $n$-gram Markov chain data, where each token in the Markov chain statistically depends on the previous $n$ tokens. We analyze a sophisticated transformer model featuring relative positional embedding, multi-head softmax attention, and a feed-forward layer with normalization. We prove that the gradient flow with respect to a cross-entropy ICL loss converges to a limiting model that performs a generalized version of the ``induction head'' mechanism with a learned feature, resulting from the congruous contribution of all the building blocks.

Chat is not available.