Poster
in
Workshop: Workshop on Theoretical Foundations of Foundation Models (TF2M)
How Do Nonlinear Transformers Acquire Generalization-Guaranteed CoT Ability?
Hongkang Li · Meng Wang · Songtao Lu · Xiaodong Cui · Pin-Yu Chen
Chain-of-Thought (CoT) is an efficient prompting method that enables the reasoning ability of large language models by augmenting the query using multiple examples with intermediate steps. Despite the empirical success, the theoretical understanding of how to train a Transformer to achieve the CoT ability remains less explored. This is primarily due to the technical challenges involved in analyzing the nonconvex optimization on nonlinear attention models. To the best of our knowledge, this work provides the first theoretical study of training Transformers with nonlinear attention to obtain the CoT generalization capability so that the resulting model can reason on unseen tasks when the input is augmented by examples of the new task. We first quantify the required training samples and iterations to train a model with CoT ability. We then prove the success of its CoT generalization on unseen tasks with distribution-shifted testing data. Moreover, we theoretically characterize the conditions for an accurate reasoning output by CoT even when the provided reasoning examples contain noises and are not always accurate. In contrast, in-context learning (ICL), which can be viewed as one-step CoT without intermediate steps, may fail to provide an accurate output when CoT does. These theoretical findings are justified through experiments.