Skip to yearly menu bar Skip to main content


Poster
in
Workshop: Next Generation of Sequence Modeling Architectures

Probing the Decision Boundaries of In-context Learning in Large Language Models

Siyan Zhao · Tung Nguyen · Aditya Grover


Abstract:

In-context learning in large language models (LLMs) enables them to generalize to new tasks by prompting with a few exemplars without explicit parameter updates. Attempts have been made to understand in-context learning as a function of model scale, pretraining data, and other factors. In this work, we propose a new mechanism to probe and understand in-context learning from the lens of decision boundaries for in-context binary classification. Decision boundaries are straightforward to visualize and qualitatively demonstrate the inductive biases of standard classifiers. Surprisingly, we find that the decision boundaries learned by current LLMs in simple binary classification tasks are irregular and non-smooth. We investigate the factors influencing these decision boundaries and explores methods to enhance their generalizability, including training-free and fine-tuning methods, the impact of model architecture, and the effectiveness of active prompting techniques for smoothing decision boundaries in a data-efficient manner. Our findings provide a deeper understanding of in-context learning dynamics and offer practical improvements for enhancing robustness and generalizability of in-context learning.

Chat is not available.