Poster
In-Context Learning as Conditioned Associative Memory Retrieval
Weimin Wu · Teng-Yun Hsiao · Jerry Yao-Chieh Hu · Wenxin Zhang · Han Liu
East Exhibition Hall A-B #E-3310
Large language models like ChatGPT can solve new problems just by being shown a few examples in a prompt. We are curious about how these models manage to “learn” so quickly without updating their internal parameters, and whether there’s a simple explanation behind this surprising behavior.We found that this process can be understood as a kind of memory retrieval. Specifically, we use a classic brain-inspired model called a Hopfield network to show how each example in the prompt subtly reshapes what the model “remembers.” This reshaping helps the model focus on the most relevant information for making predictions — just like how a person recalls different memories depending on the question they’re asked.To test this idea, we build a simplified version of a language model and run experiments with it. Our results confirm that in-context learning is stronger when the examples are similar to the test case, accurate, and drawn from a familiar setting.
Live content is unavailable. Log in and register to view live content