Uncertainty-Aware Clarification in LLM Agents with Information Gain
Mengyi DENG ⋅ Zhiwei Li ⋅ Xin Li ⋅ Tingyu ZHU ⋅ Ying Zhao ⋅ Zhijiang Guo ⋅ Wei Wang
Abstract
Large Language Model (LLM) agents often operate under underspecified user instructions, where latent uncertainty over user intent leads to erroneous tool actions. To address this challenge, we propose a goal-oriented clarification framework that aligns clarification behavior with ambiguity resolution. Central to our approach is the Information Gain Reward, a metric that quantifies the utility of clarification questions by measuring the Bayesian belief update towards the ground-truth goal induced by the clarification exchange. We train the clarifier (LLM) using this reward to optimize for high information gain, ensuring that clarifications effectively reduce uncertainty and improve task completion within the agent-tool-user environment. We validate our framework within a clarification-enhanced $\tau$-Bench environment, conducting cross-agent evaluations across five heterogeneous backbones. Empirical results demonstrate that our method consistently improves the success rate by 3.7\% over the no-clarification baseline, incurring small overhead via surgical interventions (averaging only 0.3 additional steps). Our code and data are available at the anonymous repository: \url{https://anonymous.4open.science/r/Information-Gain-Clarifier-AD68/}.
Successful Page Load