Skip to yearly menu bar Skip to main content


Poster
in
Workshop: Models of Human Feedback for AI Alignment

Comparing Few to Rank Many: Optimal Design for Learning Preferences

Kiran Thekumparampil · Gaurush Hiranandani · Kousha Kalantari · Shoham Sabach · Branislav Kveton

[ ] [ Project Page ]
Fri 26 Jul 8 a.m. PDT — 8 a.m. PDT

Abstract:

We study learning of human or agent preferences from a limited comparison feedback. This task is ubiquitous in machine learning and its applications, such as reinforcement learning with human feedback, have been transformational. We can naturally formulate this problem as learning a Plackett-Luce model over a universe of N choices. In this work, we propose a statistically and computationally efficient algorithm for collecting a limited K-way comparison feedback to learn a high-quality Plackett-Luce model. The key idea is to extend the D-optimal design for efficient data collection in regression problems to learning to rank. Our design elicits comparison feedback from a small collection of \poly(N,K) optimally chosen subsets from all feasible {N choose K} subsets. Specifically, we propose an approximate Frank-Wolfe algorithm and design an efficient oracle for approximately solving linear maximization over the simplex of all possible subsets. We analyze our algorithm and evaluate it empirically on open-source datasets.

Chat is not available.