Poster
Online learning with kernel losses
Niladri Chatterji · Aldo Pacchiano · Peter Bartlett
Pacific Ballroom #185
Keywords: [ Bandits ] [ Computational Learning Theory ] [ Online Learning ] [ Statistical Learning Theory ]
[
Abstract
]
Abstract:
We present a generalization of the adversarial linear bandits framework, where the underlying losses are kernel functions (with an associated reproducing kernel Hilbert space) rather than linear functions. We study a version of the exponential weights algorithm and bound its regret in this setting. Under conditions on the eigen-decay of the kernel we provide a sharp characterization of the regret for this algorithm. When we have polynomial eigen-decay (), we find that the regret is bounded by . While under the assumption of exponential eigen-decay () we get an even tighter bound on the regret . When the eigen-decay is polynomial we also show a \emph{non-matching} minimax lower bound on the regret of and a lower bound of when the decay in the eigen-values is exponentially fast.
We also study the full information setting when the underlying losses are kernel functions and present an adapted exponential weights algorithm and a conditional gradient descent algorithm.
Live content is unavailable. Log in and register to view live content