Date of this Version
Advances in Neural Information Processing Systems
We show a principled way of deriving online learning algorithms from a minimax analysis. Various upper bounds on the minimax value, previously thought to be non-constructive, are shown to yield algorithms. This allows us to seamlessly recover known methods and to derive new ones, also capturing such “unorthodox” methods as Follow the Perturbed Leader and the R2 forecaster. Understanding the inherent complexity of the learning problem thus leads to the development of algorithms. To illustrate our approach, we present several new algorithms, including a family of randomized methods that use the idea of a “random playout”. New versions of the Follow-the-Perturbed-Leader algorithms are presented, as well as methods based on the Littlestone’s dimension, efficient methods for matrix completion with trace norm, and algorithms for the problems of transductive learning and prediction with static experts.
Rakhlin, A., Shamir, O., & Sridharan, K. (2012). Relax and Randomize : From Value to Algorithms. Advances in Neural Information Processing Systems, 25 2150-2158. Retrieved from https://repository.upenn.edu/statistics_papers/461
Date Posted: 27 November 2017