Competing in the Dark: An Efficient Algorithm for Bandit Linear Optimization
Loading...
Penn collection
Statistics Papers
Degree type
Discipline
Subject
Statistics and Probability
Funder
Grant number
License
Copyright date
Distributor
Related resources
Author
Abernethy, Jacob D
Hazan, Elad
Rakhlin, Alexander
Contributor
Abstract
We introduce an efficient algorithm for the problem of online linear optimization in the bandit setting which achieves the optimal O*(√T)regret. The setting is a natural generalization of the nonstochastic multiarmed bandit problem, and the existence of an efficient optimal algorithm has been posed as an open problem in a number of recent papers. We show how the difficulties encountered by previous approaches are overcome by the use of a self-concordant potential function. Our approach presents a novel connection between online learning and interior point methods.
Advisor
Date of presentation
2009-01-01
Conference name
Statistics Papers
Conference dates
2023-05-17T15:27:43.000
Conference location
Date Range for Data Collection (Start Date)
Date Range for Data Collection (End Date)
Digital Object Identifier
Series name and number
Volume number
Issue number
Publisher
Publisher DOI
Comments
At the time of publication, author Alexander Rakhlin was affiliated with the University of California, Berkeley. Currently, he is a faculty member at the Statistics Department at the University of Pennsylvania.