Material Detail

Some Recent Bandit Results

Some Recent Bandit Results

This video was recorded at Large-scale Online Learning and Decision Making (LSOLDM) Workshop, Cumberland Lodge 2012. Research on multi-armed bandits is expanding in several directions. In this talk I will cover a number of recent results which reflect the variety of current approaches. The first part will be devoted to the analysis of stochastic bandits when reward distributions have heavy tails, preventing the use of standard statistical estimators. Next, I will consider bandits with switching costs and show new upper and lower bounds under different assumptions on the reward processes. The last part ofthe talk will focus on combinatorial bandits, which include several interesting special cases like ranking and multiple pulls. In this setting I will discuss merits and limitations of the three main existing algorithmic approaches (Mirror Descent, Exp2, FPL). Joint work with: Sebastien Bubeck, Ofer Dekel, Elad Hazan, Sham Kakade, Gabor Lugosi, Ohad Shamir

Quality

  • User Rating
  • Comments
  • Learning Exercises
  • Bookmark Collections
  • Course ePortfolios
  • Accessibility Info

More about this material

Comments

Log in to participate in the discussions or sign up if you are not already a MERLOT member.