This page is inactive since the closure of MSR-SVC in September 2014. The name “multi-armed bandits” comes from a\u00a0whimsical scenario in which a gambler faces\u00a0several slot machines, a.k.a. “one-armed bandits”, that look identical at first\u00a0but\u00a0produce different expected\u00a0winnings. The crucial issue here is the trade-off between acquiring new information (exploration<\/em>) and capitalizing on the information available so far (exploitation<\/em>). While the MAB problems have been studied extensively in Machine Learning, Operations Research and Economics, many exciting questions\u00a0are open. One aspect that\u00a0we are particularly interested in concerns\u00a0modeling and\u00a0efficiently using various types of side information that\u00a0may be\u00a0available to the algorithm.<\/p>\n Contact<\/u>: Alex Slivkins<\/a>.<\/p>\n Prof. S\u00e9bastien Bubeck<\/a> (Princeton) Former interns<\/strong> This is an umbrella project for several related efforts at Microsoft Research Silicon Valley that address various Multi-Armed Bandit (MAB) formulations motivated by web search and ad placement. The MAB problem is a classical paradigm in Machine Learning in which an online algorithm chooses from a set of strategies in a sequence of trials so […]<\/p>\n","protected":false},"featured_media":0,"template":"","meta":{"msr-url-field":"","msr-podcast-episode":"","msrModifiedDate":"","msrModifiedDateEnabled":false,"ep_exclude_from_search":false,"_classifai_error":"","footnotes":""},"research-area":[13561,13556,13548],"msr-locale":[268875],"msr-impact-theme":[],"msr-pillar":[],"class_list":["post-170006","msr-project","type-msr-project","status-publish","hentry","msr-research-area-algorithms","msr-research-area-artificial-intelligence","msr-research-area-economics","msr-locale-en_us","msr-archive-status-complete"],"msr_project_start":"","related-publications":[],"related-downloads":[],"related-videos":[],"related-groups":[],"related-events":[],"related-opportunities":[],"related-posts":[],"related-articles":[],"tab-content":[],"slides":[],"related-researchers":[{"type":"user_nicename","value":"slivkins","display_name":"Alex Slivkins","author_link":"Alex Slivkins<\/a>","is_active":false,"user_id":33685,"last_first":"Slivkins, Alex","people_section":0,"alias":"slivkins"}],"msr_research_lab":[199571],"msr_impact_theme":[],"_links":{"self":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-project\/170006","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-project"}],"about":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/types\/msr-project"}],"version-history":[{"count":2,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-project\/170006\/revisions"}],"predecessor-version":[{"id":795062,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-project\/170006\/revisions\/795062"}],"wp:attachment":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media?parent=170006"}],"wp:term":[{"taxonomy":"msr-research-area","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/research-area?post=170006"},{"taxonomy":"msr-locale","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-locale?post=170006"},{"taxonomy":"msr-impact-theme","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-impact-theme?post=170006"},{"taxonomy":"msr-pillar","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-pillar?post=170006"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}
\n<\/em><\/p>\n<\/div>\nResearch directions<\/h1>\n
\n
External visitors and collaborators<\/h1>\n
\nProf. Robert Kleinberg<\/a>\u00a0(Cornell)
\nFilip Radlinski<\/a>\u00a0(MSR Cambridge)
\nProf. Eli Upfal<\/a>\u00a0(Brown)<\/p>\n
\nYogi Sharma<\/a> (Cornell \u2014> Facebook; intern at MSR-SV in summer 2008)
\nUmar Syed<\/a>\u00a0(Princeton\u00a0\u2014>\u00a0Google; intern at MSR-SV in\u00a0summer 2008)
\nShaddin Dughmi<\/a>\u00a0(Stanford \u2014>USC; intern at MSR-SV in summer 2010)
\nAshwinkumar Badanidiyuru<\/a>\u00a0(Cornell –> Google;\u00a0intern at MSR-SV in summer 2012)<\/p>\nMAB\u00a0problems with similarity information<\/h1>\n
\n
\nRobert Kleinberg, Alex Slivkins and Eli Upfal (STOC 2008<\/a>)
\nAbstract<\/u> We introduce\u00a0a version of the stochastic\u00a0MAB problem, possibly with a very large set of arms, in which the expected payoffs obey a Lipschitz condition with respect to a given metric space. The goal is to minimize regret as a function of time, both in the worst case and for ‘nice’ problem instances.<\/div>\n<\/li>\n
\nRobert Kleinberg and Alex Slivkins (SODA 2010<\/a>)
\nAbstract<\/u> We focus on the connections between online learning and metric topology. The main result that the worst-case regret is either O(log t) or at least sqrt{t}, depending on whether the completion of the metric space is compact and countable. We prove a number of other dichotomy-style results, and extend them to the full-feedback (experts) version.<\/li>\n
\n<\/strong>Alex Slivkins, Filip Radlinski and Sreenivas Gollapudi (ICML 2010<\/a>)
\nAbstract<\/u> We present a learning-to-rank framework for web search that incorporates similarity and correlation between documents and thus, unlike prior work, scales to large document collections.<\/li>\n
\nAlex Slivkins (COLT 2011<\/a>)
\nAbstract<\/u> In the ‘contextual bandits’ setting, in each round nature reveals a ‘context’ x, algorithm chooses an ‘arm’ y, and the expected payoff is \u00b5(x,y). Similarity info is expressed by a metric space over the (x,y) pairs such that \u00b5 is a Lipschitz function. Our algorithms are based on adaptive (rather than uniform) partitions of the metric space which are adjusted to the popular and high-payoff regions.<\/li>\n
\nAlex Slivkins (NIPS 2011<\/a>)
\nAbstract<\/u> Suppose an MAB algorithm is given a tree-based classification of arms. This tree implicitly defines a “similarity distance” between arms, but the numeric distances are not revealed to the algorithm. Our algorithm (almost) matches the best known guarantees for the setting (Lipschitz MAB) in which the distances are revealed.<\/li>\n<\/ul>\nMAB problems in a changing environment<\/h1>\n
\n
\nAlex Slivkins and Eli Upfal (COLT 2008<\/a>)
\nAbstract<\/u> We study a version of the stochastic MAB problem in which the expected reward of each arm evolves stochastically and gradually in time, following an independent Brownian motion or a similar process. Our benchmark is a hypothetical policy that chooses the best arm in each round.<\/li>\n
\nUmar Syed, Alex Slivkins and Nina Mishra (NIPS 2009<\/a>)
\nAbstract<\/u> Query intent may shift over time. A classifier can use the available signals to predict a shift in intent. Then a bandit algorithm can be used to find the new relevant results. We present a meta-algorithm that combines such classifier with a bandit algorithm in a feedback loop.<\/li>\n
\nAlex Slivkins (COLT 2011<\/a>)
\nAbstract<\/u> Interpreting the current time as a part of the contextual information, we obtain a very general bandit framework that (in addition to similarity between arms and contexts) can include slowly changing payoffs and variable sets of arms.<\/li>\n
\nSebastien Bubeck and Alex Slivkins (COLT 2012<\/a>)
\nAbstract<\/u> We present a new bandit algorithm whose regret is optimal both for adversarial rewards and for stochastic rewards, achieving, resp., square-root regret and polylog regret.<\/li>\n<\/ul>\nExplore-exploit tradeoff in mechanism design<\/h1>\n
\n
\n<\/strong>Moshe Babaioff, Alex Slivkins and Yogi Sharma (EC 2009<\/a>)
\nAbstract<\/u> We consider a multi-round auction setting motivated by pay-per-click auctions in the Internet advertising, which can be viewed as a strategic version of the MAB problem. We investigate how the design of MAB algorithms is affected by the restriction of truthfulness. We show striking differences in terms of both the structure of an algorithm and its regret.\u00a0 <\/em><\/div>\n<\/li>\n
\nMoshe Babaioff, Robert Kleinberg and Alex Slivkins (EC 2010<\/a> Best Paper Award<\/strong>)
\nAbstract<\/u> We show that any single-parameter, monotone\u00a0allocation\u00a0function can be\u00a0truthfully implemented using\u00a0a single call to\u00a0this\u00a0function. We apply this to MAB mechanisms.<\/div>\n<\/li>\n
\nAlex Slivkins (Open Question at COLT 2011<\/u><\/a>)
\nAbstract<\/u> The reduction in the EC’10 paper opens up a problem of designing monotone\u00a0MAB allocation rules, a new twist on the familiar\u00a0MAB problem.<\/span><\/span><\/div>\n<\/li>\n
\nMoshe Babaioff, Robert Kleinberg and Alex Slivkins (EC 2013<\/a>)
\nAbstract<\/u> We\u00a0generalize the main\u00a0result of the EC’10 paper to the multi-parameter setting. We apply this to a natural multi-parameter extension of MAB mechanisms.<\/li>\n<\/ul>\nExplore-exploit learning with\u00a0limited resources<\/h1>\n
\n
\nMoshe Babaioff, Shaddin Dughmi, Robert Kleinberg and Alex Slivkins (EC 2012<\/a>)
\nAbstract<\/u> We consider dynamic pricing with limited supply and unknown demand distribution. We extend MAB techniques to the limited supply setting, and obtain optimal regret rates.<\/div>\n<\/li>\n
\nAshwinkumar Badanidiyuru, Robert Kleinberg and Alex Slivkins (FOCS 2013<\/a>)
\nAbstract<\/u> We define a broad class of explore-exploit problems with knapsack-style resource constraints, which subsumes dynamic pricing, dynamic procurement, pay-per-click ad allocation, and\u00a0many other problems. Our algorithms achieve optimal regret w.r.t. the optimal dynamic policy.<\/li>\n<\/ul>\nRisk vs. reward trade-off in MAB<\/h1>\n
\n
\nMichael Kapralov and\u00a0Rina Panigrahy (NIPS 2011<\/a>)
\nAbstract<\/u> We show that\u00a0it is theoretically possible to extract some reward in a bandit prediction game while having an exponentially small downside risk.<\/li>\n
\nAlex Andoni and Rina Panigrahy
\nAbstract<\/u> This paper\u00a0shows the role of Hermite Differential Equation in optimal risk vs reward tradeoff in prediction games.<\/li>\n<\/ul>\n<\/div>\n","protected":false},"excerpt":{"rendered":"
\n
\n
Multi-armed bandits in metric spaces<\/a><\/strong>- Sharp dichotomies for regret minimization in metric spaces<\/a><\/strong>
- Learning optimally diverse rankings over large document collections<\/a>
- Contextual bandits with similarity information<\/strong><\/a>
- Multi-armed bandits on implicit metric spaces<\/strong><\/a>
Characterizing truthful multi-armed bandit mechanisms<\/a>- \n
Truthful mechanisms with implicit payment computation<\/a><\/b>- \n
Monotone multi-armed bandit allocations<\/strong><\/a>- Multi-parameter mechanisms with implicit payment computation<\/a><\/strong>
Dynamic pricing with limited supply<\/b><\/a>- Bandits with Knapsacks<\/strong><\/a>