Algorithms · Bandit Algorithms

TitleAuthors
Blocking BanditsSoumya Basu · Rajat Sen · Sujay Sanghavi · Sanjay Shakkottai
Combinatorial Bandits with Relative FeedbackAadirupa Saha · Aditya Gopalan
Decentralized Cooperative Stochastic BanditsDavid Martínez-Rubio · Varun Kanade · Patrick Rebeschini
Doubly-Robust Lasso BanditGi-Soo Kim · Myunghee Cho Paik
Efficient Pure Exploration in Adaptive Round modeltianyuan jin · Jieming SHI · Xiaokui Xiao · Enhong Chen
Epsilon-Best-Arm Identification in Pay-Per-Reward Multi-Armed BanditsSivan Sabato
Nonstochastic Multiarmed Bandits with Unrestricted DelaysTobias Sommer Thune · Nicolò Cesa-Bianchi · Yevgeny Seldin
On the Optimality of Perturbations in Stochastic and Adversarial Multi-armed Bandit ProblemsBaekjin Kim · Ambuj Tewari
Phase Transitions and Cyclic Phenomena in Bandits with Switching ConstraintsDavid Simchi-Levi · Yunzong Xu
Polynomial Cost of Adaptation for X-Armed BanditsHedi Hadiji
Semi-Parametric Dynamic Contextual PricingVirag Shah · Ramesh Johari · Jose Blanchet
Bayesian Optimization under Heavy-tailed PayoffsSayak Ray Chowdhury · Aditya Gopalan
Connections Between Mirror Descent, Thompson Sampling and the Information RatioJulian Zimmert · Tor Lattimore
Individual Regret in Cooperative Nonstochastic Multi-Armed BanditsYogev Bar-On · Yishay Mansour
Learning Multiple Markov Chains via Adaptive AllocationMohammad Sadegh Talebi · Odalric-Ambrym Maillard
Linear Stochastic Bandits Under Safety ConstraintsSanae Amani · Mahnoosh Alizadeh · Christos Thrampoulidis
Personalizing Many Decisions with High-Dimensional CovariatesNima Hamidi · Mohsen Bayati · Kapil Gupta
Non-Asymptotic Pure Exploration by Solving GamesRémy Degenne · Wouter Koolen · Pierre Ménard
Online EXP3 Learning in Adversarial Bandits with Delayed FeedbackIlai Bistritz · Zhengyuan Zhou · Xi Chen · Nicholas Bambos · Jose Blanchet
Optimal Best Markovian Arm Identification with Fixed ConfidenceVrettos Moulos
Oracle-Efficient Algorithms for Online Linear Optimization with Bandit FeedbackShinji Ito · Daisuke Hatano · Hanna Sumita · Kei Takemura · Takuro Fukunaga · Naonori Kakimura · Ken-Ichi Kawarabayashi
Regret Bounds for Thompson Sampling in Episodic Restless Bandit ProblemsYoung Hun Jung · Ambuj Tewari
Thresholding Bandit with Optimal Aggregate RegretChao Tao · Saúl Blanco · Jian Peng · Yuan Zhou
Weighted Linear Bandits for Non-Stationary EnvironmentsYoan Russac · Claire Vernade · Olivier Cappé
Improved Regret Bounds for Bandit Combinatorial OptimizationShinji Ito · Daisuke Hatano · Hanna Sumita · Kei Takemura · Takuro Fukunaga · Naonori Kakimura · Ken-Ichi Kawarabayashi
Learning in Generalized Linear Contextual Bandits with Stochastic DelaysZhengyuan Zhou · Renyuan Xu · Jose Blanchet
Low-Rank Bandit Methods for High-Dimensional Dynamic PricingJonas Mueller · Vasilis Syrgkanis · Matt Taddy
MaxGap Bandit: Adaptive Algorithms for Approximate RankingSumeet Katariya · Ardhendu Tripathy · Robert Nowak
Model Selection for Contextual BanditsDylan Foster · Akshay Krishnamurthy · Haipeng Luo
No-Regret Learning in Unknown Games with Correlated PayoffsPier Giuseppe Sessa · Ilija Bogunovic · Maryam Kamgarpour · Andreas Krause
Nonparametric Contextual Bandits in Metric Spaces with Unknown MetricNirandika Wanigasekara · Christina Yu
Pure Exploration with Multiple Correct AnswersRémy Degenne · Wouter Koolen
Recovering BanditsCiara Pike-Burke · Steffen Grunewalder
Sequential Experimental Design for Transductive Linear BanditsLalit Jain · Kevin Jamieson · Tanner Fiez · Lillian Ratliff
SIC-MMAB: Synchronisation Involves Communication in Multiplayer Multi-Armed BanditsEtienne Boursier · Vianney Perchet
Are sample means in multi-armed bandits positively or negatively biased?Jaehyeok Shin · Aaditya Ramdas · Alessandro Rinaldo
A New Perspective on Pool-Based Active Classification and False-Discovery ControlLalit Jain · Kevin Jamieson
Bootstrapping Upper Confidence BoundBotao Hao · Yasin Abbasi Yadkori · Zheng Wen · Guang Cheng
Categorized BanditsMatthieu Jedor · Vianney Perchet · Jonathan Louedec
Censored Semi-Bandits: A Framework for Resource Allocation with Censored FeedbackArun Verma · Manjesh Hanawal · Arun Rajkumar · Raman Sankaran
Contextual Bandits with Cross-LearningSantiago Balseiro · Negin Golrezaei · Mohammad Mahdian · Vahab Mirrokni · Jon Schneider
Distribution oblivious, risk-aware algorithms for multi-armed bandits with unbounded rewardsAnmol Kagrecha · Jayakrishnan Nair · Krishna Jagannathan
Dynamic Incentive-Aware Learning: Robust Pricing in Contextual AuctionsNegin Golrezaei · Adel Javanmard · Vahab Mirrokni
Stochastic Bandits with Context DistributionsJohannes Kirschner · Andreas Krause
Thompson Sampling and Approximate InferenceMy Phan · Yasin Abbasi Yadkori · Justin Domke
Thompson Sampling for Multinomial Logit Contextual BanditsMin-hwan Oh · Garud Iyengar
Thompson Sampling with Information Relaxation PenaltiesSeungki Min · Costis Maglaras · Ciamac C Moallemi