Blocking Bandits | Soumya Basu · Rajat Sen · Sujay Sanghavi · Sanjay Shakkottai |
Combinatorial Bandits with Relative Feedback | Aadirupa Saha · Aditya Gopalan |
Decentralized Cooperative Stochastic Bandits | David Martínez-Rubio · Varun Kanade · Patrick Rebeschini |
Doubly-Robust Lasso Bandit | Gi-Soo Kim · Myunghee Cho Paik |
Efficient Pure Exploration in Adaptive Round model | tianyuan jin · Jieming SHI · Xiaokui Xiao · Enhong Chen |
Epsilon-Best-Arm Identification in Pay-Per-Reward Multi-Armed Bandits | Sivan Sabato |
Nonstochastic Multiarmed Bandits with Unrestricted Delays | Tobias Sommer Thune · Nicolò Cesa-Bianchi · Yevgeny Seldin |
On the Optimality of Perturbations in Stochastic and Adversarial Multi-armed Bandit Problems | Baekjin Kim · Ambuj Tewari |
Phase Transitions and Cyclic Phenomena in Bandits with Switching Constraints | David Simchi-Levi · Yunzong Xu |
Polynomial Cost of Adaptation for X-Armed Bandits | Hedi Hadiji |
Semi-Parametric Dynamic Contextual Pricing | Virag Shah · Ramesh Johari · Jose Blanchet |
Bayesian Optimization under Heavy-tailed Payoffs | Sayak Ray Chowdhury · Aditya Gopalan |
Connections Between Mirror Descent, Thompson Sampling and the Information Ratio | Julian Zimmert · Tor Lattimore |
Individual Regret in Cooperative Nonstochastic Multi-Armed Bandits | Yogev Bar-On · Yishay Mansour |
Learning Multiple Markov Chains via Adaptive Allocation | Mohammad Sadegh Talebi · Odalric-Ambrym Maillard |
Linear Stochastic Bandits Under Safety Constraints | Sanae Amani · Mahnoosh Alizadeh · Christos Thrampoulidis |
Personalizing Many Decisions with High-Dimensional Covariates | Nima Hamidi · Mohsen Bayati · Kapil Gupta |
Non-Asymptotic Pure Exploration by Solving Games | Rémy Degenne · Wouter Koolen · Pierre Ménard |
Online EXP3 Learning in Adversarial Bandits with Delayed Feedback | Ilai Bistritz · Zhengyuan Zhou · Xi Chen · Nicholas Bambos · Jose Blanchet |
Optimal Best Markovian Arm Identification with Fixed Confidence | Vrettos Moulos |
Oracle-Efficient Algorithms for Online Linear Optimization with Bandit Feedback | Shinji Ito · Daisuke Hatano · Hanna Sumita · Kei Takemura · Takuro Fukunaga · Naonori Kakimura · Ken-Ichi Kawarabayashi |
Regret Bounds for Thompson Sampling in Episodic Restless Bandit Problems | Young Hun Jung · Ambuj Tewari |
Thresholding Bandit with Optimal Aggregate Regret | Chao Tao · Saúl Blanco · Jian Peng · Yuan Zhou |
Weighted Linear Bandits for Non-Stationary Environments | Yoan Russac · Claire Vernade · Olivier Cappé |
Improved Regret Bounds for Bandit Combinatorial Optimization | Shinji Ito · Daisuke Hatano · Hanna Sumita · Kei Takemura · Takuro Fukunaga · Naonori Kakimura · Ken-Ichi Kawarabayashi |
Learning in Generalized Linear Contextual Bandits with Stochastic Delays | Zhengyuan Zhou · Renyuan Xu · Jose Blanchet |
Low-Rank Bandit Methods for High-Dimensional Dynamic Pricing | Jonas Mueller · Vasilis Syrgkanis · Matt Taddy |
MaxGap Bandit: Adaptive Algorithms for Approximate Ranking | Sumeet Katariya · Ardhendu Tripathy · Robert Nowak |
Model Selection for Contextual Bandits | Dylan Foster · Akshay Krishnamurthy · Haipeng Luo |
No-Regret Learning in Unknown Games with Correlated Payoffs | Pier Giuseppe Sessa · Ilija Bogunovic · Maryam Kamgarpour · Andreas Krause |
Nonparametric Contextual Bandits in Metric Spaces with Unknown Metric | Nirandika Wanigasekara · Christina Yu |
Pure Exploration with Multiple Correct Answers | Rémy Degenne · Wouter Koolen |
Recovering Bandits | Ciara Pike-Burke · Steffen Grunewalder |
Sequential Experimental Design for Transductive Linear Bandits | Lalit Jain · Kevin Jamieson · Tanner Fiez · Lillian Ratliff |
SIC-MMAB: Synchronisation Involves Communication in Multiplayer Multi-Armed Bandits | Etienne Boursier · Vianney Perchet |
Are sample means in multi-armed bandits positively or negatively biased? | Jaehyeok Shin · Aaditya Ramdas · Alessandro Rinaldo |
A New Perspective on Pool-Based Active Classification and False-Discovery Control | Lalit Jain · Kevin Jamieson |
Bootstrapping Upper Confidence Bound | Botao Hao · Yasin Abbasi Yadkori · Zheng Wen · Guang Cheng |
Categorized Bandits | Matthieu Jedor · Vianney Perchet · Jonathan Louedec |
Censored Semi-Bandits: A Framework for Resource Allocation with Censored Feedback | Arun Verma · Manjesh Hanawal · Arun Rajkumar · Raman Sankaran |
Contextual Bandits with Cross-Learning | Santiago Balseiro · Negin Golrezaei · Mohammad Mahdian · Vahab Mirrokni · Jon Schneider |
Distribution oblivious, risk-aware algorithms for multi-armed bandits with unbounded rewards | Anmol Kagrecha · Jayakrishnan Nair · Krishna Jagannathan |
Dynamic Incentive-Aware Learning: Robust Pricing in Contextual Auctions | Negin Golrezaei · Adel Javanmard · Vahab Mirrokni |
Stochastic Bandits with Context Distributions | Johannes Kirschner · Andreas Krause |
Thompson Sampling and Approximate Inference | My Phan · Yasin Abbasi Yadkori · Justin Domke |
Thompson Sampling for Multinomial Logit Contextual Bandits | Min-hwan Oh · Garud Iyengar |
Thompson Sampling with Information Relaxation Penalties | Seungki Min · Costis Maglaras · Ciamac C Moallemi |