| Blocking Bandits | Soumya Basu · Rajat Sen · Sujay Sanghavi · Sanjay Shakkottai |
| Combinatorial Bandits with Relative Feedback | Aadirupa Saha · Aditya Gopalan |
| Decentralized Cooperative Stochastic Bandits | David Martínez-Rubio · Varun Kanade · Patrick Rebeschini |
| Doubly-Robust Lasso Bandit | Gi-Soo Kim · Myunghee Cho Paik |
| Efficient Pure Exploration in Adaptive Round model | tianyuan jin · Jieming SHI · Xiaokui Xiao · Enhong Chen |
| Epsilon-Best-Arm Identification in Pay-Per-Reward Multi-Armed Bandits | Sivan Sabato |
| Nonstochastic Multiarmed Bandits with Unrestricted Delays | Tobias Sommer Thune · Nicolò Cesa-Bianchi · Yevgeny Seldin |
| On the Optimality of Perturbations in Stochastic and Adversarial Multi-armed Bandit Problems | Baekjin Kim · Ambuj Tewari |
| Phase Transitions and Cyclic Phenomena in Bandits with Switching Constraints | David Simchi-Levi · Yunzong Xu |
| Polynomial Cost of Adaptation for X-Armed Bandits | Hedi Hadiji |
| Semi-Parametric Dynamic Contextual Pricing | Virag Shah · Ramesh Johari · Jose Blanchet |
| Bayesian Optimization under Heavy-tailed Payoffs | Sayak Ray Chowdhury · Aditya Gopalan |
| Connections Between Mirror Descent, Thompson Sampling and the Information Ratio | Julian Zimmert · Tor Lattimore |
| Individual Regret in Cooperative Nonstochastic Multi-Armed Bandits | Yogev Bar-On · Yishay Mansour |
| Learning Multiple Markov Chains via Adaptive Allocation | Mohammad Sadegh Talebi · Odalric-Ambrym Maillard |
| Linear Stochastic Bandits Under Safety Constraints | Sanae Amani · Mahnoosh Alizadeh · Christos Thrampoulidis |
| Personalizing Many Decisions with High-Dimensional Covariates | Nima Hamidi · Mohsen Bayati · Kapil Gupta |
| Non-Asymptotic Pure Exploration by Solving Games | Rémy Degenne · Wouter Koolen · Pierre Ménard |
| Online EXP3 Learning in Adversarial Bandits with Delayed Feedback | Ilai Bistritz · Zhengyuan Zhou · Xi Chen · Nicholas Bambos · Jose Blanchet |
| Optimal Best Markovian Arm Identification with Fixed Confidence | Vrettos Moulos |
| Oracle-Efficient Algorithms for Online Linear Optimization with Bandit Feedback | Shinji Ito · Daisuke Hatano · Hanna Sumita · Kei Takemura · Takuro Fukunaga · Naonori Kakimura · Ken-Ichi Kawarabayashi |
| Regret Bounds for Thompson Sampling in Episodic Restless Bandit Problems | Young Hun Jung · Ambuj Tewari |
| Thresholding Bandit with Optimal Aggregate Regret | Chao Tao · Saúl Blanco · Jian Peng · Yuan Zhou |
| Weighted Linear Bandits for Non-Stationary Environments | Yoan Russac · Claire Vernade · Olivier Cappé |
| Improved Regret Bounds for Bandit Combinatorial Optimization | Shinji Ito · Daisuke Hatano · Hanna Sumita · Kei Takemura · Takuro Fukunaga · Naonori Kakimura · Ken-Ichi Kawarabayashi |
| Learning in Generalized Linear Contextual Bandits with Stochastic Delays | Zhengyuan Zhou · Renyuan Xu · Jose Blanchet |
| Low-Rank Bandit Methods for High-Dimensional Dynamic Pricing | Jonas Mueller · Vasilis Syrgkanis · Matt Taddy |
| MaxGap Bandit: Adaptive Algorithms for Approximate Ranking | Sumeet Katariya · Ardhendu Tripathy · Robert Nowak |
| Model Selection for Contextual Bandits | Dylan Foster · Akshay Krishnamurthy · Haipeng Luo |
| No-Regret Learning in Unknown Games with Correlated Payoffs | Pier Giuseppe Sessa · Ilija Bogunovic · Maryam Kamgarpour · Andreas Krause |
| Nonparametric Contextual Bandits in Metric Spaces with Unknown Metric | Nirandika Wanigasekara · Christina Yu |
| Pure Exploration with Multiple Correct Answers | Rémy Degenne · Wouter Koolen |
| Recovering Bandits | Ciara Pike-Burke · Steffen Grunewalder |
| Sequential Experimental Design for Transductive Linear Bandits | Lalit Jain · Kevin Jamieson · Tanner Fiez · Lillian Ratliff |
| SIC-MMAB: Synchronisation Involves Communication in Multiplayer Multi-Armed Bandits | Etienne Boursier · Vianney Perchet |
| Are sample means in multi-armed bandits positively or negatively biased? | Jaehyeok Shin · Aaditya Ramdas · Alessandro Rinaldo |
| A New Perspective on Pool-Based Active Classification and False-Discovery Control | Lalit Jain · Kevin Jamieson |
| Bootstrapping Upper Confidence Bound | Botao Hao · Yasin Abbasi Yadkori · Zheng Wen · Guang Cheng |
| Categorized Bandits | Matthieu Jedor · Vianney Perchet · Jonathan Louedec |
| Censored Semi-Bandits: A Framework for Resource Allocation with Censored Feedback | Arun Verma · Manjesh Hanawal · Arun Rajkumar · Raman Sankaran |
| Contextual Bandits with Cross-Learning | Santiago Balseiro · Negin Golrezaei · Mohammad Mahdian · Vahab Mirrokni · Jon Schneider |
| Distribution oblivious, risk-aware algorithms for multi-armed bandits with unbounded rewards | Anmol Kagrecha · Jayakrishnan Nair · Krishna Jagannathan |
| Dynamic Incentive-Aware Learning: Robust Pricing in Contextual Auctions | Negin Golrezaei · Adel Javanmard · Vahab Mirrokni |
| Stochastic Bandits with Context Distributions | Johannes Kirschner · Andreas Krause |
| Thompson Sampling and Approximate Inference | My Phan · Yasin Abbasi Yadkori · Justin Domke |
| Thompson Sampling for Multinomial Logit Contextual Bandits | Min-hwan Oh · Garud Iyengar |
| Thompson Sampling with Information Relaxation Penalties | Seungki Min · Costis Maglaras · Ciamac C Moallemi |