• Home
  • An Information-Theoretic Analysis of Nonstationary Bandit Learning - NUS Business School
An Information-Theoretic Analysis of Nonstationary Bandit Learning
In "Seminars and talks"

Speakers

Seungki Min
Seungki Min

Assistant Professor, Seoul National University Business School

Seungki Min is an Assistant Professor of Operations Management at Seoul National University Business School. His research focuses on bandit optimization and reinforcement learning, with an emphasis on principled frameworks for dynamic decision problems arising in business and engineering applications, including online platforms, pricing, and finance. His research has appeared in Operations Research, Management Science, and leading AI/ML conferences such as ICML and NeurIPS. He earned his Ph.D. from Columbia Business School. Prior to academia, he worked in high-frequency trading domain.


Date:
Friday, 30 January 2026
Time:
10:00 am - 11:30 am
Venue:
NUS Business School
Mochtar Riady Building BIZ1 0302
15 Kent Ridge Drive
Singapore 119245 (Map)

Abstract

In many real-world bandit learning problems, the underlying environment evolves over time, requiring decision-makers to continually acquire information and adapt their action selection accordingly. In this talk, I study Bayesian formulations of nonstationary bandit problems, where environmental dynamics are modeled as stochastic processes, and develop an information-theoretic framework for analyzing attainable performance.

Our analysis yields generic regret upper bounds that extend classical results from stationary Bayesian bandits to nonstationary settings. A key insight is that the entropy rate of the optimal action process naturally quantifies the intrinsic difficulty introduced by nonstationarity. I further connect our results to existing frequentist analyses of nonstationary bandits, showing that several well-known regret bounds in the literature can be recovered as special cases within our unified framework.