Multi-armed bandits

Multinomial Thompson Sampling for Online Sequential Decision Making with Rating Scales (Invited Seminar @ Federico II di Napoli)

Multi-armed bandit algorithms such as Thompson sampling (TS) have been put forth for decades as useful tools for optimizing sequential decision-making in online experiments. By skewing the allocation ratio towards superior arms, they can minimize …

Adaptive Experiments for Enhancing Digital Education -- Benefits and Statistical Challenges (Talk @ ICNA-STA2023)

Adaptive digital field experiments are continually increasing in their breadth of use in fields like mobile health and digital education. Using adaptive experimentation in education can help not only to explore and eventually compare various arms but …