Timezone: »

Learning Multiple Markov Chains via Adaptive Allocation
Mohammad Sadegh Talebi · Odalric-Ambrym Maillard

Tue Dec 10 05:30 PM -- 07:30 PM (PST) @ East Exhibition Hall B + C #14

We study the problem of learning the transition matrices of a set of Markov chains from a single stream of observations on each chain. We assume that the Markov chains are ergodic but otherwise unknown. The learner can sample Markov chains sequentially to observe their states. The goal of the learner is to sequentially select various chains to learn transition matrices uniformly well with respect to some loss function. We introduce a notion of loss that naturally extends the squared loss for learning distributions to the case of Markov chains, and further characterize the notion of being \emph{uniformly good} in all problem instances. We present a novel learning algorithm that efficiently balances \emph{exploration} and \emph{exploitation} intrinsic to this problem, without any prior knowledge of the chains. We provide finite-sample PAC-type guarantees on the performance of the algorithm. Further, we show that our algorithm asymptotically attains an optimal loss.

Author Information

Mohammad Sadegh Talebi (Inria)
Odalric-Ambrym Maillard (INRIA)

More from the Same Authors