搜索结果: 1-1 共查到“统计学 Markov Bandits”相关记录1条 . 查询时间(0.125 秒)
We consider the restless Markov bandit problem, in which the state of each arm evolves according to a Markov process independently of the learner's actions. We suggest an algorithm that after $T$ step...