@inproceedings{ca9e42da37ec4352818f7fbd8f62c5f9,
title = "Omega-Regular Objectives in Model-Free Reinforcement Learning",
abstract = "We provide the first solution for model-free reinforcement learning of ω -regular objectives for Markov decision processes (MDPs). We present a constructive reduction from the almost-sure satisfaction of ω -regular objectives to an almost-sure reachability problem, and extend this technique to learning how to control an unknown model so that the chance of satisfying the objective is maximized. We compile ω -regular properties into limit-deterministic B{\"u}chi automata instead of the traditional Rabin automata; this choice sidesteps difficulties that have marred previous proposals. Our approach allows us to apply model-free, off-the-shelf reinforcement learning algorithms to compute optimal strategies from the observations of the MDP. We present an experimental evaluation of our technique on benchmark learning problems.",
author = "Hahn, \{Ernst Moritz\} and Mateo Perez and Sven Schewe and Fabio Somenzi and Ashutosh Trivedi and University Liverpool",
year = "2019",
month = apr,
day = "4",
doi = "10.1007/978-3-030-17462-0\_27",
language = "English",
volume = "11427",
series = "Lecture Notes in Computer Science",
publisher = "Springer Lecture Notes in Computer Science (LNCS)",
pages = "395--412",
booktitle = "TACAS: International Conference on Tools and Algorithms for the Construction and Analysis of Systems",
}