Model-Free Reinforcement Learning for Lexicographic Omega-Regular Objectives

Ernst Moritz Hahn, Mateo Perez, Sven Schewe, Fabio Somenzi, Ashutosh Trivedi, Dominik Wojtczak

Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

6 Citations (Scopus)
96 Downloads (Pure)

Abstract

We study the problem of finding optimal strategies in Markov decision processes with lexicographic ω -regular objectives, which are ordered collections of ordinary ω -regular objectives. The goal is to compute strategies that maximise the probability of satisfaction of the first ω -regular objective; subject to that, the strategy should also maximise the probability of satisfaction of the second ω -regular objective; then the third and so forth. For instance, one may want to guarantee critical requirements first, functional ones second and only then focus on the non-functional ones. We show how to harness the classic off-the-shelf model-free reinforcement learning techniques to solve this problem and evaluate their performance on four case studies.
Original languageEnglish
Title of host publication24th International Symposium on Formal Methods, FM 2021
Subtitle of host publicationVirtual Event, November 20–26, 2021, Proceedings
EditorsMarieke Huisman, Corina S. Pasareanu, Naijun Zhan
PublisherSpringer
Pages142-159
Number of pages18
DOIs
Publication statusPublished - 10 Nov 2021
Event24th International Symposium on Formal Methods, FM 2021 - Virtual, Online
Duration: 20 Nov 202126 Nov 2021
Conference number: 24

Publication series

NameLecture Notes in Computer Science
PublisherSpringer
Volume13047

Conference

Conference24th International Symposium on Formal Methods, FM 2021
Abbreviated titleFM 2021
CityVirtual, Online
Period20/11/2126/11/21

Keywords

  • 22/1 OA procedure

Fingerprint

Dive into the research topics of 'Model-Free Reinforcement Learning for Lexicographic Omega-Regular Objectives'. Together they form a unique fingerprint.

Cite this