Toward Assured Autonomy with Model-Free Reinforcement Learning

dc.contributor.advisor

Pajic, Miroslav

dc.contributor.author

Bozkurt, Alper Kamil

dc.date.accessioned

2024-06-06T13:44:11Z

dc.date.available

2024-06-06T13:44:11Z

dc.date.issued

2024

dc.department

Computer Science

dc.description.abstract

Autonomous systems (AS), enhanced by the capabilities of reinforcement learning (RL), are expected to perform increasingly sophisticated tasks across various civilian and industrial application domains. This expectation arises from their promising ability to make decisions solely based on perception without human intervention. In addition to high efficiency, AS often require robustness and safety guarantees for real-world deployment. In this thesis, we propose model-free RL approaches that obtain controllers for AS operating in unknown, stochastic, and potentially adversarial environments directly from linear temporal logic (LTL) specifications defined on state labels, such as safety and liveness requirements. This ensures that the learned controllers satisfy the desired properties, avoiding unintended consequences, and remain robust against adversarial behavior.We first derive a novel rewarding and discounting mechanism from the LTL specifications for Markov decision processes. We show that a policy learned by a model-free RL algorithm, which maximizes the sum of these discounted rewards, also maximizes the probability of satisfying the LTL specifications. We generalize this approach to multiple objectives, where the utmost priority is given to ensuring safety. Satisfaction of the other LTL specifications takes a secondary role, and the tertiary objective is to enhance the quality of control. We then extend our results to zero-sum stochastic games to ensure the robustness of learned controllers against any unpredictable nondeterministic environment behavior. Addressing the scalability challenges inherent in learning controllers for stochastic games, we propose heuristics and approximate methods to further accelerate the learning process. We illustrate how our approach can be utilized to learn controllers that are resilient against stealthy attackers, capable of disrupting the agent's actuation without being detected. We further discuss an approach for cases where state labels are absent. This approach aims to learn a labeling function that translates raw state information into object properties applicable in LTL specifications, thereby enabling the learning of controllers from LTL specifications. We conclusively show the effectiveness of our approaches in successfully learning optimal controllers through numerous case studies. These controllers maximize the probability of satisfying LTL specifications in the worst case, thereby exhibiting resilience against adversarial behavior. Moreover, our methods demonstrate scalability across a broad spectrum of LTL specifications, consistently surpassing the performance of existing approaches.

dc.identifier.uri

https://hdl.handle.net/10161/30808

dc.rights.uri

https://creativecommons.org/licenses/by-nc-nd/4.0/

dc.subject

Artificial intelligence

dc.subject

Controller Synthesis

dc.subject

Linear Temporal Logic

dc.subject

Reinforcement Learning

dc.subject

Stochastic Games

dc.title

Toward Assured Autonomy with Model-Free Reinforcement Learning

dc.type

Dissertation

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Bozkurt_duke_0066D_17751.pdf
Size:
2.86 MB
Format:
Adobe Portable Document Format

Collections