Transition Space Distance Learning

dc.contributor.advisor

Parr, Ronald

dc.contributor.author

Nemecek, Mark William

dc.date.accessioned

2019-06-07T19:51:03Z

dc.date.available

2019-06-07T19:51:03Z

dc.date.issued

2019

dc.department

Computer Science

dc.description.abstract

The notion of distance plays and important role in many reinforcement learning (RL) techniques. This role may be explicit, as in some non-parametric approaches, or it may be implicit in the architecture of the feature space. The ability to learn distance functions tailored for RL tasks could, thus, benefit many different RL paradigms. While several approaches to learning distance functions from data do exist, they are frequently intended for use in clustering or classification tasks and typically do not take into account the inherent structure present in trajectories sampled from RL environments. For those that do, this structure is generally used to define a similarity between states rather than to represent the mechanics of the domain. Based on the idea that a good distance function in such a domain would reflect the number of transitions necessary to get to from one state to another, we detail an approach to learning distance functions which accounts for the nature of state transitions in a Markov decision process, including their inherent directionality. We then present the results of experiments performed in multiple RL environments in order to demonstrate the benefit of learning such distance functions.

dc.identifier.uri

https://hdl.handle.net/10161/18849

dc.subject

Artificial intelligence

dc.subject

Computer science

dc.subject

Distance learning

dc.subject

Feature extraction

dc.subject

Reinforcement learning

dc.title

Transition Space Distance Learning

dc.type

Master's thesis

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Nemecek_duke_0066N_15002.pdf
Size:
2.76 MB
Format:
Adobe Portable Document Format

Collections