Learning Representations With Linear-Algebraic Structure
dc.contributor.advisor | Ge, Rong | |
dc.contributor.author | Frandsen, Abraham | |
dc.date.accessioned | 2022-06-15T18:43:37Z | |
dc.date.available | 2023-05-26T08:17:14Z | |
dc.date.issued | 2022 | |
dc.department | Computer Science | |
dc.description.abstract | Representation learning is a key step for enabling algorithms to make sense of data and output good decisions. Good data representations preserve useful information, discard irrelevant features, and simplify complex relationships between data. We approach the problem of representation learning through the lens of latent variable models. In such models, the representations are directly given as unobserved variables encoding the core structure of the data. We utilize linear algebraic structure to specify the properties of the representations, which enables rigorous analysis and efficient, provable algorithms. We first consider the area of natural language processing, where the data are comprised of words. We propose a novel model for word representations that encodes compositional syntactic and semantic structure as latent multilinear structure. We prove that the representations can be efficiently recovered and develop a practical learning algorithm. We show that learning the word embedding model is closely connected to the Tucker decomposition, an important basic operation in tensor analysis that also arises in the context of other latent variable models. We formulate the Tucker decomposition as a nonconvex optimization problem and prove that its landscape is benign. We then give a local search algorithm that provably finds the global optimum. We finally consider the area of reinforcement learning and control, where the time dynamics of the data are vital. We propose a model in which the state observations are high-dimensional with nonlinear dynamics, but depend on a latent low-dimensional linear control system. We develop state representation learning algorithms based on both forward and inverse dynamics that provably and efficiently recover the hidden linear system. | |
dc.identifier.uri | ||
dc.subject | Computer science | |
dc.title | Learning Representations With Linear-Algebraic Structure | |
dc.type | Dissertation | |
duke.embargo.months | 11.342465753424657 |
Files
Original bundle
- Name:
- Frandsen_duke_0066D_16663.pdf
- Size:
- 967.2 KB
- Format:
- Adobe Portable Document Format