2021
DOI: 10.48550/arxiv.2112.08360
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

How to Learn and Represent Abstractions: An Investigation using Symbolic Alchemy

Abstract: Alchemy is a new meta-learning environment rich enough to contain interesting abstractions, yet simple enough to make fine-grained analysis tractable. Further, Alchemy provides an optional symbolic interface that enables meta-RL research without a large compute budget. In this work, we take the first steps toward using Symbolic Alchemy to identify design choices that enable deep-RL agents to learn various types of abstraction. Then, using a variety of behavioral and introspective analyses we investigate how ou… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
0
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(3 citation statements)
references
References 11 publications
0
0
0
Order By: Relevance
“…To test whether the general pattern of results described above hold for other kinds of neural network architectures, we repeated our experiments for three more architectures (Figs 4 and S7): Episodic Planning Networks [43,44], transformers [45], and Compositional Relational Networks [46]. Below is a description of each architecture.…”
Section: Performance Of Other Neural Network Architecturesmentioning
confidence: 99%
See 2 more Smart Citations
“…To test whether the general pattern of results described above hold for other kinds of neural network architectures, we repeated our experiments for three more architectures (Figs 4 and S7): Episodic Planning Networks [43,44], transformers [45], and Compositional Relational Networks [46]. Below is a description of each architecture.…”
Section: Performance Of Other Neural Network Architecturesmentioning
confidence: 99%
“…Episodic Planning Networks augment recurrent-based reinforcement learning agents [26] with an external episodic memory implemented through a self-attention mechanism [44]. Our implementation was closely inspired by the implementation of AlKhammasi et al [43]. It is an augmentation of the LSTM-based meta-learner [26] used in our previous experiment with one additional input.…”
Section: Performance Of Other Neural Network Architecturesmentioning
confidence: 99%
See 1 more Smart Citation