2021
DOI: 10.48550/arxiv.2102.07456
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Neuro-algorithmic Policies enable Fast Combinatorial Generalization

Marin Vlastelica,
Michal Rolínek,
Georg Martius
Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(3 citation statements)
references
References 11 publications
0
3
0
Order By: Relevance
“…As shown in Tables 2 and 3, most methods focus on changing the loss function or algorithmic approach. Architectural changes informed by inductive biases are less well studied, with notable examples coming from the work of Cobbe et al (2020b), Singh and Zheng (2021), Tang et al (2020), Vlastelica et al (2021), Zambaldi et al (2019). More work can be done on investigating different architectures, either taking inspiration from supervised learning or creating RL-specific architectures.…”
Section: Future Work On Methods For Zero-shot Generalisationmentioning
confidence: 99%
“…As shown in Tables 2 and 3, most methods focus on changing the loss function or algorithmic approach. Architectural changes informed by inductive biases are less well studied, with notable examples coming from the work of Cobbe et al (2020b), Singh and Zheng (2021), Tang et al (2020), Vlastelica et al (2021), Zambaldi et al (2019). More work can be done on investigating different architectures, either taking inspiration from supervised learning or creating RL-specific architectures.…”
Section: Future Work On Methods For Zero-shot Generalisationmentioning
confidence: 99%
“…As shown in Table 2, most methods focus on changing the loss function or algorithmic approach. Architectural changes informed by inductive biases are less well studied, with notable examples being [154,155,54,124,91]. More work can be done on investigating different architectures, either taking inspiration from supervised learning or creating RL-specific architectures.…”
Section: Future Work On Methods For Generalisationmentioning
confidence: 99%
“…Higgins et al [15,DARLA] uses β-VAEs [123] to encode the inductive bias of disentanglement into the representations of the policy, improving zero-shot performance on various visual variations. Vlastelica et al [124,NAP] incorporates a black-box shortest-path solver to improve generalisation performance in hard navigation problems. Zambaldi et al [125,91] incorporate a relational inductive bias into the model architecture which aids in generalising along ordinal axes of variation, including extrapolation performance.…”
Section: Handling Differences Between Training and Testingmentioning
confidence: 99%