2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) 2020
DOI: 10.1109/iros45743.2020.9341325
|View full text |Cite
|
Sign up to set email alerts
|

Encoding formulas as deep networks: Reinforcement learning for zero-shot execution of LTL formulas

Abstract: We demonstrate a reinforcement learning agent which uses a compositional recurrent neural network that takes as input an LTL formula and determines satisfying actions. The input LTL formulas have never been seen before, yet the network performs zero-shot generalization to satisfy them. This is a novel form of multi-task learning for RL agents where agents learn from one diverse set of tasks and generalize to a new set of diverse tasks. The formulation of the network enables this capacity to generalize. We demo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
13
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 14 publications
(13 citation statements)
references
References 13 publications
0
13
0
Order By: Relevance
“…Camacho et al [8] show that one can generate RMs from temporal specifications but RMs generated this way lead to sparse rewards. Kuo et al [23] propose a compositional model for zero-shot execution of LTL formulas but training such a model requires a lot of samples even for relatively simpler environments. There has also been recent work on using temporal logic specifications for multi-agent RL [13,29].…”
Section: Related Workmentioning
confidence: 99%
“…Camacho et al [8] show that one can generate RMs from temporal specifications but RMs generated this way lead to sparse rewards. Kuo et al [23] propose a compositional model for zero-shot execution of LTL formulas but training such a model requires a lot of samples even for relatively simpler environments. There has also been recent work on using temporal logic specifications for multi-agent RL [13,29].…”
Section: Related Workmentioning
confidence: 99%
“…Formally, the first component of the SM is the extractor E, which transforms the complex formula T into a list K consisting of all the sequences of atomic tasks α that satisfy T . As it is common in the literature (Kuo et al, 2020;Vaezipoor et al, 2021), we assume that the SM have access to an internal labelling function L I : Z → 2 AP . L I is the restriction of L to the observation space and maps the observations of the agent in Z into the set AP of atoms.…”
Section: Neuro-symbolic Agentsmentioning
confidence: 99%
“…Beyond TL, we find methods such as reward machines (a type of finite state machine) Xu et al, 2020;, or RL-specific formal languages such as SPECTRL (Jothimurugan et al, 2019). Closer to our line of work, Kuo et al (2020) presents a novel RL framework to follow OOD combinations of known tasks expressed in linear-time temporal logic (LTL) by training multiple networks (one per LTL operator and per object). In a similar line, Araki et al (2021) introduces a hierarchical reinforcement learning framework aimed to learn policies that are optimal and composable while relying on different neural networks each specialized in one subtask.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…This manuscript is an extension of “Encoding formulas as deep networks: Reinforcement learning for zero-shot execution of LTL formulas” ( Kuo et al, 2020b ) by the same authors published at the Conference on Intelligent Robots and Systems (IROS) 2020. The manuscript contains over 30% new material, including additional technical details and a new domain, Fetch, which required substantive advances.…”
mentioning
confidence: 99%