2021
DOI: 10.1109/tac.2020.3006967
|View full text |Cite
|
Sign up to set email alerts
|

Learning-Based Probabilistic LTL Motion Planning With Environment and Motion Uncertainties

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
21
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 50 publications
(21 citation statements)
references
References 28 publications
0
21
0
Order By: Relevance
“…where U tr π and U rec π are the expected return of states in transient and recurrent classes under policy π, respectively. In (13), P π (T , T ) ∈ R r×r is the probability transition matrix between states in T π , and…”
Section: Base Rewardmentioning
confidence: 99%
See 1 more Smart Citation
“…where U tr π and U rec π are the expected return of states in transient and recurrent classes under policy π, respectively. In (13), P π (T , T ) ∈ R r×r is the probability transition matrix between states in T π , and…”
Section: Base Rewardmentioning
confidence: 99%
“…However, scalability is a pressing issue for applying model-based approaches due to the need to store the learned model. On the other hand, by relaxing the need to construct an MDP model, model-free RL is recently adopted where appropriate reward shaping schemes are proposed [12][13][14][15][16][17][18][19][20].…”
Section: Introductionmentioning
confidence: 99%
“…Puri elaborated on some of the drones available in the market that can be used for agricultural monitoring and observation to get better crop quality and to prevent various forms of damage to the fields (Puri et al, 2017). Currently, most human-robot skill systems focus on the service robot (Hu et al, 2019), manufacturing (Cai et al, 2020), medical rehabilitation (Su et al, 2020c) and robotic teleoperation (Yang et al, 2018c). The research of the humanrobot skill transfer system in agriculture still is one of the most challenging (Yingbai et al, 2020).…”
Section: Introductionmentioning
confidence: 99%
“…On reinforcement learning with MDP, various methods exist to improve its behavior and performance for safety objectives (e.g. [24]), to deal with unknown stochastic behavior [5,25], and with linear-time logic specifications (e.g. [5,21,45]).…”
Section: Introductionmentioning
confidence: 99%
“…[24]), to deal with unknown stochastic behavior [5,25], and with linear-time logic specifications (e.g. [5,21,45]).…”
Section: Introductionmentioning
confidence: 99%