2018
DOI: 10.1177/1729881418817162
|View full text |Cite
|
Sign up to set email alerts
|

Car-following method based on inverse reinforcement learning for autonomous vehicle decision-making

Abstract: There are still some problems need to be solved though there are a lot of achievements in the fields of automatic driving. One of those problems is the difficulty of designing a car-following decision-making system for complex traffic conditions. In recent years, reinforcement learning shows the potential in solving sequential decision optimization problems. In this article, we establish the reward function R of each driver data based on the inverse reinforcement learning algorithm, and r visualization is carr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
29
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
8
2

Relationship

0
10

Authors

Journals

citations
Cited by 55 publications
(29 citation statements)
references
References 27 publications
0
29
0
Order By: Relevance
“…Lifelong or continuous learning has been a long-standing challenge for machine learning and autonomous systems. [8][9][10] Mimicking humans and animals that continuously acquire new knowledge and transfer them to new tasks throughout their lifetime, continuous learning builds an adaptive system that is capable of learning from a continuous stream of information. However, dilemma between plasticity and catastrophic forgetting 11,12 is the main challenge due to inefficiency and poor performances when relearning from scratch for new tasks.…”
Section: Related Workmentioning
confidence: 99%
“…Lifelong or continuous learning has been a long-standing challenge for machine learning and autonomous systems. [8][9][10] Mimicking humans and animals that continuously acquire new knowledge and transfer them to new tasks throughout their lifetime, continuous learning builds an adaptive system that is capable of learning from a continuous stream of information. However, dilemma between plasticity and catastrophic forgetting 11,12 is the main challenge due to inefficiency and poor performances when relearning from scratch for new tasks.…”
Section: Related Workmentioning
confidence: 99%
“…Additionally, tests to measure performance quantitatively during these tasks are implemented. These applications integrate field tests and simulated tests—which are correlated—and whose assessment is conducted through quantitative performance indexes for each predefined task [ 13 , 14 , 15 ].…”
Section: Introductionmentioning
confidence: 99%
“…The error existing in the single time-step prediction will accumulate and gradually increase in the sequential decision-making process, which may cause the model to reach unseen states, making the model have even worse predictions. To avoid this problem, some researchers have begun to use reinforcement learning (RL) methods [ 11 , 14 ].…”
Section: Introductionmentioning
confidence: 99%