2020
DOI: 10.3390/app10093079
|View full text |Cite
|
Sign up to set email alerts
|

Optimized YOLOv3 Algorithm and Its Application in Traffic Flow Detections

Abstract: In the intelligent traffic system, real-time and accurate detections of vehicles in images and video data are very important and challenging work. Especially in situations with complex scenes, different models, and high density, it is difficult to accurately locate and classify these vehicles during traffic flows. Therefore, we propose a single-stage deep neural network YOLOv3-DL, which is based on the Tensorflow framework to improve this problem. The network structure is optimized by introducing the idea of s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
42
0
3

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 92 publications
(45 citation statements)
references
References 18 publications
0
42
0
3
Order By: Relevance
“…However, the MSE loss function fails to reflect the relationship between the information, but uses it as an independent variable. In order to improve it, IoU loss is proposed, which considers the area of the predicted bounding box (BBOX) and the ground truth bounding box [28,29]. YOLOv4 uses CIoU loss instead of MSE loss, which includes the shape and direction of the object and also considers the overlap area, the distance between the center points, and the aspect ratio, which are defined as follows.…”
Section: The Loss Of Yolov4mentioning
confidence: 99%
“…However, the MSE loss function fails to reflect the relationship between the information, but uses it as an independent variable. In order to improve it, IoU loss is proposed, which considers the area of the predicted bounding box (BBOX) and the ground truth bounding box [28,29]. YOLOv4 uses CIoU loss instead of MSE loss, which includes the shape and direction of the object and also considers the overlap area, the distance between the center points, and the aspect ratio, which are defined as follows.…”
Section: The Loss Of Yolov4mentioning
confidence: 99%
“…Fifty videos were recorded. The length of each collected video was 40 s, a value chosen according to related studies [78]. The videos are 34.25 FPS and were captured with an EOS 550D camera at four different locations, under three occlusion statuses.…”
Section: Validation With Real-time Videosmentioning
confidence: 99%
“…The pioneering work of region-based target detection began with the region-based convolutional neural network (R-CNN), including three modules: regional proposal, vector transformation, and classification [ 15 , 16 ]. Spatial pyramid pooling (SPP)-net optimized the R-CNN and improved detection performance [ 16 , 17 ]. Fast R-CNN combines the essence of SPP-net and R-CNN, and introduces a multi-task loss function, which is what makes the training and testing of the whole network so functional [ 16 , 18 ].…”
Section: Introductionmentioning
confidence: 99%