2020
DOI: 10.1007/s12145-020-00516-y
|View full text |Cite
|
Sign up to set email alerts
|

Person identification with aerial imaginary using SegNet based semantic segmentation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
10
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
7
2
1

Relationship

0
10

Authors

Journals

citations
Cited by 20 publications
(10 citation statements)
references
References 20 publications
0
10
0
Order By: Relevance
“…Therefore, training of U-Net is relatively slow compared to SegNet ( 59 ). Moreover, in the study by Manickam et al ( 60 ), in which semantic human detection from UAV images was made, SegNet provided more successful results than many other deep models. Similarly, in another study for brain tissue segmentation by Kumar et al ( 59 ), SegNet provided more successful classification than U-Net.…”
Section: Experimental Studies and Resultsmentioning
confidence: 98%
“…Therefore, training of U-Net is relatively slow compared to SegNet ( 59 ). Moreover, in the study by Manickam et al ( 60 ), in which semantic human detection from UAV images was made, SegNet provided more successful results than many other deep models. Similarly, in another study for brain tissue segmentation by Kumar et al ( 59 ), SegNet provided more successful classification than U-Net.…”
Section: Experimental Studies and Resultsmentioning
confidence: 98%
“…The SegNet model outperformed the other two models and had an overall segmentation accuracy of 92.7%. Similarly, Manickam et al [69] compared four models (VGG16, GoogleNet, ResNet, and SegNet) for personal identification with aerial imaginary. Their results revealed that SegNet achieved the highest overall accuracy of 91.04%.…”
Section: Evaluation Of Segmentation Methodsmentioning
confidence: 99%
“…In the Encoder Network Layer of SegNet, the pre-trained VGG16 network is used to remove the full connection layer, leaving only the previous convolutional layers as the coding layer. The main reason for this is to maintain higher resolution and reduce network parameters to improve training efficiency [47]. There is a novel Hierarchical Transformer Encoder in the SegFormer network architecture, which enables the network to output multi-scale features and does not require location coding.…”
Section: Classifier Model Architecturesmentioning
confidence: 99%