2017
DOI: 10.1007/978-3-319-54427-4_22
|View full text |Cite
|
Sign up to set email alerts
|

Multi-view Automatic Lip-Reading Using Neural Network

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

5
53
1

Year Published

2017
2017
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 47 publications
(59 citation statements)
references
References 18 publications
5
53
1
Order By: Relevance
“…The most common confusion pair 2 for the OuluVS2 dataset is between "Hello" (3rd phrase) and "Thank you" (8th phrase) which is consistent with confusions presented in (Petridis et al, 2017a;Lee et al, 2016). The most frequently confused pairs in the CUAVE dataset are zero and two, and six and nine and this is consistent with (Petridis et al, 2017a).…”
Section: Methodssupporting
confidence: 75%
“…The most common confusion pair 2 for the OuluVS2 dataset is between "Hello" (3rd phrase) and "Thank you" (8th phrase) which is consistent with confusions presented in (Petridis et al, 2017a;Lee et al, 2016). The most frequently confused pairs in the CUAVE dataset are zero and two, and six and nine and this is consistent with (Petridis et al, 2017a).…”
Section: Methodssupporting
confidence: 75%
“…The only combination which outperformed the frontal view is frontal + 30 • . Similarly to [19], the combination of all views led to worse performance than most individual views.…”
Section: Related Workmentioning
confidence: 97%
“…The frontal view was found to be the best by Saitoh et al [29], the profile view by Lee at al. [19], the 30 • view by Zimmermann et al [34] and the 60 • view was found to be the best performing in Three different convolutional neural networks (CNNs), GoogLeNet, AlexNet and Network in Network, were trained on OuluVS2 using data augmentation. Each model led to different performance across the views.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…The final individual weights are a combination of these coarser weights. Finally, some recent research has applied cross-view analysis to 3D-AAMs [30] and used channel, image and feature fusion for multiple-and cross-view analysis [31].…”
Section: Related Workmentioning
confidence: 99%