2017 3rd International Conference on Pattern Recognition and Image Analysis (IPRIA) 2017
DOI: 10.1109/pria.2017.7983045
|View full text |Cite
|
Sign up to set email alerts
|

Lip-reading via a DNN-HMM hybrid system using combination of the image-based and model-based features

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
9
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 28 publications
(10 citation statements)
references
References 8 publications
0
9
0
Order By: Relevance
“…viseme appearance features, were utilized in our system, resulting in the base algorithm of accuracy 69.8%. It should be noted that our deep CNN classifier outperforms the CNN model presented in [7], where accuracy of 76.6% is attained in [7], in spite of using large viseme dataset. In our work, the accuracy is www.ijacsa.thesai.org increased by about 21%.…”
Section: Discussionmentioning
confidence: 83%
See 1 more Smart Citation
“…viseme appearance features, were utilized in our system, resulting in the base algorithm of accuracy 69.8%. It should be noted that our deep CNN classifier outperforms the CNN model presented in [7], where accuracy of 76.6% is attained in [7], in spite of using large viseme dataset. In our work, the accuracy is www.ijacsa.thesai.org increased by about 21%.…”
Section: Discussionmentioning
confidence: 83%
“…The authors in [7], proposed lip viseme modalities through multimodal learning methods. In [8], two deep CNNs are trained using text labels and video frames and their final layers are fused to extract mutual features to be classified by a deep network.…”
Section: Introductionmentioning
confidence: 99%
“…Ngiam et al [6] achieved a 68.70% word recognition rate using an RBM-Autoencoder. Rahmani [141] extracted deep bottleneck features, and then used a GMM-HMM for the language model to achieve a WAR of 63.40%. Petridis et al [78] achieved a WAR of 78.60% using the dual flow method.…”
Section: Performances In Lip-readingmentioning
confidence: 99%
“…A HMM model which predicts septic shock for ICU patients [10]. Rehabilitation of a deaf person is done by a DNN-HMM hybrid system for lip-reading and audio visual speech recognition (AVSR) [11], fall detection and real-life home monitoring for senior citizens [12], emotion classification by a combined SVM-HMM classifier to recognize human emotion states based on EEG signals [13]. Already a stacked HMM model has found its applications in robotics for motion intention recognition based on motion trajectories [14].…”
Section: Introductionmentioning
confidence: 99%