2019 IEEE International Conference on Healthcare Informatics (ICHI) 2019
DOI: 10.1109/ichi.2019.8904713
|View full text |Cite
|
Sign up to set email alerts
|

Multimodal Attention Network for Trauma Activity Recognition from Spoken Language and Environmental Sound

Abstract: Trauma activity recognition aims to detect, recognize, and predict the activities (or tasks) during a trauma resuscitation. Previous work has mainly focused on using various sensor data including image, RFID, and vital signals to generate the trauma event log. However, spoken language and environmental sound, which contain rich communication and contextual information necessary for trauma team cooperation, are still largely ignored. In this paper, we propose a multimodal attention network (MAN) that uses both … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
16
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
2
2
1

Relationship

2
3

Authors

Journals

citations
Cited by 5 publications
(16 citation statements)
references
References 15 publications
0
16
0
Order By: Relevance
“…Text-based activity recognition employed the transcript of the verbal communication between the medical team to predict the activity type. Recent research applied a multi-head attention architecture [13] to predict a speech-reliant activity from the transcripts and the environmental sound [6]. The drawback in this approach is that obtaining the text requires additional automatic speech recognition (ASR).…”
Section: Related Workmentioning
confidence: 99%
See 4 more Smart Citations
“…Text-based activity recognition employed the transcript of the verbal communication between the medical team to predict the activity type. Recent research applied a multi-head attention architecture [13] to predict a speech-reliant activity from the transcripts and the environmental sound [6]. The drawback in this approach is that obtaining the text requires additional automatic speech recognition (ASR).…”
Section: Related Workmentioning
confidence: 99%
“…The audio modality was used as an auxiliary to other modalities in works [5], [6]. These papers analyzed the audio ability to improve the accuracy of the activity recognition.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations