2016
DOI: 10.1371/journal.pone.0154486
|View full text |Cite
|
Sign up to set email alerts
|

I Hear You Eat and Speak: Automatic Recognition of Eating Condition and Food Type, Use-Cases, and Impact on ASR Performance

Abstract: We propose a new recognition task in the area of computational paralinguistics: automatic recognition of eating conditions in speech, i. e., whether people are eating while speaking, and what they are eating. To this end, we introduce the audio-visual iHEARu-EAT database featuring 1.6 k utterances of 30 subjects (mean age: 26.1 years, standard deviation: 2.66 years, gender balanced, German speakers), six types of food (Apple, Nectarine, Banana, Haribo Smurfs, Biscuit, and Crisps), and read as well as spontaneo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
39
0
1

Year Published

2016
2016
2020
2020

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 38 publications
(40 citation statements)
references
References 61 publications
0
39
0
1
Order By: Relevance
“…We performed our experiments on the iHEARu-EAT database [27], which contains the utterances of 30 people recorded while speaking during eating. Six types of food were used along with the "no food" class, resulting in seven classes overall.…”
Section: The Ihearu-eat Corpusmentioning
confidence: 99%
See 1 more Smart Citation
“…We performed our experiments on the iHEARu-EAT database [27], which contains the utterances of 30 people recorded while speaking during eating. Six types of food were used along with the "no food" class, resulting in seven classes overall.…”
Section: The Ihearu-eat Corpusmentioning
confidence: 99%
“…For each speaker and food type, seven utterances were recorded; some subjects refused to eat certain types of foods, resulting in a total of 1414 utterances in German. Although this dataset can be used primarily to test machine learning and signal processing techniques, Hantke et al also anticipated several possible future applications [27]. This dataset was later used in the Interspeech ComParE 2015 Eating Condition Sub-Challenge [15].…”
Section: The Ihearu-eat Corpusmentioning
confidence: 99%
“…Degree of Nativeness (DN) Sub-Challenge В аудиофайлах из базы данных iHEARu-EAT [52,60], предоставленной организаторами из Германии, присутствовал один из 6 типов пищи, которую человек ел во время разговора, или она отсутствовала: 1. яблоко (жесткий фрукт), класс «Apple»; 2. банан (мягкий фрукт), класс «Banana»; 3. нектарин (полумягкий фрукт), класс «Nectarine»; 4. бисквит, класс «Biscuit»; 5. чипсы (хрустящая еда), класс «Crisp»; 6. жевательные конфеты, класс «Haribo», а также 7. без пищи во рту, класс «No Food».…”
Section: конкурсunclassified
“…Whereas Automatic Speech Recognition (ASR) tries to determine which words are spoken, CP attempts to discover how those words are spoken and thereby gain knowledge about the various aspects and conditions of the speakers, e.g., age, gender, sleepiness, friendliness, etc. Hantke et al, 2016). The Sleepiness Sub-Challenge employs the openSMILE toolkit (Eyben et al, 2010) to generate the 4368 baseline acoustic features from the "Sleepy Language Corpus" (SLC) (Schuller et al, 2011).…”
Section: Introductionmentioning
confidence: 99%