DOI: 10.1007/978-3-540-85853-9_14
|View full text |Cite
|
Sign up to set email alerts
|

Detection of Laughter-in-Interaction in Multichannel Close-Talk Microphone Recordings of Meetings

Abstract: Abstract. Laughter is a key element of human-human interaction, occurring surprisingly frequently in multi-party conversation. In meetings, laughter accounts for almost 10% of vocalization effort by time, and is known to be relevant for topic segmentation and the automatic characterization of affect. We present a system for the detection of laughter, and its attribution to specific participants, which relies on simultaneously decoding the vocal activity of all participants given multi-channel recordings. The p… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
27
0

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 22 publications
(28 citation statements)
references
References 21 publications
1
27
0
Order By: Relevance
“…In the first group, there are usually two approaches: 1) laughter detection/segmentation, e.g., [27], [29], [31], where the aim is to segment an unsegmented audio stream into laughter and nonlaughter episodes; 2) laughter-versus-speech classification/discrimination, e.g., [33], [57], [60], where the aim is to correctly classify presegmented episodes of laughter and speech. One of the first works on laughter detection is that of Kennedy and Ellis [27], who trained SVMs with MFCCs, spatial cues, and modulation spectrum features (MSFs) to detect group laughter, i.e., when more than a certain percentage of participants are laughing.…”
Section: B Automatic Laughter Classification/detectionmentioning
confidence: 99%
See 1 more Smart Citation
“…In the first group, there are usually two approaches: 1) laughter detection/segmentation, e.g., [27], [29], [31], where the aim is to segment an unsegmented audio stream into laughter and nonlaughter episodes; 2) laughter-versus-speech classification/discrimination, e.g., [33], [57], [60], where the aim is to correctly classify presegmented episodes of laughter and speech. One of the first works on laughter detection is that of Kennedy and Ellis [27], who trained SVMs with MFCCs, spatial cues, and modulation spectrum features (MSFs) to detect group laughter, i.e., when more than a certain percentage of participants are laughing.…”
Section: B Automatic Laughter Classification/detectionmentioning
confidence: 99%
“…GMMs were trained for speech, laughter, and silence, and the system was evaluated on the ICSI corpus achieving an EER of 10.9%. Laskowski and Schultz [31] present a system for the detection of laughter and its attribution to specific participants in multi-channel recordings. Each participant can be in one of the three states (silence, speech, laughter) and the aim is to decode the vocal activity of all participants simultaneously.…”
Section: B Automatic Laughter Classification/detectionmentioning
confidence: 99%
“…Knox and Mirghafori [94], Knox et al [95] used neural networks and HMMs in combination with MFCCs and prosodic features for laughter segmentation. Laskowski and Schultz [104] used a multiparticipant 3-state vocal activity recognition module to detect so-called laughter-in-interaction. Recently, audiovisual approaches to laughter detection have been undertaken by Reuderink et al [147] and Petridis and Pantic [130,131,132]: according to their work, fusion between the visual and auditory modalities helps, but it remains unclear how this fusion between visual and auditory information should work.…”
Section: Related Workmentioning
confidence: 99%
“…However, note that they only used annotated vocalized segments; silence and other sounds were thus discarded. Laskowski and Schultz [104] segmented laughter based on three distinct states, namely laughter, speech and non-vocalizations. Their system is not only based on the acoustic characteristics of laughter, but also makes use of the vocal activity of multiple participants by constraining the number of simultaneous speakers and the number of simultaneous laughter.…”
Section: Laughter Segmentationmentioning
confidence: 99%
See 1 more Smart Citation