2020
DOI: 10.1007/978-3-030-52237-7_19
|View full text |Cite
|
Sign up to set email alerts
|

Improving Affect Detection in Game-Based Learning with Multimodal Data Fusion

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
27
0
3

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 27 publications
(30 citation statements)
references
References 14 publications
0
27
0
3
Order By: Relevance
“…While depictions of early and late fusion styles have been relatively consistent across multiple papers [9,17,37,41,75,84,119], there are still cases where other terms have been used. In [31], one network architecture described as multi-view-one-network is essentially early fusion and one-view-one-network could be considered late fusion.…”
Section: Applying the Taxonomymentioning
confidence: 97%
“…While depictions of early and late fusion styles have been relatively consistent across multiple papers [9,17,37,41,75,84,119], there are still cases where other terms have been used. In [31], one network architecture described as multi-view-one-network is essentially early fusion and one-view-one-network could be considered late fusion.…”
Section: Applying the Taxonomymentioning
confidence: 97%
“…It is also interesting to note the fusion of multimedia data (audio and video) together with data from students' hands while they performed certain tasks (Worsley, 2014). The only researchers who contributed with several papers to this survey were Henderson et al, whose studies focused on student posture and movement along with other data that varied from one study to the next (N. Henderson et al, 2020; N. L. Henderson, Rowe, Mott, & Lester, 2019). It was also interesting to see the fusion between physical and digital data gathered via webcam and the learning platform in Ma et al (2015).…”
Section: Multimodal Datamentioning
confidence: 99%
“…Five of the studies which appeared in the early fusion category stand out for going beyond simple concatenation of features with rather more detailed procedures. Four of those studies were configured to select the best features of each data source (Chango, Cerezo, & Romero, 2021; Chango, Cerezo, Sanchez‐Santillan, et al, 2021; N. L. Henderson, Rowe, Mott, Brawner, et al, 2019; N. Henderson et al, 2020). In contrast, N. L. Henderson, Rowe, Mott, and Lester (2019), reduced the dimensionality of the features using principal component analysis (PCA) in two different configurations: (a) they concatenated all of the features of the sources and applied PCA to the resulting vector; (b) they applied PCA to the features of each source first and concatenated the results following the reduction of dimensionality.…”
Section: Data Fusion Techniques In Multimodal La/edmmentioning
confidence: 99%
“…For the DEAP and DREAMER datasets, class imbalance occurred when AroVal scores were divided into Low and High classes by setting the threshold at the center of the scale. The synthetic minority oversampling technique (SMOTE) [47] addresses this problem by generating new data points from minority classes, and has been used in applications with sparse [48] and noisy [49] samples and multimodal physiological signals for affect classification [50] [51] [52]. In some cases [50] [51], models trained with SMOTE-generated data did not perform well due to the socalled "cold start" problem.…”
Section: Explainable Affect Recognitionmentioning
confidence: 99%