“…In the first one, methods extract a number of features from audio (related to prosody and turn-taking) [47], [26], [44], video (related to head and body activity or gaze) [37], [24], and wearable sensors (related to body motion or physical proximity) [35], [38]. In the second step, these features are used as input to supervised or unsupervised learning methods to infer traits like dominance [48], [28], extroversion and locus of control [44], [37]; relations like roles [58], [16], [17] or status [47], [26]; group attitudes like cooperation and competition [30], tasks like brainstorming [29]; and concepts like collective intelligence [57]. Other works use the extracted features to create interactive systems that, through various visualizations of behavioral cues, affect the interaction itself [13], [35], [52], [6].…”