2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) 2015
DOI: 10.1109/iros.2015.7354164
|View full text |Cite
|
Sign up to set email alerts
|

Audio-visual beat tracking based on a state-space model for a music robot dancing with humans

Abstract: This paper presents an audio-visual beat-tracking method for an entertainment robot that can dance in synchronization with music and human dancers. Conventional music robots have focused on either music audio signals or dancing movements of humans for detecting and predicting beat times in real time. Since a robot needs to record music audio signals by using its own microphones, however, the signals are severely contaminated with loud environmental noise and reverberant sounds. Moreover, it is difficult to vis… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
13
0

Year Published

2016
2016
2021
2021

Publication Types

Select...
3
2
1

Relationship

1
5

Authors

Journals

citations
Cited by 13 publications
(13 citation statements)
references
References 19 publications
0
13
0
Order By: Relevance
“…Instead, the method keeps all the possibilities of tempos and beat times. If a unique tempo were extracted from music audio signals as in [15], tempo estimation failure would severely degrade the overall performance. We therefore formulate a nonlinear state-space model that has a tempo and a beat time as latent variables and acoustic and skeleton features as observed variables.…”
Section: Proposed Methodsmentioning
confidence: 99%
See 4 more Smart Citations
“…Instead, the method keeps all the possibilities of tempos and beat times. If a unique tempo were extracted from music audio signals as in [15], tempo estimation failure would severely degrade the overall performance. We therefore formulate a nonlinear state-space model that has a tempo and a beat time as latent variables and acoustic and skeleton features as observed variables.…”
Section: Proposed Methodsmentioning
confidence: 99%
“…We compared the proposed audio-visual beat-tracking method with two conventional audio beat-tracking methods [5,15]. The method [5] is implemented in HARK [29] robot audition software, and its parameters are set to the default values except for m = 90.…”
Section: Experimental Conditionsmentioning
confidence: 99%
See 3 more Smart Citations