2005
DOI: 10.1109/tsmca.2004.838480
|View full text |Cite
|
Sign up to set email alerts
|

Dynamic Context Capture and Distributed Video Arrays for Intelligent Spaces

Abstract: Abstract-Intelligent environments can be viewed as systems where humans and machines (rooms) collaborate. Intelligent (or smart) environments need to extract and maintain an awareness of a wide range of events and human activities occurring in these spaces. This requirement is crucial for supporting efficient and effective interactions among humans as well as humans and intelligent spaces. Visual information plays an important role for developing accurate and useful representation of the static and dynamic sta… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
36
0

Year Published

2006
2006
2015
2015

Publication Types

Select...
4
4
2

Relationship

0
10

Authors

Journals

citations
Cited by 76 publications
(36 citation statements)
references
References 36 publications
0
36
0
Order By: Relevance
“…can be applied to captured audio-visual signals to extract higher-level semantic information, such as identification and location in real time. A system that combines face and audio based identification along with motion detection, person tracking and audio based localization has been proposed in the literature [21]. Such a system applies state-of-the-art methods to process results from each individual modality and uses particle filtering to fuse both modalities for providing robust identification and localization.…”
Section: Simultaneous Identification and Localizationmentioning
confidence: 99%
“…can be applied to captured audio-visual signals to extract higher-level semantic information, such as identification and location in real time. A system that combines face and audio based identification along with motion detection, person tracking and audio based localization has been proposed in the literature [21]. Such a system applies state-of-the-art methods to process results from each individual modality and uses particle filtering to fuse both modalities for providing robust identification and localization.…”
Section: Simultaneous Identification and Localizationmentioning
confidence: 99%
“…In our ongoing research on driver-assistance systems [27], [28] and audiovisual scene understanding [29], [30], emotion recognition using multimodality will certainly help us improve the interface. With this motivation, we have put in significant effort on collection of audiovisual affect database in a challenging ambient of car settings [31].…”
Section: B Audio-visual Affect Database (Cvrrcar-avdb)mentioning
confidence: 99%
“…The method of video-only tracking [4], [5] is generally reliable and accurate when the targets are in the camera field of view, but limitations are introduced when the targets are occluded by other speakers, when they disappear from the camera field of view, or the appearance of the targets or illumination is changed [3], [6]. Audio tracking [7], [8], [9] is not restricted by these limitations, however, audio data is intermittent over time and may be corrupted by background noise and room reverberations, which may introduce non-negligible tracking errors.…”
Section: Introductionmentioning
confidence: 99%