2001 IEEE International Conference on Acoustics, Speech, and Signal Processing. Proceedings (Cat. No.01CH37221)
DOI: 10.1109/icassp.2001.940754
|View full text |Cite
|
Sign up to set email alerts
|

MiPad: a multimodal interaction prototype

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
24
0
1

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 30 publications
(25 citation statements)
references
References 5 publications
0
24
0
1
Order By: Relevance
“…For example, whenever hands are occupied (e.g., while driving), or where hand-based interfaces are bulky (using personal digital assistances (PDAs) or cell phones), ASR will undeniably succeed. Indeed, ASR is increasingly used on hand-held devices [14] -some PDA-based ASR systems are starting to appear commercially such as the IBM personal speech assistant [4] and the Microsoft MiPad [8] (others are listed in [14]). A number of wireless communication companies also launched their products integrated with ASR systems, like Motorala's voiceXML and Nokia 9000 series.…”
Section: Introductionmentioning
confidence: 99%
“…For example, whenever hands are occupied (e.g., while driving), or where hand-based interfaces are bulky (using personal digital assistances (PDAs) or cell phones), ASR will undeniably succeed. Indeed, ASR is increasingly used on hand-held devices [14] -some PDA-based ASR systems are starting to appear commercially such as the IBM personal speech assistant [4] and the Microsoft MiPad [8] (others are listed in [14]). A number of wireless communication companies also launched their products integrated with ASR systems, like Motorala's voiceXML and Nokia 9000 series.…”
Section: Introductionmentioning
confidence: 99%
“…Before deep learning methods were adopted, there had already been numerous efforts in multimodal and multitask learning. For example, a prototype called MiPad for multimodal interactions involving capturing, leaning, coordinating, and rendering a mix of speech, touch, and visual information was developed and reported in [113,164]. In [165,166], mixed sources of information from multiple-sensory microphones with separate bone-conductive and air-born paths were exploited to de-noise speech.…”
Section: B) a Selected Review On Deep Learning For Multimodal Processingmentioning
confidence: 99%
“…A wide range of noise-robust techniques developed over past 30 years can be analyzed and categorized using five different criteria: (1) feature-domain versus model-domain processing, (2) the use of prior knowledge about the acoustic environment distortion, (3) the use of explicit environment-distortion models, (4) deterministic versus uncertainty processing, and (5) the use of acoustic models trained jointly with the same feature enhancement or model adaptation process used in the testing stage. See a comprehensive review in [109,110] and additional review literature or original work in [111][112][113][114].…”
Section: D) a New Level Of Noise Robustnessmentioning
confidence: 99%
“…Among several use-cases such as mobile interaction, interactive maps, graphic design applications and systems for augmented dual-task environments, mobile systems are selected for detailed discussion, underpinning of pattern candidates and pattern identification. Examples for multimodal mobile interaction are personal assistants for e-mail and web access such as MiPad [22], Personal Speech Assistant [13], tourist guides and city information systems such as SmartKom mobile [26], MATCH [21,24], MUST [3] or COMPASS [4].…”
Section: Identification Of Multimodal User Interface Patterns Based Omentioning
confidence: 99%
“…Mobile Systems such as Microsoft's MiPad [22] and IBM's Personal Speech Assistant [13] are good examples.…”
Section: Known Usesmentioning
confidence: 99%