Proceedings of the 5th International Conference on Multimodal Interfaces 2003
DOI: 10.1145/958432.958467
|View full text |Cite
|
Sign up to set email alerts
|

Augmenting user interfaces with adaptive speech commands

Abstract: We present a system that augments any unmodified Java application with an adaptive speech interface. The augmented system learns to associate spoken words and utterances with interface actions such as button clicks. Speech learning is constantly active and searches for correlations between what the user says and does. Training the interface is seamlessly integrated with using the interface. As the user performs normal actions, she may optionally verbally describe what she is doing. By using a phoneme recognize… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2004
2004
2016
2016

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 13 publications
(4 citation statements)
references
References 8 publications
0
4
0
Order By: Relevance
“…These high level commands are manually created and PIXELTONE may be better served by learning from examples [10] or mining online tutorials.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…These high level commands are manually created and PIXELTONE may be better served by learning from examples [10] or mining online tutorials.…”
Section: Discussionmentioning
confidence: 99%
“…Since these templates are based on grammatical structures of sentences, it is possible to semi-automate the process of defining templates by analyzing grammatical patterns on large corpora of users' utterances. Alternatively, users can train the system directly by verbalizing what they are doing as they are doing [10].…”
Section: Scalabilitymentioning
confidence: 99%
“…We are also continuing work in applying our results from grounded language systems to multimodal interface design (Gorniak & Roy, 2003). We recently demonstrated an application of the Bishop system as described in this paper to the problem of referent resolution in the graphical user interface for a 3D modelling application, Blender (Blender Foundation , 2003).…”
Section: Lexical Entriesmentioning
confidence: 90%
“…Speech-based user interfaces are considered a solid foundation when handling natural interfaces regarding people with some sort of disability: It permits the input of information without the resort to a keyboard or even in the event of the inexistence of a monitor; it facilitates tasks where hands and/or eyes of the users are busy; and it relieves the need for writing for people with motor or intellectual disabilities [34]. Some systems have been developed that take advantage of speech interfaces in order to replace quick commands given by the mouse or keyboard in graphical interfaces, with very positive results [12,20].…”
Section: Related Workmentioning
confidence: 99%