Multimodal interactive spaces: MagicTV and magicMap

Marcelo Worsley*, Michael Johnston

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contribution

2 Scopus citations


Through the growing popularity of voice-enabled search, multimodal applications are finally starting to get into the hands of consumers. However, these applications are principally for mobile platforms and generally involve highly-moded interaction where the user has to click or hold a button in order to speak. Significant technical challenges remain in bringing multimodal interaction to other environments such as smart living rooms and classrooms, where users speech and gesture is directed toward large displays or interactive kiosks and the microphone and other sensors are 'always on'. In this demonstration, we present a framework combining low cost hardware and open source software that lowers the barrier of entry for exploration of multimodal interaction in smart environments. Specifically, we will demonstrate the combination of infrared tracking, face detection, and open microphone speech recognition for media search (magicTV) and map navigation (magicMap).

Original languageEnglish (US)
Title of host publication2010 IEEE Workshop on Spoken Language Technology, SLT 2010 - Proceedings
Number of pages2
StatePublished - Dec 1 2010
Event2010 IEEE Workshop on Spoken Language Technology, SLT 2010 - Berkeley, CA, United States
Duration: Dec 12 2010Dec 15 2010

Publication series

Name2010 IEEE Workshop on Spoken Language Technology, SLT 2010 - Proceedings


Other2010 IEEE Workshop on Spoken Language Technology, SLT 2010
Country/TerritoryUnited States
CityBerkeley, CA


  • Gesture recognition
  • Multimodal integration
  • Open microphone
  • Speech recognition

ASJC Scopus subject areas

  • Language and Linguistics


Dive into the research topics of 'Multimodal interactive spaces: MagicTV and magicMap'. Together they form a unique fingerprint.

Cite this