Title
Towards SMIL as a Foundation for Multimodal, Multimedia Applications
Abstract
Rich and interactive multimedia applications, where audio, video, graphics and text are precisely synchronized under timing constraints are becoming ubiquitous. Multimodal applications further extend the concept of user interaction combining different modalities, like speech recognition, speech synthesis and gestures. However, authoring dialog-capable multimodal, multimedia services is a very difficult task. Fortunately, the W3C has sponsored the development of SMIL, an elegant notation for multimedia applications, which has been embraced by both Microsoft and RealNetworks. In this paper, we argue that SMIL is an ideal substrate for extending multimedia applications with multimodal facilities. SMIL as it stands is not a general notation for controlling media and input mode resources. We show that all what is needed are few natural extensions to SMIL along with the addition of a simple reactive programming language that we call ReX. Our language is designed to be maximally compatible with existing W3C recommendations through a generic event system based on DOM and an expression language based on XPATH. It is also designed to be simple so that the fundamental notion of seeking time (e.g. going backwards and forwards in presentations) is preserved.
Year
Venue
Field
2001
INTERSPEECH
Graphics,Notation,Speech synthesis,Gesture,Computer science,XPath,Reactive programming,Interactive media,Multimedia,Synchronized Multimedia Integration Language
DocType
Citations 
PageRank 
Conference
6
1.02
References 
Authors
1
4
Name
Order
Citations
PageRank
Jennifer L. Beckham161.02
Giuseppe DiFabbrizio2557.78
Nils Klarlund364565.43
Giuseppe Di Fabbrizio433044.45