Empirically based research in multimodality starts with the collection of video data relevant for a specific research question. This data must then be enriched by human coders, who annotate the corpus and who also explore the potential interconnections between modalities (e.g., face, gesture, speech, posture, etc.). The data must then be validated for consistency using quantitative measures -- systematic analysis of the observed multimodal behaviors is the central characteristic of multimodal corpora. For analysis and modeling tasks, exploratory qualitative and quantitative analysis tools are needed for browsing, viewing, extraction, and modeling. At the same time, the intercultural dimension must be taken into account, since multimodal interaction is always embedded in a social setting.
This state-of-the-art survey documents the scientific outcome of the International Workshop on "Multimodal Corpora: From Models of Natural Interaction to Systems and Applications", held in conjunction with the 6th International Conference for Language Resources and Evaluation (LREC), on May 27, 2008, in Marrakech, Morocco. It presents current research from fields as diverse as psychology, artificial intelligence, robotics, signal processing, computational linguistics, and human-computer interaction.
The 6 selected workshop contributions are complemented by 7 invited research articles, including contributions from major international multimodal corpus projects like AMI and SmartWeb. All papers underwent a special review process for this volume, resulting in significant revisions and extensions based on the experts' advice.