Courses Organising Committee Programme Committee Preliminary Schedule

Motivation

Information Fusion is an everyday necessity in complex speech and language systems although it is rarely noticed as such. Systems in this area are always composed of individual components which need to co-operate towards a common goal. The reason for such modularisation is obvious: due to the many different layers and knowledge sources usually involved, system components need to be developed independently by different people with different areas and degrees of specialization. Moreover, there might be the option or even the necessity to train such components on vastly different kinds of data sets. Often it is the case that a range of solution alternatives exists for one and the same processing task, each of them providing a partial and unreliable but perhaps complementary contribution to the overall behaviour of the system. Here, the problem arises of how to achieve a synergy between such competitive approaches, even for tasks where the desired processing result is no longer trivial one.

Since all available solutions for speech and language processing are approximations to a conveyed ideal, system design has to account for the inherent uncertainty of processing results on all levels. this makes system integration a problem of information fusion, which can be considered as solved in some cases but it is still an open research issue in others: for speech recognition the contributions of the acoustic models and the language model need to be adjusted properly, whereas in translation a target sentence has to be composed of partial structures produced by e.g. an example-based component and a deep-linguistic one. Other examples can be taken from more ambitious task like the integration of acoustic (speech and noise) and visual data (lip movements, hand gestures and facial expressions) in complex multimodal environments.

It is always astonishing to notice, how little effort humansspend to integrate the available information from such a diversity of sources. Even more, multimodal information processing in many cases leads to a facilitatory effect in the sense that using evidence from a range of sensory channels is faster than relying on a single one. this, obviously, is contrary to the behaviour of all the techniques we currently have at our disposal when designing a complex speech and language system. Therefore, the summer school will depart from a survey of phenomena and mechanisms for information fusion. It continues with studying various approaches for sensor-data fusion in technical systems, like robots. Finally it will investigate the issue of information fusion from the perspective of a range of speech and language processing tasks, namely
  • speech recognition and spoken language systems
  • machine translation
  • distributed and multilingual information systems
  • information retrieval in heterogenous sources
  • parsing
  • multimodal speech and language systems
How different these application areas might look, the underlying principles of and the approaches towards information fusion seem to be comparable it not even highly similar. It is the goal of the summer school to highlight such similarities and to inspire the cross-disciplinary transfer of ideas and solutions

-- CristinaVertan -- 05 Mar 2006
 
This site is powered by FoswikiCopyright © by the contributing authors. All material on this collaboration platform is the property of the contributing authors.
Ideas, requests, problems regarding Foswiki? Send feedback