Posted on behalf of Noureddine Elouazizi, Yulia Bachvarova and Anton Nijholt
(Apologies for multiple postings)
The submission deadline for the International Workshop on Linguistic engineering meets cognitive engineering in the interface design of multimodal systems (LECEMS) has been extended.
The deadline is now **** July 22nd **** (no more extensions will be possible).
Please see details below.
CALL FOR PAPERS
Linguistic Engineering Meets Cognitive Engineering in the Interface Design of Multimodal Systems (LECEMS)
Workshop at the 7th International Conference on Multimodal Interfaces October 3, 2005, Trento, Italy http://icmi05.itc.it http://hmi.ewi.utwente.nl/conference/LECEMS
Multimodal systems demand innovative approaches in terms of information selection, extraction and presentation (fission), software architectures, interactive techniques and devices.
As such, various kinds of rules, guidelines and strategies have been proposed for deciding how to present information in multimodal interfaces, making the best use of multimodal options.
Yet, many cognitive aspects as related to the user's interaction with information are overlooked. This is due to the absence of formal cognitive models which can model and map such interactions.
Many of those who create multimodal systems, are often unaware of the formal cognitive techniques that can be applied to deepen our understanding of how the representation of information works, how it can support reasoning, how it activates the user's channels of perception in an optimal way, etc. Therefore, it follows that the application of the formal theories of cognition to the principles of design, implementation and evaluation of multimodal interface technology is needed more than ever. For, there is a growing need to formally understand the cognitive processes, which underlie human-machine multimodal communication.
This will help in predicting the usability preferences of the user, the optimal vs. non-optimal aspects of multimodal information processing, etc.
Moreover, the attempts to use multiple media to represent information in an automatically generated environment have far led to new problems.
For example, there is a lack of a unified meta-model which formalizes the cross-media references. The engineering of concepts from natural language processing (eg. cohesion, anaphora, referring expressions etc.) can allow a useful transfer of concepts to implement in the design of the output of multimodal systems, however; the discourse of an artificial multimodal system requires an explicit representation of the syntax and semantics of the visual, auditory and textual discourse to conform to the optimal processing requirements of human-cognitive channels of perception. In this respect, among the questions that the workshop will focus on: (i) Are the existing models of linguistic reference transferable to the context of multimodal reference? (e.g. in automatically generated multimedia systems, etc. ), (ii) Are the processes of cognitive referring and reference resolution in natural languages akin to those in artificial multimodal systems?
The questions related to how to determine, cognitive relevance in the processes of information representation in multimodal systems by stepping on referring models leaves much to desire.
In sum, the task of bringing "intelligence" into the technology of multimodal interfaces is channeled through a relevant, useful and productive combination of "cognitive engineering"
and "linguistic engineering".
Paper submission deadline: July 15, 2005
Acceptance notification: July 29, 2005
Camera ready papers due: September 15, 2005
Early registration due: August 31, 2005
Workshop day: October 3, 2005
Falling within the general guidelines described in the main theme of the workshop above, we invite scientific contributions (papers) on the following topics and the related ones:
· Formal models of multimodal referring
· Computational modeling for the generation of referring expressions
in a multimodal discourse environment
· Handling and presenting of arbitrary and ambiguous information in different media
· Cognitive processes of visual and auditory identification, reasoning
and resolution in multimodal environments
· The integration of natural language and speech processing technology
with other modalities (e.g. graphics, etc.)
· Knowledge representation of referential relations in multimodal systems
· Visualization and interpretation of referential concepts in multimodal systems,
which combine speech and gesture, etc.
· NLP Features- engineering for multimedia automatic generation
· Embodied conversational agents and multimodal systems
· Intelligent presentation of information retrieved from heterogeneous multimedia databases
· Reasoning in/on multimodal interactive systems
· Multimodal architectures for heterogeneous reasoning in multimodal interactive systems
· Representation of knowledge about modalities for multimodal/multimedia systems
Noureddine Elouazizi (University of Leiden) n.elouazizi at let.leidenuniv.nl
Yulia Bachvarova (University of Twente)
y.s.bachvarova at ewi.utwente.nl
Anton Nijholt (University of Twente)
A.Nijholt at ewi.utwente.nl
1. Elisabeth André (Universität Augsburg, Germany)
2. Antal van den Bosch (Tilburg University, The Netherlands)
3. Harry Bunt (Tilburg University, The Netherlands)
4. Thierry Declerck, (DFKI GmbH, Germany)
5. Eduard Hovy (University of Southern California, USA)
6. Franciska de Jong (Twente University, the Netherlands)
7. Karon MacLean (University of British Columbia, Canada)
8. Mark Maybury (MITRE, USA)
9. Sharon Oviatt (Oregon Health & Sciences University , USA)
10. Jon Oberlander (University of Edinburgh, Scotland)
11. Francis Jeffry Pelletier ( Simon Fraser University, Canada)
12. Fred Popowhich (Simon Fraser University, Canada)
13. Laurent Romary (LORIA, France)
14. Mark Steedman (University of Edinburgh, Scotland)
15. Keith Stenning (University of Edinburgh, Scotland)
16. Hari Sundaram (Arizona State University, USA
Professor Sharon Oviatt from the Oregon Health & Sciences University (USA) will give an invited talk.
We invite research papers (maximum 8 pages).
Submitted papers must be original, containing new
and original results. All submissions will be reviewed
and selected based on originality and contribution to the workshop topics.
Submissions are welcome from the academic and industrial institutes.
Please send your submissions as a single PDF or PS file
to the following address [log in to unmask]
In the body of your paper submission email, please include
the following identification information:
* Author(s) name(s), affiliation(s), and e-mail address(es)
* Abstract: short summary (up to 5 lines maxi.)
For the format, we strongly recommend to use ACM SIG Proceedings Templates.
More detailed information about the format of submissions can be found here:
All papers accepted for the workshop will be published in the
Workshop Proceedings provided by ITC. In addition, we plan that a selected number of the
accepted workshop papers will be expanded and revised for inclusion
into a possible post-workshop special journal issue.
To unsubscribe, send an empty email to
mailto:[log in to unmask]
For further details of CHI lists see http://sigchi.org/listserv