ACM SIGCHI General Interest Announcements (Mailing List)


Options: Use Forum View

Use Monospaced Font
Show Text Part by Default
Show All Mail Headers

Message: [<< First] [< Prev] [Next >] [Last >>]
Topic: [<< First] [< Prev] [Next >] [Last >>]
Author: [<< First] [< Prev] [Next >] [Last >>]

Print Reply
Leimin Tian <[log in to unmask]>
Reply To:
Leimin Tian <[log in to unmask]>
Fri, 29 Nov 2019 11:38:27 +1100
text/plain (109 lines)
ICMI 2020: Call for Long and Short Papers
25-29 Oct 2020, Utrecht, The Netherlands

Call for Long and Short Papers

The 22nd International Conference on Multimodal Interaction (ICMI 2020)
will be held in Utrecht, the Netherlands.  ICMI is the premier
international forum for multidisciplinary research on multimodal
human-human and human-computer interaction, interfaces, and system
development. The conference focuses on theoretical and empirical
foundations, component technologies, and combined multimodal processing
techniques that define the field of multimodal interaction analysis,
interface design, and system development.

We are keen to showcase novel input and output modalities and interactions
to the ICMI community. ICMI 2020 will feature a single-track main
conference which includes: keynote speakers, technical full and short
papers (including oral and poster presentations), demonstrations, exhibits
and doctoral spotlight papers. The conference will also feature workshops
and grand challenges. The proceedings of ICMI 2020 will be published by ACM
as part of their series of International Conference Proceedings and Digital

We also want to welcome conference papers from behavioral and social
sciences. These papers allow us to understand how technology can be used to
increase our scientific knowledge and may focus less on presenting
technical or algorithmic novelty. For this reason, the "novelty" criteria
used during ICMI 2020 review will be based on two sub-criteria (i.e.,
scientific novelty and technical novelty as described below). Accepted
papers at ICMI 2020 only need to be novel on one of these sub-criteria. In
other words, a paper which is strong on scientific knowledge contribution
but low on algorithmic novelty should be ranked similarly to a paper that
is high on algorithmic novelty but low on knowledge discovery.

- Scientific Novelty: Papers should bring some new knowledge to the
scientific community. For example, discovering new behavioral markers that
are predictive of mental health or how new behavioral patterns relate to
children’s interactions during learning. It is the responsibility of the
authors to perform a proper literature review and clearly discuss the
novelty in the scientific discoveries made in their paper.
- Technical Novelty: Papers reviewed with this sub-criterion should include
novelty in their computational approach for recognizing, generating or
modeling data. Examples include: novelty in the learning and prediction
algorithms, in the neural architecture, or in the data representation.
Novelty can also be associated to a new usage of an existing approach.

Please see the Submission Guidelines for Authors for detailed submission

This year’s conference theme: In this information age, technological
innovation is at the core of our lives and rapidly transforming and
impacting the state of the world in art, culture, and society, and science
as well - the borders between classical disciplines such as humanities and
computer science are fading. In particular, we wonder how multimodal
processing of human behavioural data can create meaningful impact in art,
culture, and society practices. And vice versa, how does art, culture, and
society influence our approaches and techniques in multimodal processing?
As such, this year, ICMI welcomes contributions on our theme for Multimodal
processing and representation of Human Behaviour in Art, Culture, and

Additional topics of interest include but are not limited to:

- Affective computing and interaction
- Cognitive modeling and multimodal interaction
- Gesture, touch and haptics
- Healthcare, assistive technologies
- Human communication dynamics
- Human-robot/agent multimodal interaction
- Interaction with smart environment
- Machine learning for multimodal interaction
- Mobile multimodal systems
- Multimodal behavior generation
- Multimodal datasets and validation
- Multimodal dialogue modeling
- Multimodal fusion and representation
- Multimodal interactive applications
- Speech behaviors in social interaction
- System components and multimodal platforms
- Visual behaviours in social interaction
- Virtual/augmented reality and multimodal interaction

Important Dates

Paper Submission: May 4, 2020 (11:59pm GMT-7)
Reviews to authors: July 3, 2020
Rebuttal due: July 10, 2020 (11:59pm GMT-7)
Paper notification: July 20, 2020
Camera ready paper: August 17, 2020
Presenting at main conference: October 25-29, 2020


    For news of CHI books, courses & software, join CHI-RESOURCES
     mailto: [log in to unmask]

    To unsubscribe from CHI-ANNOUNCEMENTS send an email to
     mailto:[log in to unmask]

    For further details of CHI lists see