ACM SIGCHI General Interest Announcements (Mailing List)


Options: Use Forum View

Use Monospaced Font
Show HTML Part by Default
Show All Mail Headers

Message: [<< First] [< Prev] [Next >] [Last >>]
Topic: [<< First] [< Prev] [Next >] [Last >>]
Author: [<< First] [< Prev] [Next >] [Last >>]

Print Reply
William Hudson <[log in to unmask]>
Reply To:
Stephen Brewster <[log in to unmask]>
Tue, 2 Dec 2003 16:17:59 -0000
text/plain (91 lines)
Posted on behalf of Stephen Brewster [mailto:[log in to unmask]]

Special Issue of Personal and Ubiquitous Computing 


Editors: Stephen Brewster, University of Glasgow, UK and Matt Jones, 
University of Waikato, New Zealand 

Mobile devices have been one of the major success stories of computing 
in recent years. Large numbers of people now carry a sophisticated 
computing device with them all the time in the form of a mobile 
telephone or a personal digital assistant. More sophisticated and 
powerful wearable computers are becoming available and in the future 
will be embedded into the clothes and accessories a user wears; these 
wearable computers have more power and flexibility for use in different 
situations and with access to different media and services. 

One of the key problems for interaction with mobile and wearable devices 
is their impoverished user interface when compared desktop computers. 
Screens are often small or non existent and head-mounted displays are 
not always an option, so output is limited. Small keyboards or touch 
screens can be hard to use for input when on the move, making 
interaction hard. To make mobile and wearable devices more effective and 
acceptable to users, and to allow a whole new range of services to be 
delivered to people when on the move, the challenges of input and output 
must be addressed. Multimodal interaction is one way to enhance 

The combination of vision, hearing and speech, and touch and gesture 
have great possibilities for increasing the bandwidth of communication 
between user and device (taste and smell are potential future 
contributors too). Flexible multimodal interfaces may allow users to 
interact in a more natural manner. This is particularly important for 
mobile and wearable devices as they are used by a wide range of 
different people in a wide range of different situations. They also have 
great possibilities for users with disabilities as they provide 
alternatives to the standard GUI model which can be problematic. 

Key issues remain about how the different senses should be used, what 
each sense is good for, how they should be combined and how to assess 
their performance in a mobile setting. For this special issue we are 
keen to gather papers on the state of the art in multimodal interface 
design for mobile and wearable devices to try and answer some of these 
questions and to look towards the future of multimodal interface design.

We are soliciting papers that discuss novel multimodal techniques, 
methods, models and tools to overcome the impoverished interfaces of the 
current generation of mobile devices. We would like papers that bring 
together the following sorts of issues in the mobile and wearable 

.       Auditory interfaces using speech and non-speech sounds 
.       Role and efficacy of speech recognition 
.       Gestural, graspable, tactile, haptic and tangible interfaces 
.       Vision based interaction (e.g. navigation through glancing) 
.       Physiological input/output 
.       Effective combination of multiple modalities (both in theory and 
.       Evaluation of multimodal systems 
.       Multimodal interfaces for disabled users 
.       Novel sensors and output devices to facilitate multimodal 
.       Novel mobile services using multimodal interfaces 

This list is not exclusive; we are keen to receive papers on any novel 
combination of modalities. 

Submission format and schedule 
Submissions should be e-mailed as a PDF file (including images) directly 
to the editors of the special issue. Information regarding journal 
submissions and formats is available at:

The deadline for receiving submissions is 26th January, 2004. All 
contributions will be peer reviewed to the journal's usual standard. We 
encourage potential authors to contact the editors well before the final 

For further information or to discuss a possible contribution, please 
contact the special issue editors, Steve Brewster 
([log in to unmask])  and Matt Jones ([log in to unmask]) 
Paper Submission: 26th January, 2004 
Notification of acceptance: 9th April, 2004 
Final Corrections to papers: 9th July, 2004 
Publication of special issue: September 2004