VisHCI 2006

                  HCSNet Workshop on the Use of Vision in HCI

           Australian National University, Canberra, 1-3 November 2006

        In cooperation with RSISE and NICTA Canberra Research Laboratory


The Human Communication Sciences Network (HCSNet) is supporting a 
workshop on the use of computer vision in human-computer interaction 
(HCI). Visual communication - such as hand and body gestures, facial 
expressions, auditory-visual speech, sign language etc. - is a major 
communication channel for humans. The availability of low-cost camera 
technology has led to an increased use of visual evidence in HCI.

This workshop aims to bring together researchers, practitioners and 
students from a number of disciplines related to using vision and visual 
evidence in HCI. Visual HCI research is multi-disciplinary and both 
computer vision and HCI research have a strong tradition in Australia. 
Relevant disciplines include computer science, engineering, IT, 
psychology and spoken language research to name a few. Our goal is to 
provide an informal, Australian-based, yet international forum for the 
presentation and discussion of current trends and recent ideas and 
results from leading national and international scientists to foster 
scholarly exchange across disciplines and future collaborations in 
computer vision and the human communication sciences.

The workshop will take place over two and a half days and provide ample 
time to discuss recent ideas and results as well as time for networking. 
The program will include keynote addresses by Iain Matthews 
(Carnegie-Mellon University, Pittsburgh), Gerasimos Potamianos (IBM TJ 
Watson Research Center, Yorktown), and David Powers (Flinders 
University, Adelaide).

The workshop will be divided into several sessions that will run in a 
single-track fashion. Original contributions are invited in all related
areas, including but not limited to:

- Hand and body gestures
- Human motion and pose recognition
- Visual object tracking (e.g. face, body, hands)
- Facial expression analysis and recognition
- Face recognition
- Affective computing
- Non-rigid object structure recovery (e.g. structure from motion for
   hand and face shapes, active shape models, active appearance models)
- Auditory-visual speech processing
- Vision processing in human factors analysis
- Event detection and recognition
- Visual interface design
- Multimodal interfaces and integration
- Perceptual user interfaces
- Sign language analysis and recognition
- Behaviour analysis
- User and context modelling
- Applications (e.g. person authentication, user monitoring,
   driver assistance technology, AV automatic speech recognition)
- Data corpora of visual HCI events

Please check the VisHCI website for paper submission details and 
deadlines. The paper submission is open to both Australian-based and 
international participants.

Workshop participation is open internationally to anyone interested in 
the areas covered by the workshop. Thanks to the generous support from 
HCSNet, registration for VisHCI 2006 will be free. A number of travel 
awards will be made available to help covering the cost of travelling to 
Canberra. Details will be made available on the workshop website shortly.

One of the aims of VisHCI 2006 is in particular to support student 
participation in the workshop which we see as an important opportunity 
for students to present their work and to network with both national and 
international researchers. Such networking forms an important part of 
their studies and is a core goal of the ARC research networks such as 
HCSNet. A number of specifically allocated student travel awards will be
available for Australian-based, interstate students.

Roland Goecke (NICTA / ANU, Canberra)
Antonio Robles-Kelly (NICTA / ANU, Canberra)

If you any questions, please do not hesitate to contact us at
[log in to unmask]

VisHCI 2006
HCSNet Workshop on the Use of Vision in HCI

                To unsubscribe, send an empty email to
     mailto:[log in to unmask]
    For further details of CHI lists see