MM-INTEREST Archives

ACM SIGMM Interest List

MM-INTEREST@LISTSERV.ACM.ORG

Options: Use Forum View

Use Monospaced Font
Show Text Part by Default
Show All Mail Headers

Message: [<< First] [< Prev] [Next >] [Last >>]
Topic: [<< First] [< Prev] [Next >] [Last >>]
Author: [<< First] [< Prev] [Next >] [Last >>]

Print Reply
Subject:
From:
Bjoern Schuller <[log in to unmask]>
Reply To:
Bjoern Schuller <[log in to unmask]>
Date:
Tue, 24 Jan 2012 21:32:58 +0000
Content-Type:
text/plain
Parts/Attachments:
text/plain (138 lines)
Dear List,

We would like to bring to your attention the2nd International Audio/Visual Emotion Challenge and Workshop (AVEC2012), organised in conjunction with ACM ICMI2012 this year. See below for the CFP - apologies for potential cross-posting.

_____________________________________________________________

Call for Participation / Papers

2nd International Audio/Visual Emotion Challenge and Workshop (AVEC 2012)

in conjunction with ACM ICMI 2012, October 22-26, Santa Monica, California, USA

http://sspnet.eu/avec2012/
http://www.acm.org/icmi/2012/ 

Register and download data and features:
http://avec-db.sspnet.eu/accounts/register/ 

_____________________________________________________________

Scope

The Audio/Visual Emotion Challenge and Workshop (AVEC 2012) will be the second competition event aimed at comparison of multimedia processing and machine learning methods for automatic audio, visual and audiovisual emotion analysis, with all participants competing under strictly the same conditions. The goal of the Challenge is to provide a common benchmark test set for individual multimodal information processing and to bring together the audio and video emotion recognition communities, to compare the relative merits of the two approaches to emotion recognition under well-defined and strictly comparable conditions and establish to what extent fusion of the approaches is possible and beneficial. A second motivation is the need to advance emotion recognition systems to be able to deal with naturalistic behavior in large volumes of un-segmented, non-prototypical and non-preselected data as this is exactly the type of data that both multimedia retrieval and human-machine/human-robot communication interfaces have to face in the real world.

We are calling for teams to participate in emotion recognition from acoustic audio analysis, linguistic audio analysis, video analysis, or any combination of these. As benchmarking database the SEMAINE database of naturalistic video and audio of human-agent interactions, along with labels for four affect dimensions will be used. Emotion will have to be recognized in terms of continuous time, continuous valued dimensional affect in the dimensions arousal, expectation, power and valence. Two Sub-Challenges are addressed: The Word-Level Sub-Challenge requires participants to predict the level of affect at word-level and only when the user is speaking. The Fully Continuous Sub-Challenge involves fully continuous affect recognition, where the level of affect has to be predicted for every moment of the recording.

Besides participation in the Challenge we are calling for papers addressing the overall topics of this workshop, in particular works that address the differences between audio and video processing of emotive data, and the issues concerning combined audio-visual emotion recognition

Topics include, but are not limited to:

Audio/Visual Emotion Recognition:
. Audio-based Emotion Recognition
. Linguistics-based Emotion Recognition
. Video-based Emotion Recognition
. Social Signals in Emotion Recognition
. Multi-task learning of Multiple Dimensions
. Novel Fusion Techniques as by Prediction
. Cross-corpus Feature Relevance
. Agglomeration of Learning Data
. Semi- and Unsupervised Learning
. Synthesized Training Material
. Context in Audio/Visual Emotion Recognition
. Multiple Rater Ambiguity

Application:
. Multimedia Coding and Retrieval
. Usability of Audio/Visual Emotion Recognition
. Real-time Issues


Important Dates
___________________________________________

Paper submission
July 21, 2012

Notification of acceptance
August 6, 2012

Camera ready paper and final challenge result submission
August 15, 2012

Workshop
October 22 or 26, 2012


Organisers
___________________________________________

Björn Schuller (Tech. Univ. Munich, Germany)
Michel Valstar University of Nottingham, UK)
Roddy Cowie (Queen's University Belfast, UK)
Maja Pantic (Imperial College London, UK)


Program Committee
___________________________________________

Elisabeth André, Universität Augsburg, Germany
Anton Batliner, FAU Erlangen-Nuremberg, Germany
Felix Burkhardt, Deutsche Telekom, Germany
Fang Chen, NICTA, Australia
Mohamed Chetouani, UPMC, France
Laurence Devillers, U. Paris-Sorbonne 4, France
Julien Epps, University of New South Wales, Australia
Anna Esposito, IIASS, Italy
Raul Fernandez, IBM, USA
Roland Göcke, Australian National Univ., Australia
Hatice Gunes, Queen Mary University London, UK
Julia Hirschberg, Columbia University, USA
Aleix Martinez, Ohio State University, USA
Marc Méhu, University of Geneva, Switzerland
Marcello Mortillaro, University of Geneva, Switzerland
Matti Pietikainen, University of Oulu, Finland
Peter Robinson, University of Cambridge, UK
Stefan  Steidl, FAU Erlangen-Nuremberg, Germany
Jianhua Tao, Chinese Academy of Sciences, China
Fernando de la Torre, Carnegie Mellon University, USA
Mohan Trivedi, Univ. of California San Diego, USA
Matthew Turk, Univ. of California Santa Barbara, USA
Alessandro Vinciarelli, University of Glasgow, UK
Rama Chellappa, University of Maryland, USA
Stefanos Zafeiriou, Imperial College London, UK


Please regularly visit our website http://sspnet.eu/avec2012 for more information and excuse cross-postings,


Thank you very much and all the best,



Björn  Schuller, Roddy Cowie, Maja Pantic, and Michel Valstar



___________________________________________

Dr. Björn Schuller
Senior Lecturer

Technische Universität München
Institute for Human-Machine Communication
D-80333 München
Germany
+49-(0)89-289-28548

[log in to unmask]
www.mmk.ei.tum.de/~sch
___________________________________________

############################

To unsubscribe from the MM-INTEREST list:
write to: mailto:[log in to unmask]
or click the following link:
http://listserv.acm.org/SCRIPTS/WA-ACMLPX.EXE?SUBED1=MM-INTEREST&A=1

ATOM RSS1 RSS2