[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

2nd Audio/Visual Emotion Challenge and Workshop (AVEC 2012)

Dear List,

We would like to bring to your attention the2nd International Audio/Visual Emotion Challenge and Workshop (AVEC2012), organised in conjunction with ACM ICMI2012 this year. See below for the CFP - apologies for potential cross-posting.


Call for Participation / Papers

2nd International Audio/Visual Emotion Challenge and Workshop (AVEC 2012)

in conjunction with ACM ICMI 2012, October 22-26, Santa Monica, California, USA

http://www.acm.org/icmi/2012/ ;

Register and download data and features:
http://avec-db.sspnet.eu/accounts/register/ ;



The Audio/Visual Emotion Challenge and Workshop (AVEC 2012) will be the second competition event aimed at comparison of multimedia processing and machine learning methods for automatic audio, visual and audiovisual emotion analysis, with all participants competing under strictly the same conditions. The goal of the Challenge is to provide a common benchmark test set for individual multimodal information processing and to bring together the audio and video emotion recognition communities, to compare the relative merits of the two approaches to emotion recognition under well-defined and strictly comparable conditions and establish to what extent fusion of the approaches is possible and beneficial. A second motivation is the need to advance emotion recognition systems to be able to deal with naturalistic behavior in large volumes of un-segmented, non-prototypical and non-preselected data as this is exactly the type of data that both multimedia retrieval and human-machine/human-robot communication interfaces have to face in the real world.

We are calling for teams to participate in emotion recognition from acoustic audio analysis, linguistic audio analysis, video analysis, or any combination of these. As benchmarking database the SEMAINE database of naturalistic video and audio of human-agent interactions, along with labels for four affect dimensions will be used. Emotion will have to be recognized in terms of continuous time, continuous valued dimensional affect in the dimensions arousal, expectation, power and valence. Two Sub-Challenges are addressed: The Word-Level Sub-Challenge requires participants to predict the level of affect at word-level and only when the user is speaking. The Fully Continuous Sub-Challenge involves fully continuous affect recognition, where the level of affect has to be predicted for every moment of the recording.

Besides participation in the Challenge we are calling for papers addressing the overall topics of this workshop, in particular works that address the differences between audio and video processing of emotive data, and the issues concerning combined audio-visual emotion recognition

Topics include, but are not limited to:

Audio/Visual Emotion Recognition:
. Audio-based Emotion Recognition
. Linguistics-based Emotion Recognition
. Video-based Emotion Recognition
. Social Signals in Emotion Recognition
. Multi-task learning of Multiple Dimensions
. Novel Fusion Techniques as by Prediction
. Cross-corpus Feature Relevance
. Agglomeration of Learning Data
. Semi- and Unsupervised Learning
. Synthesized Training Material
. Context in Audio/Visual Emotion Recognition
. Multiple Rater Ambiguity

. Multimedia Coding and Retrieval
. Usability of Audio/Visual Emotion Recognition
. Real-time Issues

Important Dates

Paper submission
July 21, 2012

Notification of acceptance
August 6, 2012

Camera ready paper and final challenge result submission
August 15, 2012

October 22 or 26, 2012


Björn Schuller (Tech. Univ. Munich, Germany)
Michel Valstar University of Nottingham, UK)
Roddy Cowie (Queen's University Belfast, UK)
Maja Pantic (Imperial College London, UK)

Program Committee

Elisabeth André, Universität Augsburg, Germany
Anton Batliner, FAU Erlangen-Nuremberg, Germany
Felix Burkhardt, Deutsche Telekom, Germany
Fang Chen, NICTA, Australia
Mohamed Chetouani, UPMC, France
Laurence Devillers, U. Paris-Sorbonne 4, France
Julien Epps, University of New South Wales, Australia
Anna Esposito, IIASS, Italy
Raul Fernandez, IBM, USA
Roland Göcke, Australian National Univ., Australia
Hatice Gunes, Queen Mary University London, UK
Julia Hirschberg, Columbia University, USA
Aleix Martinez, Ohio State University, USA
Marc Méhu, University of Geneva, Switzerland
Marcello Mortillaro, University of Geneva, Switzerland
Matti Pietikainen, University of Oulu, Finland
Peter Robinson, University of Cambridge, UK
Stefan  Steidl, FAU Erlangen-Nuremberg, Germany
Jianhua Tao, Chinese Academy of Sciences, China
Fernando de la Torre, Carnegie Mellon University, USA
Mohan Trivedi, Univ. of California San Diego, USA
Matthew Turk, Univ. of California Santa Barbara, USA
Alessandro Vinciarelli, University of Glasgow, UK
Rama Chellappa, University of Maryland, USA
Stefanos Zafeiriou, Imperial College London, UK

Please regularly visit our website http://sspnet.eu/avec2012 for more information and excuse cross-postings,

Thank you very much and all the best,

Björn  Schuller, Roddy Cowie, Maja Pantic, and Michel Valstar


Dr. Björn Schuller
Senior Lecturer

Technische Universität München
Institute for Human-Machine Communication
D-80333 München