ELSNET-list archive

Category:   E-Conf
Subject:   2nd Audio/Visual Emotion Challenge and Workshop (AVEC 2012)
Email:   schuller_(on)_tum.de
Date received:   24 Jan 2012
Start date:   22 Oct 2012

Dear List, We would like to bring to your attention the2nd International Audio/Visual Emotion Challenge and Workshop (AVEC2012), organised in conjunction with ACM ICMI2012 this year. See below for the CFP - apologies for potential cross-posting. _____________________________________________________________ Call for Participation / Papers 2nd International Audio/Visual Emotion Challenge and Workshop (AVEC 2012) in conjunction with ACM ICMI 2012, October 22-26, Santa Monica, California, USA http://sspnet.eu/avec2012/ http://www.acm.org/icmi/2012/ Register and download data and features: http://avec-db.sspnet.eu/accounts/register/ _____________________________________________________________ Scope The Audio/Visual Emotion Challenge and Workshop (AVEC 2012) will be the second competition event aimed at comparison of multimedia processing and machine learning methods for automatic audio, visual and audiovisual emotion analysis, with all participants competing under strictly the same conditions. The goal of the Challenge is to provide a common benchmark test set for individual multimodal information processing and to bring together the audio and video emotion recognition communities, to compare the relative merits of the two approaches to emotion recognition under well-defined and strictly comparable conditions and establish to what extent fusion of the approaches is possible and beneficial. A second motivation is the need to advance emotion recognition systems to be able to deal with naturalistic behavior in large volumes of un-segmented, non-prototypical and non-preselected data as this is exactly the type of data that both multimedia retrieval and human-machine/human-robot communication interfaces have to face in the real world. We are calling for teams to participate in emotion recognition from acoustic audio analysis, linguistic audio analysis, video analysis, or any combination of these. As benchmarking database the SEMAINE database of naturalistic video and audio of human-agent interactions, along with labels for four affect dimensions will be used. Emotion will have to be recognized in terms of continuous time, continuous valued dimensional affect in the dimensions arousal, expectation, power and valence. Two Sub-Challenges are addressed: The Word-Level Sub-Challenge requires participants to predict the level of affect at word-level and only when the user is speaking. The Fully Continuous Sub-Challenge involves fully continuous affect recognition, where the level of affect has to be predicted for every moment of the recording. Besides participation in the Challenge we are calling for papers addressing the overall topics of this workshop, in particular works that address the differences between audio and video processing of emotive data, and the issues concerning combined audio-visual emotion recognition Topics include, but are not limited to: Audio/Visual Emotion Recognition: . Audio-based Emotion Recognition . Linguistics-based Emotion Recognition . Video-based Emotion Recognition . Social Signals in Emotion Recognition . Multi-task learning of Multiple Dimensions . Novel Fusion Techniques as by Prediction . Cross-corpus Feature Relevance . Agglomeration of Learning Data . Semi- and Unsupervised Learning . Synthesized Training Material . Context in Audio/Visual Emotion Recognition . Multiple Rater Ambiguity Application: . Multimedia Coding and Retrieval . Usability of Audio/Visual Emotion Recognition . Real-time Issues Important Dates ___________________________________________ Paper submission July 21, 2012 Notification of acceptance August 6, 2012 Camera ready paper and final challenge result submission August 15, 2012 Workshop October 22 or 26, 2012 Organisers ___________________________________________ Björn Schuller (Tech. Univ. Munich, Germany) Michel Valstar University of Nottingham, UK) Roddy Cowie (Queen's University Belfast, UK) Maja Pantic (Imperial College London, UK) Program Committee ___________________________________________ Elisabeth André, Universität Augsburg, Germany Anton Batliner, FAU Erlangen-Nuremberg, Germany Felix Burkhardt, Deutsche Telekom, Germany Fang Chen, NICTA, Australia Mohamed Chetouani, UPMC, France Laurence Devillers, U. Paris-Sorbonne 4, France Julien Epps, University of New South Wales, Australia Anna Esposito, IIASS, Italy Raul Fernandez, IBM, USA Roland Göcke, Australian National Univ., Australia Hatice Gunes, Queen Mary University London, UK Julia Hirschberg, Columbia University, USA Aleix Martinez, Ohio State University, USA Marc Méhu, University of Geneva, Switzerland Marcello Mortillaro, University of Geneva, Switzerland Matti Pietikainen, University of Oulu, Finland Peter Robinson, University of Cambridge, UK Stefan Steidl, FAU Erlangen-Nuremberg, Germany Jianhua Tao, Chinese Academy of Sciences, China Fernando de la Torre, Carnegie Mellon University, USA Mohan Trivedi, Univ. of California San Diego, USA Matthew Turk, Univ. of California Santa Barbara, USA Alessandro Vinciarelli, University of Glasgow, UK Rama Chellappa, University of Maryland, USA Stefanos Zafeiriou, Imperial College London, UK Please regularly visit our website http://sspnet.eu/avec2012 for more information and excuse cross-postings, Thank you very much and all the best, Björn Schuller, Roddy Cowie, Maja Pantic, and Michel Valstar ___________________________________________ Dr. Björn Schuller Senior Lecturer Technische Universität München Institute for Human-Machine Communication D-80333 München Germany +49-(0)89-289-28548 schuller_(at)_tum.de www.mmk.ei.tum.de/~sch ___________________________________________ __________________________________________ - ELSNET mailing list Elsnet-list_(at)_elsnet.org - To manage your subscription go to: http://mailman.elsnet.org/mailman/listinfo/elsnet-list

[print/pda] [no frame] [navigation table] [navigation frame]     Page generated 27-07-2012 by Steven Krauwer Disclaimer / Contact ELSNET