ELSNET-List Message
Elsnet
 
   


ELSNET-List Message

Subject: [ E-Conf ] 2nd Audio/Visual Emotion Challenge and Workshop (AVEC 2012)
From: <schuller_(on)_tum.de>
Date received: 24 Jan 2012
Deadline: -
Start date: 22 Oct 2012





Dear List,

We would like to bring to your attention the2nd International
Audio/Visual Emotion Challenge and Workshop (AVEC2012), organised
in conjunction with ACM ICMI2012 this year. See below for the CFP
- apologies for potential cross-posting.

_____________________________________________________________

Call for Participation / Papers

2nd International Audio/Visual Emotion Challenge and Workshop
(AVEC 2012)

in conjunction with ACM ICMI 2012, October 22-26, Santa Monica,
California, USA

http://sspnet.eu/avec2012/ http://www.acm.org/icmi/2012/

Register and download data and features:
http://avec-db.sspnet.eu/accounts/register/

_____________________________________________________________

Scope

The Audio/Visual Emotion Challenge and Workshop (AVEC 2012) will
be the second competition event aimed at comparison of multimedia
processing and machine learning methods for automatic audio,
visual and audiovisual emotion analysis, with all participants
competing under strictly the same conditions. The goal of the
Challenge is to provide a common benchmark test set for
individual multimodal information processing and to bring
together the audio and video emotion recognition communities, to
compare the relative merits of the two approaches to emotion
recognition under well-defined and strictly comparable conditions
and establish to what extent fusion of the approaches is possible
and beneficial. A second motivation is the need to advance
emotion recognition systems to be able to deal with naturalistic
behavior in large volumes of un-segmented, non-prototypical and
non-preselected data as this is exactly the type of data that
both multimedia retrieval and human-machine/human-robot
communication interfaces have to face in the real world.

We are calling for teams to participate in emotion recognition
from acoustic audio analysis, linguistic audio analysis, video
analysis, or any combination of these. As benchmarking database
the SEMAINE database of naturalistic video and audio of
human-agent interactions, along with labels for four affect
dimensions will be used. Emotion will have to be recognized in
terms of continuous time, continuous valued dimensional affect in
the dimensions arousal, expectation, power and valence. Two
Sub-Challenges are addressed: The Word-Level Sub-Challenge
requires participants to predict the level of affect at
word-level and only when the user is speaking. The Fully
Continuous Sub-Challenge involves fully continuous affect
recognition, where the level of affect has to be predicted for
every moment of the recording.

Besides participation in the Challenge we are calling for papers
addressing the overall topics of this workshop, in particular
works that address the differences between audio and video
processing of emotive data, and the issues concerning combined
audio-visual emotion recognition

Topics include, but are not limited to:

Audio/Visual Emotion Recognition:
. Audio-based Emotion Recognition
. Linguistics-based Emotion Recognition
. Video-based Emotion Recognition
. Social Signals in Emotion Recognition
. Multi-task learning of Multiple Dimensions
. Novel Fusion Techniques as by Prediction
. Cross-corpus Feature Relevance
. Agglomeration of Learning Data
. Semi- and Unsupervised Learning
. Synthesized Training Material
. Context in Audio/Visual Emotion Recognition
. Multiple Rater Ambiguity

Application:
. Multimedia Coding and Retrieval
. Usability of Audio/Visual Emotion Recognition
. Real-time Issues

Important Dates
___________________________________________

Paper submission July 21, 2012

Notification of acceptance August 6, 2012

Camera ready paper and final challenge result submission August
15, 2012

Workshop October 22 or 26, 2012

Organisers
___________________________________________

Björn Schuller (Tech. Univ. Munich, Germany) Michel Valstar
University of Nottingham, UK) Roddy Cowie (Queen's University
Belfast, UK) Maja Pantic (Imperial College London, UK)

Program Committee
___________________________________________

Elisabeth André, Universität Augsburg, Germany Anton Batliner,
FAU Erlangen-Nuremberg, Germany Felix Burkhardt, Deutsche
Telekom, Germany Fang Chen, NICTA, Australia Mohamed Chetouani,
UPMC, France Laurence Devillers, U. Paris-Sorbonne 4, France
Julien Epps, University of New South Wales, Australia Anna
Esposito, IIASS, Italy Raul Fernandez, IBM, USA Roland Göcke,
Australian National Univ., Australia Hatice Gunes, Queen Mary
University London, UK Julia Hirschberg, Columbia University, USA
Aleix Martinez, Ohio State University, USA Marc Méhu, University
of Geneva, Switzerland Marcello Mortillaro, University of Geneva,
Switzerland Matti Pietikainen, University of Oulu, Finland Peter
Robinson, University of Cambridge, UK Stefan Steidl, FAU
Erlangen-Nuremberg, Germany Jianhua Tao, Chinese Academy of
Sciences, China Fernando de la Torre, Carnegie Mellon University,
USA Mohan Trivedi, Univ. of California San Diego, USA Matthew
Turk, Univ. of California Santa Barbara, USA Alessandro
Vinciarelli, University of Glasgow, UK Rama Chellappa, University
of Maryland, USA Stefanos Zafeiriou, Imperial College London, UK

Please regularly visit our website http://sspnet.eu/avec2012 for
more information and excuse cross-postings,

Thank you very much and all the best,

Björn Schuller, Roddy Cowie, Maja Pantic, and Michel Valstar

___________________________________________

Dr. Björn Schuller Senior Lecturer

Technische Universität München Institute for Human-Machine
Communication D-80333 München Germany +49-(0)89-289-28548

schuller_(at)_tum.de www.mmk.ei.tum.de/~sch
___________________________________________

__________________________________________
- ELSNET mailing list Elsnet-list_(at)_elsnet.org
- To manage your subscription go to:
  http://mailman.elsnet.org/mailman/listinfo/elsnet-list


	

[ Search | Events calendar | Deadline calendar ]

 

Page generated 20-12-2014 by Steven Krauwer Disclaimer / Contact ELSNET