ELSNET-List Message

ELSNET-List Message

Subject: [ E-Announce ] CfP INTERSPEECH 2013 Computational Paralinguistics Challenge
From: <schuller_(on)_tum.de>
Date received: 22 Dec 2012
Deadline: 18 Mar 2013
Start date: -

Dear List,

For those of you interested, let us announce the opening of the
first Sub-Challenge of the

INTERSPEECH 2013 Computational Paralinguistics Challenge:

Call for Participation



Social Signals, Conflict, Emotion, Autism


First Sub-Challenge now open - obtain the data:


The Challenge

After four consecutive Challenges at INTERSPEECH, there still
exists a multiplicity of not yet covered, but highly relevant
paralinguistic phenomena. In the last instalments, we focused on
single speakers. With a new task, we now want to broaden to
analysing discussion of multiple speakers in the Conflict
Sub-Challenge. A further novelty is introduced by the Social
Signals Sub-Challenge: For the first time, non-linguistic events
have to be classified and localised - laughter and fillers. In
the Emotion Sub-Challenge we are literally "going back to the
roots". However, by intention, we use acted material for the
first time to fuel the ever on-going discussion on differences
between naturalistic and acted material and hope to highlight the
differences. Finally, the Autism Sub-Challenge picks up on Autism
Spectrum Condition in children's speech in this year. Apart from
intelligent and socially competent future agents and robots, main
applications are found in the medical domain and surveillance.
The Challenge corpora feature rich annotation such as speaker
meta-data, orthographic transcript, phonemic transcript, and
segmentation. All four are given with distinct definitions of
test, development, and training partitions, incorporating speaker
independence as needed in most real-life settings. Benchmark
results of the most popular approaches will be provided as in the
years before. In these respects, the INTERSPEECH 2013
bridging the gap between excellent research on paralinguistic
information in spoken language and low compatibility of results.

In summary, four Sub-Challenges are addressed:

 In the Social Signals Sub-Challenge, non-linguistic events -
laughter and fillers - of a speaker have to be classified and
localised based on acoustics.

 In the Conflict Sub-Challenge, group discussions have to be
automatically evaluated aiming at retrieving conflicts.

 In the Emotion Sub-Challenge, the emotion of a speaker's voice
has to be determined by a suited learning algorithm and acoustic

 In the Autism Sub-Challenge, the type of pathology of a speaker
has to be determined by a suited classification algorithm and
acoustic features.

The measures of competition will be Unweighted Average Area Under
receiver operating Curve and Recall. All Sub-Challenges allow
contributors to find their own features with their own machine
learning algorithm. However, a standard feature set will be
provided per corpus that may be used. Participants will have to
stick to the definition of training, development, and test sets.
They may report on results obtained on the development set, but
have only five trials to upload their results on the test sets,
whose labels are unknown to them. Each participation will be
accompanied by a paper presenting the results that undergoes
peer-review and has to be accepted for the conference in order to
participate in the Challenge. The organisers preserve the right
to re-evaluate the findings, but will not participate themselves
in the Challenge. Participants are encouraged to compete in all

Overall, contributions using the provided or equivalent data are
sought for (but not limited to):

 Participation in a Sub-Challenge

 Contributions focussing on Computational Paralinguistics
centred around the Challenge topics

The results of the Challenge will be presented at Interspeech
2013 in Lyon, France. Prizes will be awarded to the Sub-Challenge
winners. If you are interested and planning to participate in
INTERSPEECH 2013 ComParE, or if you want to be kept informed
about the Challenge, please send the organisers an e-mail to
indicate your interest and visit the homepage:


Björn Schuller (TUM, Germany)

Stefan Steidl (FAU Erlangen-Nuremberg, Germany)

Anton Batliner (TUM, Germany)

Alessandro Vinciarelli (University of Glasgow, UK)

Klaus Scherer (Swiss Center for Affective Sciences, Switzerland)

Fabien Ringeval (University of Fribourg, Switzerland)

Mohamed Chetouani (Université Pierre et Marie Curie, France)


Paper Submission 18 March 2013

Final Result Upload 24 May 2013

Camera-ready Paper 29 May 2013


HUMAINE Association (http://emotion-research.net/)

SSPNet (http://sspnet.eu/)

ASC-Inclusion (http://www.asc-inclusion.eu/)


PD Dr. habil. DI Björn W. Schuller

Head Machine Intelligence & Signal Processing Group

Institute for Human-Machine Communication

Technische Universität München

D-80333 München







- ELSNET mailing list Elsnet-list_(at)_elsnet.org
- To manage your subscription go to:


[ Search | Events calendar | Deadline calendar ]


Page generated 23-04-2014 by Steven Krauwer Disclaimer / Contact ELSNET