Presentazione di PowerPoint - PowerPoint PPT Presentation

1 / 40
About This Presentation
Title:

Presentazione di PowerPoint

Description:

Audio Features implementing and analysing a set of low level features to ... Emotion recognition analysing the acoustic features in order to recognize ... – PowerPoint PPT presentation

Number of Views:52
Avg rating:3.0/5.0
Slides: 41
Provided by: Gia107
Category:

less

Transcript and Presenter's Notes

Title: Presentazione di PowerPoint


1
Conveying Affectiveness in Leading-edge Living
Adaptive Systems
2
  • The CALLAS Project
  • http//www.callas-newmedia.eu
  • Fiore Basile
  • Metaware
  • f.basile_at_metaware.it
  • Diego Arnone
  • Engineering Ingegneria Informatica
  • diego.arnone_at_eng.it

3
Partners
Engineering (ITA) VTT Electronics (FI) BBC
(UK) MetaWare (ITA) Studio Azzurro (ITA) XIM
(UK) Digital Video (ITA) Humanware (ITA) Nexture
(ITA) University of Augsburg (DE) ICCS / NTUA
(GR)
University of Mons (BE) University of Teeside
(UK) Helsinki Univ. Technology (FI) Paris 8
(FR) Scuola Normale Sup. (IT) University of
Reading (UK) Fondazione Teatro Massimo (IT) HIT
Laboratory (NZ)
4
Overview
  • Emotions and affectiveness a fundamental element
    for a rich human-machine interaction and
    communication
  • Affective Interfaces are central to new media
    experience, and especially for entertainment
  • Digital Theatre
  • Interactive TV
  • Augmented Reality Art
  • Interactive public performances

5
Objectives
  • Develop specific re-usable technologies for the
    multimodal processing of the emotional experience
    associated to Arts and Entertainment
  • Handle new and innovative categories of emotions,
    as well as improve the performance for existing
    modalities at the input level
  • Promote technology transfer for these results in
    particular towards SMEs (Small and Medium
    Enterprises) in the new media sector.

6
Working Areas
  • CALLAS Shelf
  • CALLAS Framework
  • CALLAS Showcases

7
The CALLAS Shelf
8
The CALLAS Shelf
  • Its a set of components, each one processing one
    or more modalities, taking into account affective
    aspects of the interaction
  • Many input modalities are foreseen speech,
    gesture, sound, facial expressions, sensors, etc.
  • All components aim to be reusable and capable of
    real-time processing
  • The Shelf components will be aggregated according
    to commonly used fusion patterns, and composed
    through a visual authoring environment

9
The CALLAS Shelf
  • Input Shelf Components

10
FPMS I-Component
University of Mons
The word uttered and a corresponding emotion
(active - passive) between a small list of
Ekmanian and not Ekmanian emotions
ESR Emotional Speech Recognition
Speech
Main Concrete Elements of Work () Emotion
Robustness proposing and implementing
approaches for robust emotional speech
recognition Emotion recognition implementing
speech-related features for supporting extraction
of emotional information
11
VTT I-Components(1/3)
VTT Electronics
speech music constant noise environmental
sound silence
SCA Sound Capture and Analysis
Audio streaming
Main Concrete Elements of Work Sound Capture
designed for different situations and events in
order to provide high quality audio capture Sound
Analysis in order to provide low and high level
information from the audio input (e.g. mapping 6
the states in emotional states )
12
VTT I-Components(2/3)
VTT Electronics
VFE Video Features Extraction
Fast/slow Lot of/little amount Upper/lower
body movement
Video streaming on wide spaces
Main Concrete Elements of Work Video Features
implementing and analysing a set of low level
video features Audio Features implementing and
analysing a set of low level features to support
in extracting contextual and emotional
information
13
VTT I-Components(3/3)
VTT Electronics
S
GBMT Gesture and Body Motion Tracking
Sensors ()
Hand movement
Video streaming
Main Concrete Elements of Work Gesture
recognition data acquisition with different
sensors and positions Emotion recognition -
Outlining relevant emotion related gestures and
body motions
14
UOA I-Components(1/2)
University of Augsburg
ESR-speech Emotional Speech Recognition (based
on acoustic features)
Emotional class recognition
Audio files
Main Concrete Elements of Work Emotion
recognition analysing the acoustic features in
order to recognize emotional aspects
15
UOA I-Components(2/2)
University of Augsburg
Emotional class recognition 2-9 emotional
classes ()
ESR-linguistic Emotional Speech Recognition
(based on linguistic features)
Text files
Main Concrete Elements of Work Implementation of
a fusion model for linguistic and acoustic
features
16
ICCS I-Components(1/4)
ICCS/NTUA
What the user is looking at Up, down, right,
left OR Degree of direction (by means of sensors)
S
GzR Gaze Recognition
High resolution images of frontal faces and
signals coming from many sensors
17
ICCS I-Components(2/4)
ICCS/NTUA
  • Whissels quadrant or Expressions
  • Neutral
  • Anger
  • Disgust
  • Fear
  • Sadness
  • Joy
  • Surprise

R-bFER Rule-based Facial Expression Recognition
Static images, Videos?
FFD Facial Features Detection
Coordinates of interest points of the face
18
ICCS I-Components(3/4)
ICCS/NTUA
GR Gesture Recognition
One of six output states Representing recognized
gestures
Static images, Videos
Coordinates of hands and head
HHDT Hands and Head Detection and Tracking
19
ICCS I-Components(4/4)
ICCS/NTUA
GEA Gesture Expressivity Analysis
Expressivity features for performed gestures
Static images, Videos
Coordinates of hands and head
HHDT Hands and Head Detection and Tracking
20
ICCS I-Components
  • Main Concrete Elements of Work
  • Try to overcome issues introduced by imperfect
    recording conditions and personalized
    expressivity.
  • Enhance existing feature extraction with the
    introduction of measures of confidence on the
    feature values and the final emotion estimation.
  • Evaluate features with different emotional models
    according to specific application requirements on
    expected results.

21
Output Shelf Components
22
ICCS O-Component
ICCS/NTUA
An expressive model of users behaviour (by ECA)
ES Expressivity Synthesis
S
Image sequences, sensors, history and personality
details
23
UOA O-Component
University of Augsburg
NLG Natural Language Generation
A text containing the desired utterance
Attributes and values pairs describing the
emotions
24
PAR8 O-Component
Eye / head / gaze directions
Virtual env.
EA-ECA Emotional Attentive ECA
APML expressive gestures
  • Main concrete elements of work
  • develop an ECA that is sensitive to and
    expressive
  • through aspects relating to emotion and
    attention

25
HIT O-Component
ARToolKit is a software library for building
Augmented Reality (AR) applications
Main Concrete Elements of Work
  • Develop an interface for visual programming of AR
    applications
  • Add support for speech and gesture
  • Extend AR toolkit to include natural feature
    tracking

26
The CALLAS Framework
27
The CALLAS Framework
  • Flexible application framework, based on the
    following approach
  • Theory-neutral in terms of Modalities integration
  • Supporting the semantic processing of modalities
  • Supporting the development of applications
    featuring
  • Blackboards
  • Custom pre-defined multimodal fusion models
  • Fusion of affective modalities

28
The CALLAS Framework
  • The CALLAS Framework aims to
  • Ease the development of a specific kind of
    applications targeted to entertainment and arts
  • Ease the aggregation of shelf components into
    easy-to-reuse building blocks
  • Provide an intuitive metaphor suitable for
    non-technical users (mainly artists) willing to
    adapt and repurpose the CALLAS Showcases
    applications or their high-level components

29
The CALLAS Showcases
30
The CALLAS Showcases
  • The CALLAS Showcases are an on-going laboratory
    for experimenting with the fusion of affective
    modalities and their impact on digital arts
  • The target audience includes digital arts,
    entertainment and digital theatre
  • They also serve as testbeds for the CALLAS Shelf
    components and the CALLAS Framework

31
The CALLAS Showcases
  • Augmented Reality for Art, Entertainment, and
    Digital Theatre
  • support the development of Augmented Reality Art
    installations in which user interaction is
    mediated by the detection of user emotions
  • demonstrate how real-time detection of the mood
    and the affective state of the people involved in
    a live performance (directors, actors, audience)
    can generate a new genre of Digitally-enhanced
    performances.
  • Interactive Installations for Public Spaces
  • Next-Generation Interactive Television    

32
The CALLAS Showcases
  • Augmented Reality for Art, Entertainment, and
    Digital Theatre
  • Interactive Installations for Public Spaces
  • Explore how emotional states of members of a
    group can be conveyed to other members through
    mixed reality configurations and traces
  • Explore, in intensive group experiences, the
    implications of adding awareness of emotional
    states of remote or collocated members in
    addition to other contextual features.
  • Develop mixed reality applications combining
    sensors and user-controlled mechanisms
  • Next-Generation Interactive Television    

33
The CALLAS Showcases
  • Augmented Reality for Art, Entertainment, and
    Digital Theatre
  • Interactive Installations for Public Spaces
  • Next-Generation Interactive Television  
  • Develop the concept of affective Interactive TV
  • Based on the generation of affective content by
    ECA
  • ECA gets inputs from the broadcasted content and
    the users perceived viewing experience  

34
Some examples..
35
Emotional Tree (e-Tree)
Dynamic growth, a function of perceived Affective
relation
ARToolkit Table top installation
Multimodal Interaction motion, non-verbal
behaviour, interaction history, spoken comments
Artistic concept by Maurice Benayoun
36
Interactive TV
37
Technical Approach
User
Virtual spectator
Interactive Story
Emotional Speech
Keyword spotting
Affective categories
Paralinguistic speech
Non-verbal (body attitude)
Interactive Storytelling Engine
Multimodal Affective Analysis
38
(No Transcript)
39
Collaboration ideas
  • CALLAS results will be very focused on Digital
    Arts and Entertainment
  • The project may become a viable channel for
    experimenting other project results in these
    sectors
  • CALLAS is open for collaboration both on the
    technical and evaluation sides

40
  • THANK YOU
  • Fiore Basile,
  • Metaware, Pisa, Italy
  • Email f.basile_at_metaware.it
  • Diego Arnone,
  • Engineering I. I., Roma, Italy
  • Email diego.arnone_at_eng.it
Write a Comment
User Comments (0)
About PowerShow.com