Speech-driven Animation with Meaningful Behaviors

08/04/2017
by   Najmeh Sadoughi, et al.
0

Conversational agents (CAs) play an important role in human computer interaction. Creating believable movements for CAs is challenging, since the movements have to be meaningful and natural, reflecting the coupling between gestures and speech. Studies in the past have mainly relied on rule-based or data-driven approaches. Rule-based methods focus on creating meaningful behaviors conveying the underlying message, but the gestures cannot be easily synchronized with speech. Data-driven approaches, especially speech-driven models, can capture the relationship between speech and gestures. However, they create behaviors disregarding the meaning of the message. This study proposes to bridge the gap between these two approaches overcoming their limitations. The approach builds a dynamic Bayesian network (DBN), where a discrete variable is added to constrain the behaviors on the underlying constraint. The study implements and evaluates the approach with two constraints: discourse functions and prototypical behaviors. By constraining on the discourse functions (e.g., questions), the model learns the characteristic behaviors associated with a given discourse class learning the rules from the data. By constraining on prototypical behaviors (e.g., head nods), the approach can be embedded in a rule-based system as a behavior realizer creating trajectories that are timely synchronized with speech. The study proposes a DBN structure and a training approach that (1) models the cause-effect relationship between the constraint and the gestures, (2) initializes the state configuration models increasing the range of the generated behaviors, and (3) captures the differences in the behaviors across constraints by enforcing sparse transitions between shared and exclusive states per constraint. Objective and subjective evaluations demonstrate the benefits of the proposed approach over an unconstrained model.

READ FULL TEXT

page 4

page 5

page 6

page 7

page 11

research
01/11/2021

A Review of Evaluation Practices of Gesture Generation in Embodied Conversational Agents

Embodied Conversational Agents (ECA) take on different forms, including ...
research
03/04/2021

Toward Automated Generation of Affective Gestures from Text:A Theory-Driven Approach

Communication in both human-human and human-robot interac-tion (HRI) con...
research
08/29/2021

A Hybrid Rule-Based and Data-Driven Approach to Driver Modeling through Particle Filtering

Autonomous vehicles need to model the behavior of surrounding human driv...
research
07/01/2021

Passing a Non-verbal Turing Test: Evaluating Gesture Animations Generated from Speech

In real life, people communicate using both speech and non-verbal signal...
research
05/18/2023

AMII: Adaptive Multimodal Inter-personal and Intra-personal Model for Adapted Behavior Synthesis

Socially Interactive Agents (SIAs) are physical or virtual embodied agen...
research
08/12/2021

To Rate or Not To Rate: Investigating Evaluation Methods for Generated Co-Speech Gestures

While automatic performance metrics are crucial for machine learning of ...
research
08/08/2023

TranSTYLer: Multimodal Behavioral Style Transfer for Facial and Body Gestures Generation

This paper addresses the challenge of transferring the behavior expressi...

Please sign up or login with your details

Forgot password? Click here to reset