DBATES: DataBase of Audio features, Text, and visual Expressions in competitive debate Speeches

03/26/2021
by   Taylan K. Sen, et al.
8

In this work, we present a database of multimodal communication features extracted from debate speeches in the 2019 North American Universities Debate Championships (NAUDC). Feature sets were extracted from the visual (facial expression, gaze, and head pose), audio (PRAAT), and textual (word sentiment and linguistic category) modalities of raw video recordings of competitive collegiate debaters (N=717 6-minute recordings from 140 unique debaters). Each speech has an associated competition debate score (range: 67-96) from expert judges as well as competitor demographic and per-round reflection surveys. We observe the fully multimodal model performs best in comparison to models trained on various compositions of modalities. We also find that the weights of some features (such as the expression of joy and the use of the word we) change in direction between the aforementioned models. We use these results to highlight the value of a multimodal dataset for studying competitive, collegiate debate.

READ FULL TEXT

page 1

page 7

page 11

research
11/30/2020

Detecting expressions with multimodal transformers

Developing machine learning algorithms to understand person-to-person en...
research
06/21/2023

Visual-Aware Text-to-Speech

Dynamically synthesizing talking speech that actively responds to a list...
research
08/05/2022

Hybrid Multimodal Feature Extraction, Mining and Fusion for Sentiment Analysis

In this paper, we present our solutions for the Multimodal Sentiment Ana...
research
07/03/2018

Getting the subtext without the text: Scalable multimodal sentiment classification from visual and acoustic modalities

In the last decade, video blogs (vlogs) have become an extremely popular...
research
09/18/2017

Depression Scale Recognition from Audio, Visual and Text Analysis

Depression is a major mental health disorder that is rapidly affecting l...
research
03/03/2023

Meme Sentiment Analysis Enhanced with Multimodal Spatial Encoding and Facial Embedding

Internet memes are characterised by the interspersing of text amongst vi...
research
09/28/2022

Multimodal Prediction of Spontaneous Humour: A Novel Dataset and First Results

Humour is a substantial element of human affect and cognition. Its autom...

Please sign up or login with your details

Forgot password? Click here to reset