Multimodal models for contextual affect assessment in real-time

Jordan Vice, Masood Mehmood Khan, Svetlana Yanushkevich

Research output: Chapter in Book/Conference paperConference paperpeer-review

5 Citations (Scopus)

Abstract

Most affect classification schemes rely on near accurate single-cue models resulting in less than required accuracy under certain peculiar conditions. We investigate how the holism of a multimodal solution could be exploited for affect classification. This paper presents the design and implementation of a prototype, stand-alone, real-time multimodal affective state classification system. The presented system utilizes speech and facial muscle movements to create a holistic classifier. The system combines a facial expression classifier and a speech classifier that analyses speech through paralanguage and propositional content. The proposed classification scheme includes a Support Vector Machine (SVM) - paralanguage; a K-Nearest Neighbor (KNN) - propositional content and an InceptionV3 neural network - facial expressions of affective states. The SVM and Inception models boasted respective validation accuracies of 99.2% and 92.78%.

Original languageEnglish
Title of host publicationProceedings - 2019 IEEE 1st International Conference on Cognitive Machine Intelligence, CogMI 2019
PublisherIEEE, Institute of Electrical and Electronics Engineers
Pages87-92
Number of pages6
ISBN (Electronic)9781728167374
ISBN (Print)978-1-7281-6738-1
DOIs
Publication statusPublished - Dec 2019
Externally publishedYes
Event1st IEEE International Conference on Cognitive Machine Intelligence, CogMI 2019 - Los Angeles, United States
Duration: 12 Dec 201914 Dec 2019

Publication series

NameProceedings - 2019 IEEE 1st International Conference on Cognitive Machine Intelligence, CogMI 2019

Conference

Conference1st IEEE International Conference on Cognitive Machine Intelligence, CogMI 2019
Country/TerritoryUnited States
CityLos Angeles
Period12/12/1914/12/19

Fingerprint

Dive into the research topics of 'Multimodal models for contextual affect assessment in real-time'. Together they form a unique fingerprint.

Cite this