Face and Voice Recognition-Based Emotion Analysis System (EAS) to Minimize Heterogeneity in the Metaverse

The metaverse, where users interact through avatars, is evolving to closely mirror the real world, requiring realistic object responses based on users’ emotions. While technologies like eye-tracking and hand-tracking transfer physical movements into virtual spaces, accurate emotion detection remains...

Full description

Saved in:
Bibliographic Details
Main Authors: Surak Son, Yina Jeong
Format: Article
Language:English
Published: MDPI AG 2025-01-01
Series:Applied Sciences
Subjects:
Online Access:https://www.mdpi.com/2076-3417/15/2/845
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1832589163051876352
author Surak Son
Yina Jeong
author_facet Surak Son
Yina Jeong
author_sort Surak Son
collection DOAJ
description The metaverse, where users interact through avatars, is evolving to closely mirror the real world, requiring realistic object responses based on users’ emotions. While technologies like eye-tracking and hand-tracking transfer physical movements into virtual spaces, accurate emotion detection remains challenging. This study proposes the “Face and Voice Recognition-based Emotion Analysis System (EAS)” to bridge this gap, assessing emotions through both voice and facial expressions. EAS utilizes a microphone and camera to gauge emotional states, combining these inputs for a comprehensive analysis. It comprises three neural networks: the Facial Emotion Analysis Model (FEAM), which classifies emotions using facial landmarks; the Voice Sentiment Analysis Model (VSAM), which detects vocal emotions even in noisy environments using MCycleGAN; and the Metaverse Emotion Recognition Model (MERM), which integrates FEAM and VSAM outputs to infer overall emotional states. EAS’s three primary modules—Facial Emotion Recognition, Voice Emotion Recognition, and User Emotion Analysis—analyze facial features and vocal tones to detect emotions, providing a holistic emotional assessment for realistic interactions in the metaverse. The system’s performance is validated through dataset testing, and future directions are suggested based on simulation outcomes.
format Article
id doaj-art-099e7111baee467da011fb306d6a0fae
institution Kabale University
issn 2076-3417
language English
publishDate 2025-01-01
publisher MDPI AG
record_format Article
series Applied Sciences
spelling doaj-art-099e7111baee467da011fb306d6a0fae2025-01-24T13:21:02ZengMDPI AGApplied Sciences2076-34172025-01-0115284510.3390/app15020845Face and Voice Recognition-Based Emotion Analysis System (EAS) to Minimize Heterogeneity in the MetaverseSurak Son0Yina Jeong1Department of Software, College of Engineering, Catholic Kwandong University, Gangneung 25601, Republic of KoreaDepartment of Software, College of Engineering, Catholic Kwandong University, Gangneung 25601, Republic of KoreaThe metaverse, where users interact through avatars, is evolving to closely mirror the real world, requiring realistic object responses based on users’ emotions. While technologies like eye-tracking and hand-tracking transfer physical movements into virtual spaces, accurate emotion detection remains challenging. This study proposes the “Face and Voice Recognition-based Emotion Analysis System (EAS)” to bridge this gap, assessing emotions through both voice and facial expressions. EAS utilizes a microphone and camera to gauge emotional states, combining these inputs for a comprehensive analysis. It comprises three neural networks: the Facial Emotion Analysis Model (FEAM), which classifies emotions using facial landmarks; the Voice Sentiment Analysis Model (VSAM), which detects vocal emotions even in noisy environments using MCycleGAN; and the Metaverse Emotion Recognition Model (MERM), which integrates FEAM and VSAM outputs to infer overall emotional states. EAS’s three primary modules—Facial Emotion Recognition, Voice Emotion Recognition, and User Emotion Analysis—analyze facial features and vocal tones to detect emotions, providing a holistic emotional assessment for realistic interactions in the metaverse. The system’s performance is validated through dataset testing, and future directions are suggested based on simulation outcomes.https://www.mdpi.com/2076-3417/15/2/845artificial intelligenceemotion recognitionfacial expression analysisvoice analysisfacial-voice matching
spellingShingle Surak Son
Yina Jeong
Face and Voice Recognition-Based Emotion Analysis System (EAS) to Minimize Heterogeneity in the Metaverse
Applied Sciences
artificial intelligence
emotion recognition
facial expression analysis
voice analysis
facial-voice matching
title Face and Voice Recognition-Based Emotion Analysis System (EAS) to Minimize Heterogeneity in the Metaverse
title_full Face and Voice Recognition-Based Emotion Analysis System (EAS) to Minimize Heterogeneity in the Metaverse
title_fullStr Face and Voice Recognition-Based Emotion Analysis System (EAS) to Minimize Heterogeneity in the Metaverse
title_full_unstemmed Face and Voice Recognition-Based Emotion Analysis System (EAS) to Minimize Heterogeneity in the Metaverse
title_short Face and Voice Recognition-Based Emotion Analysis System (EAS) to Minimize Heterogeneity in the Metaverse
title_sort face and voice recognition based emotion analysis system eas to minimize heterogeneity in the metaverse
topic artificial intelligence
emotion recognition
facial expression analysis
voice analysis
facial-voice matching
url https://www.mdpi.com/2076-3417/15/2/845
work_keys_str_mv AT surakson faceandvoicerecognitionbasedemotionanalysissystemeastominimizeheterogeneityinthemetaverse
AT yinajeong faceandvoicerecognitionbasedemotionanalysissystemeastominimizeheterogeneityinthemetaverse