Enhancing Emotion Recognition in Speech Based on Self-Supervised Learning: Cross-Attention Fusion of Acoustic and Semantic Features
Speech Emotion Recognition has gained considerable attention in speech processing and machine learning due to its potential applications in human-computer interaction, mental health monitoring, and customer service. However, state-of-the-art models for speech emotion recognition use many parameters,...
Saved in:
| Main Authors: | Bashar M. Deeb, Andrey V. Savchenko, Ilya Makarov |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
IEEE
2025-01-01
|
| Series: | IEEE Access |
| Subjects: | |
| Online Access: | https://ieeexplore.ieee.org/document/10938083/ |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
AFT-SAM: Adaptive Fusion Transformer with a Sparse Attention Mechanism for Audio–Visual Speech Recognition
by: Na Che, et al.
Published: (2024-12-01) -
Knowledge enhancement for speech emotion recognition via multi-level acoustic feature
by: Huan Zhao, et al.
Published: (2024-12-01) -
Dual-feature speech emotion recognition fusion algorithm based on wavelet scattering transform and MFCC
by: YING Na, et al.
Published: (2024-05-01) -
End-to-end feature fusion for jointly optimized speech enhancement and automatic speech recognition
by: Mohamed Medani, et al.
Published: (2025-07-01) -
Hierarchical Multi-Task Learning Based on Interactive Multi-Head Attention Feature Fusion for Speech Depression Recognition
by: Yujuan Xing, et al.
Published: (2025-01-01)