Multi-Detection-Based Speech Emotion Recognition Using Autoencoder in Mobility Service Environment
- Abstract
- In mobility service environments, recognizing the user condition and driving status is critical in driving safety and experiences. While speech emotion recognition is one of the possible features to predict the driver status, current emotion recognition models have a fundamental limitation: they target to classify only single emotion classes, not multi-classes. It prevents the comprehensive understanding of the driver’s condition and intention during driving. In addition, mobility devices inherently generate noises that might affect speech emotion recognition performances in the mobility service. Considering mobility service environments, we investigate possible models that detect multiple emotions while mitigating noise issues. In this paper, we propose a speech-emotion recognition model based on the autoencoder for multi-emotion detection. First, we analyze the Mel Frequency Cepstral Coefficients (MFCCs) to design the specific features. We also develop a multi-emotion detection scheme based on an autoencoder to detect multiple emotions with substantial flexibility compared to existing models. With our proposed scheme, we investigate and analyze mobility noise impacts and mitigation approaches to evaluate performance results.
- Author(s)
- 김준영; 오정민; 김진관
- Issued Date
- 2025-05-08
- Type
- Article
- Keyword
- 공학
- DOI
- 10.3390/electronics14101915
- URI
- http://repository.sungshin.ac.kr/handle/2025.oak/8780
- Publisher
- MDPI
- ISSN
- 2079-9292
-
Appears in Collections:
- AI융합학부 > 학술논문
- 공개 및 라이선스
-
- 파일 목록
-
Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.