OAK

Multi-Detection-Based Speech Emotion Recognition Using Autoencoder in Mobility Service Environment

Metadata Downloads
Abstract
In mobility service environments, recognizing the user condition and driving status is critical in driving safety and experiences. While speech emotion recognition is one of the possible features to predict the driver status, current emotion recognition models have a fundamental limitation: they target to classify only single emotion classes, not multi-classes. It prevents the comprehensive understanding of the driver’s condition and intention during driving. In addition, mobility devices inherently generate noises that might affect speech emotion recognition performances in the mobility service. Considering mobility service environments, we investigate possible models that detect multiple emotions while mitigating noise issues. In this paper, we propose a speech-emotion recognition model based on the autoencoder for multi-emotion detection. First, we analyze the Mel Frequency Cepstral Coefficients (MFCCs) to design the specific features. We also develop a multi-emotion detection scheme based on an autoencoder to detect multiple emotions with substantial flexibility compared to existing models. With our proposed scheme, we investigate and analyze mobility noise impacts and mitigation approaches to evaluate performance results.
Author(s)
김준영오정민김진관
Issued Date
2025-05-08
Type
Article
Keyword
공학
DOI
10.3390/electronics14101915
URI
http://repository.sungshin.ac.kr/handle/2025.oak/8780
Publisher
MDPI
ISSN
2079-9292
Appears in Collections:
AI융합학부 > 학술논문
공개 및 라이선스
  • 공개 구분공개
파일 목록
  • 관련 파일이 존재하지 않습니다.

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.