Gumelar, Agustinus Bimo and Yuniarno, Eko Mulyanto and Adi, Derry Pramono and Setiawan, Rudi and Sugiarto, Indar and Purnomo, Mauridhi Hery (2023) Transformer-CNN Automatic Hyperparameter Tuning for Speech Emotion Recognition. In: 022 IEEE International Conference on Imaging Systems and Techniques (IST), 23-06-2023 - 23-06-2023, Kaohsiung - Taiwan.
![]() | PDF Download (572Kb) |
![]() | PDF Download (2492Kb) |
Abstract
Given the high number of hyperparameters in deep learning models, there is a need to tune automatically deep learning models in specific research cases. Deep learning models require hyperparameters because they substantially influence the models behavior. As a result, optimizing any given model with a hyperparameter optimization technique will improve model efficiency significantly. This paper discusses the hyperparameter-optimized Speech Emotion Recognition (SER) research case using Transformer-CNN deep learning model. Each speech samples are transformed into spectrogram data using the RAVDESS dataset, which contains 1,536 speech samples (192 samples per eight emotion classes). We use the Gaussian Noise augmentation technique to reduce the overfitting problem in training data. After augmentation, the RAVDESS dataset yields a total of 2,400 emotional speech samples (300 samples per eight emotion classes). For SER model, we combine the Transformer and CNN for temporal and spatial speech feature processing. However, our Transformer-CNN must be thoroughly tested, as different hyperparameter settings result in varying accuracy performance. We experiment with Naive Bayes to optimize many hyperparameters of Transformer-CNN (it could be categorical or numerical), such as learning rate, dropouts, activation function, weight initialization, epoch, even the best split data scale of training and testing. Consequently, our automatically tuned Transformer-CNN achieves 97.3 % of accuracy.
Item Type: | Conference or Workshop Item (Paper) |
---|---|
Uncontrolled Keywords: | Speech Emotion Recognition, Automatic Hyperparameter Tuning, Transformer-CNN, Naive Bayes Optimization |
Subjects: | Q Science > QA Mathematics > QA75 Electronic computers. Computer science |
Divisions: | Faculty of Industrial Technology > Electrical Engineering Department |
Depositing User: | Admin |
Date Deposited: | 16 Sep 2025 22:59 |
Last Modified: | 18 Sep 2025 21:53 |
URI: | https://repository.petra.ac.id/id/eprint/21851 |
Actions (login required)
View Item |