Please use this identifier to cite or link to this item: https://hdl.handle.net/20.500.14365/1445
Title: Speech emotion recognition: Emotional models, databases, features, preprocessing methods, supporting modalities, and classifiers
Authors: Akcay, Mehmet Berkehan
Oguz, Kaya
Keywords: Speech emotion recognition
Survey
Speech features
Classification
Speech databases
Voice Quality
Communicating Emotion
Spectral Features
Neural-Networks
Classification
Valence
Expression
Arousal
Adversarial
Audio
Publisher: Elsevier
Abstract: Speech is the most natural way of expressing ourselves as humans. It is only natural then to extend this communication medium to computer applications. We define speech emotion recognition (SER) systems as a collection of methodologies that process and classify speech signals to detect the embedded emotions. SER is not a new field, it has been around for over two decades, and has regained attention thanks to the recent advancements. These novel studies make use of the advances in all fields of computing and technology, making it necessary to have an update on the current methodologies and techniques that make SER possible. We have identified and discussed distinct areas of SER, provided a detailed survey of current literature of each, and also listed the current challenges.
URI: https://doi.org/10.1016/j.specom.2019.12.001
https://hdl.handle.net/20.500.14365/1445
ISSN: 0167-6393
1872-7182
Appears in Collections:Scopus İndeksli Yayınlar Koleksiyonu / Scopus Indexed Publications Collection
WoS İndeksli Yayınlar Koleksiyonu / WoS Indexed Publications Collection

Files in This Item:
File SizeFormat 
492.pdf
  Restricted Access
1.16 MBAdobe PDFView/Open    Request a copy
Show full item record



CORE Recommender

SCOPUSTM   
Citations

500
checked on Nov 6, 2024

WEB OF SCIENCETM
Citations

319
checked on Nov 6, 2024

Page view(s)

130
checked on Nov 11, 2024

Download(s)

6
checked on Nov 11, 2024

Google ScholarTM

Check




Altmetric


Items in GCRIS Repository are protected by copyright, with all rights reserved, unless otherwise indicated.