Selecting Emotion Specific Speech Features To Distinguish One Emotion From Others
| dc.contributor.author | Ozkan C. | |
| dc.contributor.author | Oguz K. | |
| dc.date.accessioned | 2023-06-16T15:00:49Z | |
| dc.date.available | 2023-06-16T15:00:49Z | |
| dc.date.issued | 2021 | |
| dc.description | Kocaeli University;Kocaeli University Technopark | en_US |
| dc.description | 2021 International Conference on INnovations in Intelligent SysTems and Applications, INISTA 2021 -- 25 August 2021 through 27 August 2021 -- 172175 | en_US |
| dc.description.abstract | Speech is one of the most studied modalities of emotion recognition. Most studies use one or more labeled data sets that contain multiple emotions to extract and select speech features to be trained by machine learning algorithms. Instead of this multi-class approach, our study focuses on selecting features that most distinguish an emotion from others. This requires a one-against-all (OAA) binary classification approach. The features that are extracted and selected for the multi-class case is compared to features extracted for seven one-against-all cases using a standard backpropagation feedforward neural network (BFNN). The results while OAA distinguishes some of the emotions better than the multi-class BFNN configurations, this is not true for all cases. However, when multi-class BFNN is tested with all emotions, the error rate is as high as 16.48. © 2021 IEEE. | en_US |
| dc.identifier.doi | 10.1109/INISTA52262.2021.9548533 | |
| dc.identifier.isbn | 9.78E+12 | |
| dc.identifier.scopus | 2-s2.0-85116665979 | |
| dc.identifier.uri | https://doi.org/10.1109/INISTA52262.2021.9548533 | |
| dc.identifier.uri | https://hdl.handle.net/20.500.14365/3568 | |
| dc.language.iso | en | en_US |
| dc.publisher | Institute of Electrical and Electronics Engineers Inc. | en_US |
| dc.relation.ispartof | 2021 International Conference on INnovations in Intelligent SysTems and Applications, INISTA 2021 - Proceedings | en_US |
| dc.rights | info:eu-repo/semantics/closedAccess | en_US |
| dc.subject | Artificial neural network | en_US |
| dc.subject | Feature selection | en_US |
| dc.subject | Speech emotion recognition | en_US |
| dc.subject | Feedforward neural networks | en_US |
| dc.subject | Learning algorithms | en_US |
| dc.subject | Machine learning | en_US |
| dc.subject | Speech | en_US |
| dc.subject | Speech recognition | en_US |
| dc.subject | Binary Classification Approach | en_US |
| dc.subject | Data set | en_US |
| dc.subject | Emotion recognition | en_US |
| dc.subject | Error rate | en_US |
| dc.subject | Features selection | en_US |
| dc.subject | Labeled data | en_US |
| dc.subject | Machine learning algorithms | en_US |
| dc.subject | Neural network configurations | en_US |
| dc.subject | Speech emotion recognition | en_US |
| dc.subject | Speech features | en_US |
| dc.subject | Feature extraction | en_US |
| dc.title | Selecting Emotion Specific Speech Features To Distinguish One Emotion From Others | en_US |
| dc.type | Conference Object | en_US |
| dspace.entity.type | Publication | |
| gdc.author.scopusid | 57289727300 | |
| gdc.bip.impulseclass | C5 | |
| gdc.bip.influenceclass | C5 | |
| gdc.bip.popularityclass | C5 | |
| gdc.coar.access | metadata only access | |
| gdc.coar.type | text::conference output | |
| gdc.collaboration.industrial | false | |
| gdc.description.departmenttemp | Ozkan, C., Izmir University of Economics, Department of Computer Engineering, Izmir, Turkey; Oguz, K., Izmir University of Economics, Department of Computer Engineering, Izmir, Turkey | en_US |
| gdc.description.endpage | 5 | |
| gdc.description.publicationcategory | Konferans Öğesi - Uluslararası - Kurum Öğretim Elemanı | en_US |
| gdc.description.scopusquality | N/A | |
| gdc.description.startpage | 1 | |
| gdc.description.wosquality | N/A | |
| gdc.identifier.openalex | W3204659206 | |
| gdc.index.type | Scopus | |
| gdc.oaire.diamondjournal | false | |
| gdc.oaire.impulse | 1.0 | |
| gdc.oaire.influence | 2.619421E-9 | |
| gdc.oaire.isgreen | false | |
| gdc.oaire.popularity | 2.4000324E-9 | |
| gdc.oaire.publicfunded | false | |
| gdc.oaire.sciencefields | 0202 electrical engineering, electronic engineering, information engineering | |
| gdc.oaire.sciencefields | 02 engineering and technology | |
| gdc.openalex.collaboration | National | |
| gdc.openalex.fwci | 0.447 | |
| gdc.openalex.normalizedpercentile | 0.56 | |
| gdc.opencitations.count | 1 | |
| gdc.plumx.mendeley | 5 | |
| gdc.plumx.scopuscites | 2 | |
| gdc.scopus.citedcount | 2 | |
| gdc.virtual.author | Oğuz, Kaya | |
| relation.isAuthorOfPublication | 352071e4-5cb7-4239-be4d-3132ba33986c | |
| relation.isAuthorOfPublication.latestForDiscovery | 352071e4-5cb7-4239-be4d-3132ba33986c | |
| relation.isOrgUnitOfPublication | b4714bc5-c5ae-478f-b962-b7204c948b70 | |
| relation.isOrgUnitOfPublication | 26a7372c-1a5e-42d9-90b6-a3f7d14cad44 | |
| relation.isOrgUnitOfPublication | e9e77e3e-bc94-40a7-9b24-b807b2cd0319 | |
| relation.isOrgUnitOfPublication.latestForDiscovery | b4714bc5-c5ae-478f-b962-b7204c948b70 |
Files
Original bundle
1 - 1 of 1
