Deep Learning Based Facial Emotion Recognition System
| dc.contributor.author | Ozdemir M.A. | |
| dc.contributor.author | Elagoz B. | |
| dc.contributor.author | Alaybeyoglu Soy A. | |
| dc.contributor.author | Akan A. | |
| dc.date.accessioned | 2023-06-16T15:01:50Z | |
| dc.date.available | 2023-06-16T15:01:50Z | |
| dc.date.issued | 2020 | |
| dc.description | 2020 Medical Technologies Congress, TIPTEKNO 2020 -- 19 November 2020 through 20 November 2020 -- 166140 | en_US |
| dc.description.abstract | In this study, it was aimed to recognize the emotional state from facial images using the deep learning method. In the study, which was approved by the ethics committee, a custom data set was created using videos taken from 20 male and 20 female participants while simulating 7 different facial expressions (happy, sad, surprised, angry, disgusted, scared, and neutral). Firstly, obtained videos were divided into image frames, and then face images were segmented using the Haar library from image frames. The size of the custom data set obtained after the image preprocessing is more than 25 thousand images. The proposed convolutional neural network (CNN) architecture which is mimics of LeNet architecture has been trained with this custom dataset. According to the proposed CNN architecture experiment results, the training loss was found as 0.0115, the training accuracy was found as 99.62%, the validation loss was 0.0109, and the validation accuracy was 99.71%. © 2020 IEEE. | en_US |
| dc.identifier.doi | 10.1109/TIPTEKNO50054.2020.9299256 | |
| dc.identifier.isbn | 9.78E+12 | |
| dc.identifier.scopus | 2-s2.0-85099442002 | |
| dc.identifier.uri | https://doi.org/10.1109/TIPTEKNO50054.2020.9299256 | |
| dc.identifier.uri | https://hdl.handle.net/20.500.14365/3638 | |
| dc.language.iso | tr | en_US |
| dc.publisher | Institute of Electrical and Electronics Engineers Inc. | en_US |
| dc.relation.ispartof | TIPTEKNO 2020 - Tip Teknolojileri Kongresi - 2020 Medical Technologies Congress, TIPTEKNO 2020 | en_US |
| dc.rights | info:eu-repo/semantics/closedAccess | en_US |
| dc.subject | Deep Learning | en_US |
| dc.subject | Emotion Recognition | en_US |
| dc.subject | Facial Expression | en_US |
| dc.subject | Biomedical engineering | en_US |
| dc.subject | Convolutional neural networks | en_US |
| dc.subject | Face recognition | en_US |
| dc.subject | Learning systems | en_US |
| dc.subject | Network architecture | en_US |
| dc.subject | Emotional state | en_US |
| dc.subject | Ethics committee | en_US |
| dc.subject | Facial emotions | en_US |
| dc.subject | Facial Expressions | en_US |
| dc.subject | Facial images | en_US |
| dc.subject | Image preprocessing | en_US |
| dc.subject | Learning methods | en_US |
| dc.subject | Training accuracy | en_US |
| dc.subject | Deep learning | en_US |
| dc.title | Deep Learning Based Facial Emotion Recognition System | en_US |
| dc.title.alternative | Derin Ögrenme Tabanli Yüz Duygulari Tanima Sistemi | en_US |
| dc.type | Conference Object | en_US |
| dspace.entity.type | Publication | |
| gdc.author.scopusid | 57206479576 | |
| gdc.author.scopusid | 57221553007 | |
| gdc.author.scopusid | 35617283100 | |
| gdc.bip.impulseclass | C4 | |
| gdc.bip.influenceclass | C5 | |
| gdc.bip.popularityclass | C4 | |
| gdc.coar.access | metadata only access | |
| gdc.coar.type | text::conference output | |
| gdc.collaboration.industrial | false | |
| gdc.description.departmenttemp | Ozdemir, M.A., Izmir Katip Celebi University, Department of Biomedical Engineering, Izmir, Turkey; Elagoz, B., Izmir Katip Celebi University, Department of Biomedical Technologies, Izmir, Turkey; Alaybeyoglu Soy, A., Izmir Katip Celebi University, Department of Computer Engineering, Izmir, Turkey; Akan, A., Izmir University of Economics, Department of Electrical and Electronics Engineering, Izmir, Turkey | en_US |
| gdc.description.publicationcategory | Konferans Öğesi - Uluslararası - Kurum Öğretim Elemanı | en_US |
| gdc.description.scopusquality | N/A | |
| gdc.description.wosquality | N/A | |
| gdc.identifier.openalex | W3118012581 | |
| gdc.identifier.wos | WOS:000659419900042 | |
| gdc.index.type | WoS | |
| gdc.index.type | Scopus | |
| gdc.oaire.diamondjournal | false | |
| gdc.oaire.impulse | 7.0 | |
| gdc.oaire.influence | 3.1313974E-9 | |
| gdc.oaire.isgreen | false | |
| gdc.oaire.popularity | 7.885164E-9 | |
| gdc.oaire.publicfunded | false | |
| gdc.oaire.sciencefields | 03 medical and health sciences | |
| gdc.oaire.sciencefields | 0302 clinical medicine | |
| gdc.oaire.sciencefields | 0202 electrical engineering, electronic engineering, information engineering | |
| gdc.oaire.sciencefields | 02 engineering and technology | |
| gdc.openalex.collaboration | National | |
| gdc.openalex.fwci | 1.7084 | |
| gdc.openalex.normalizedpercentile | 0.85 | |
| gdc.opencitations.count | 7 | |
| gdc.plumx.crossrefcites | 4 | |
| gdc.plumx.mendeley | 37 | |
| gdc.plumx.scopuscites | 10 | |
| gdc.scopus.citedcount | 10 | |
| gdc.virtual.author | Akan, Aydın | |
| gdc.wos.citedcount | 3 | |
| relation.isAuthorOfPublication | 9b1a1d3d-05af-4982-b7d1-0fefff6ac9fd | |
| relation.isAuthorOfPublication.latestForDiscovery | 9b1a1d3d-05af-4982-b7d1-0fefff6ac9fd | |
| relation.isOrgUnitOfPublication | b02722f0-7082-4d8a-8189-31f0230f0e2f | |
| relation.isOrgUnitOfPublication | 26a7372c-1a5e-42d9-90b6-a3f7d14cad44 | |
| relation.isOrgUnitOfPublication | e9e77e3e-bc94-40a7-9b24-b807b2cd0319 | |
| relation.isOrgUnitOfPublication.latestForDiscovery | b02722f0-7082-4d8a-8189-31f0230f0e2f |
Files
Original bundle
1 - 1 of 1
