Show simple item record

dc.creatorSpyrou E., Vernikos I., Nikopoulou R., Mylonas P.en
dc.date.accessioned2023-01-31T10:01:35Z
dc.date.available2023-01-31T10:01:35Z
dc.date.issued2019
dc.identifier10.1109/IISA.2018.8633644
dc.identifier.isbn9781538637319
dc.identifier.urihttp://hdl.handle.net/11615/79346
dc.description.abstractOne of the most important issues in several aspects of human-computer interaction is the understanding of the users' emotional state. In several applications such as monitoring of humans in assistive living environments, or assessing students' affective state during a course, it is imperative to use an unobtrusive method, so as to avoid discomforting or distracting the user. Thus, one should opt for approaches that use either visual or audio sensors which may observe users without any kind of direct contact. In this work, our goal is to recognize the emotional state of humans using only the non-linguistic aspect of speech information, i.e., the acoustic properties of speech. Therefore, we propose an emotion classification that is based on the bag-of-visual words model that has been previously applied in many computer vision tasks. A given audio segment is transformed to a spectrogram, i.e., a visual representation of its spectrum. From this representation we first extract SURF features and using a previously constructed visual vocabulary, we quantize them into a set of visual words. Then a histogram is constructed per image; These feature vectors are used to train SVM classifiers. Our approach is evaluated using a) 3 publicly available datasets that contain speech from different languages and b) a custom dataset that has been constructed during a real-life classroom experiments, involving middle-school students. ©2018 IEEEen
dc.language.isoenen
dc.source2018 9th International Conference on Information, Intelligence, Systems and Applications, IISA 2018en
dc.source.urihttps://www.scopus.com/inward/record.uri?eid=2-s2.0-85062861017&doi=10.1109%2fIISA.2018.8633644&partnerID=40&md5=013f27a62a341e06db33c59b306230e7
dc.subjectAcoustic propertiesen
dc.subjectClassification (of information)en
dc.subjectHuman computer interactionen
dc.subjectLinguisticsen
dc.subjectStudentsen
dc.subjectBag-of-visual-wordsen
dc.subjectEmotion classificationen
dc.subjectHuman emotion recognitionen
dc.subjectLinguistic approachen
dc.subjectMiddle school studentsen
dc.subjectSpeech informationen
dc.subjectVisual representationsen
dc.subjectVisual vocabulariesen
dc.subjectSpeech recognitionen
dc.subjectInstitute of Electrical and Electronics Engineers Inc.en
dc.titleA non-linguistic approach for human emotion recognition from speechen
dc.typeconferenceItemen


Files in this item

FilesSizeFormatView

There are no files associated with this item.

This item appears in the following Collection(s)

Show simple item record