Archives and Documentation Center
Digital Archives

Fusing acoustic and linguistic parameters for multilingual emotion recognition

Show simple item record

dc.contributor Graduate Program in Electrical and Electronic Engineering.
dc.contributor.advisor Arslan, Levent M.
dc.contributor.author Erden, Mustafa.
dc.date.accessioned 2023-03-16T10:17:35Z
dc.date.available 2023-03-16T10:17:35Z
dc.date.issued 2011.
dc.identifier.other EE 2011 E73
dc.identifier.uri http://digitalarchive.boun.edu.tr/handle/123456789/12780
dc.description.abstract Emotion recognition from speech can be used for detection of customer problems in call centers, agent performance monitoring, improving automatic speech recognition accuracies, enhancing human robot as well as human machine interaction. In this thesis two di erent spontaneous databases are investigated in terms of binary emotion classi cation. On Turkish call center dataset (CCD) which consists of human-human dialogs, emotion recognition problem is de ned on angry and non-angry classes. On Fau Aibo dataset (FAD) which is composed of recordings of children playing with a pet robot, the negative and idle classes are considered. For extracting acoustic information we have implemented Support Vector Machines with utterance level features and Gaussian Mixture Models with frame level features. In terms of language modeling we compared word based, stem-only and stem+ending structures using manual transcriptions. Stem+ending based system resulted in the highest accuracies on CCD whereas the word based LM performed the best on FAD. This can be mainly attributed to the agglutinative nature of Turkish language. When we fused the acoustic and LM classi ers using a Multi Layer Perceptron (MLP) we could achieve 89% and 69% correct detection of both classes for CCD and FAD respectively.
dc.format.extent 30cm.
dc.publisher Thesis (M.S.) - Bogazici University. Institute for Graduate Studies in Science and Engineering, 2011.
dc.subject.lcsh Discourse analysis.
dc.subject.lcsh Historical linguistics.
dc.title Fusing acoustic and linguistic parameters for multilingual emotion recognition
dc.format.pages xii, 49 leaves ;


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search Digital Archive


Browse

My Account