Development of a speech emotion recognizer for large-scale child-centered audio recordings from a hospital environment

dc.contributor.authorVaaras Einari
dc.contributor.authorAhlqvist-Björkroth Sari
dc.contributor.authorDrossos Konstantinos
dc.contributor.authorLehtonen Liisa
dc.contributor.authorRäsänen Okko
dc.contributor.organizationfi=lastentautioppi|en=Paediatrics and Adolescent Medicine|
dc.contributor.organizationfi=psykologia|en=Psychology|
dc.contributor.organizationfi=tyks, vsshp|en=tyks, varha|
dc.contributor.organization-code1.2.246.10.2458963.20.15586825505
dc.contributor.organization-code1.2.246.10.2458963.20.40612039509
dc.converis.publication-id179057384
dc.converis.urlhttps://research.utu.fi/converis/portal/Publication/179057384
dc.date.accessioned2025-08-27T22:22:58Z
dc.date.available2025-08-27T22:22:58Z
dc.description.abstract<p>In order to study how early emotional experiences shape infant development, one approach is to analyze the emotional content of <a href="https://www.sciencedirect.com/topics/social-sciences/logopedics" title="Learn more about speech from ScienceDirect's AI-generated Topic Pages">speech</a> heard by infants, as captured by child-centered daylong recordings, and as analyzed by automatic speech emotion recognition (SER) systems. However, since large-scale daylong audio is initially unannotated and differs from typical speech corpora from controlled environments, there are no existing in-domain SER systems for the task. Based on existing literature, it is also unclear what is the best approach to deploy a SER system for a new domain. Consequently, in this study, we investigated alternative strategies for deploying a SER system for large-scale child-centered <a href="https://www.sciencedirect.com/topics/computer-science/audio-recording" title="Learn more about audio recordings from ScienceDirect's AI-generated Topic Pages">audio recordings</a> from a neonatal hospital environment, comparing cross-corpus <a href="https://www.sciencedirect.com/topics/computer-science/generalization" title="Learn more about generalization from ScienceDirect's AI-generated Topic Pages">generalization</a>, active learning (AL), and <a href="https://www.sciencedirect.com/topics/computer-science/domain-adaptation" title="Learn more about domain adaptation from ScienceDirect's AI-generated Topic Pages">domain adaptation</a> (DA) methods in the process. We first conducted simulations with existing emotion-labeled speech corpora to find the best strategy for SER system deployment. We then tested how the findings generalize to our new initially unannotated dataset. As a result, we found that the studied AL method provided overall the most consistent results, being less dependent on the specifics of the training corpora or speech features compared to the alternative methods. However, in situations without the possibility to annotate data, unsupervised DA proved to be the best approach. We also observed that deployment of a SER system for real-world daylong child-centered audio recordings achieved a SER performance level comparable to those reported in literature, and that the amount of human effort required for the system deployment was overall relatively modest.<br></p>
dc.format.pagerange22
dc.format.pagerange9
dc.identifier.eissn1872-7182
dc.identifier.jour-issn0167-6393
dc.identifier.olddbid202079
dc.identifier.oldhandle10024/185106
dc.identifier.urihttps://www.utupub.fi/handle/11111/45394
dc.identifier.urlhttps://doi.org/10.1016/j.specom.2023.02.001
dc.identifier.urnURN:NBN:fi-fe2023033033889
dc.language.isoen
dc.okm.affiliatedauthorAhlqvist-Björkroth, Sari
dc.okm.affiliatedauthorLehtonen, Liisa
dc.okm.affiliatedauthorDataimport, tyks, vsshp
dc.okm.discipline515 Psychologyen_GB
dc.okm.discipline515 Psykologiafi_FI
dc.okm.internationalcopublicationnot an international co-publication
dc.okm.internationalityInternational publication
dc.okm.typeA1 ScientificArticle
dc.publisherElsevier B.V.
dc.publisher.countryNetherlandsen_GB
dc.publisher.countryAlankomaatfi_FI
dc.publisher.country-codeNL
dc.relation.doi10.1016/j.specom.2023.02.001
dc.relation.ispartofjournalSpeech Communication
dc.relation.volume148
dc.source.identifierhttps://www.utupub.fi/handle/10024/185106
dc.titleDevelopment of a speech emotion recognizer for large-scale child-centered audio recordings from a hospital environment
dc.year.issued2023

Tiedostot

Näytetään 1 - 1 / 1
Ladataan...
Name:
1-s2.0-S0167639323000262-main (1).pdf
Size:
1.25 MB
Format:
Adobe Portable Document Format