Publication:
UniC: a dataset for emotion analysis of videos with multimodal and unimodal labels
| cris.virtual.department | #PLACEHOLDER_PARENT_METADATA_VALUE# | |
| cris.virtual.orcid | 0000-0002-9901-5768 | |
| cris.virtualsource.department | df6c83d3-392b-4c86-82f0-1f3fadc2f1fd | |
| cris.virtualsource.orcid | df6c83d3-392b-4c86-82f0-1f3fadc2f1fd | |
| dc.contributor.author | Du, Quanqi | |
| dc.contributor.author | Labat, Sofie | |
| dc.contributor.author | Demeester, Thomas | |
| dc.contributor.author | Hoste, Veronique | |
| dc.contributor.imecauthor | Demeester, Thomas | |
| dc.contributor.orcidimec | Demeester, Thomas::0000-0002-9901-5768 | |
| dc.date.accessioned | 2025-05-26T13:09:59Z | |
| dc.date.available | 2025-05-25T05:33:57Z | |
| dc.date.available | 2025-05-26T13:09:59Z | |
| dc.date.issued | 2025 | |
| dc.description.abstract | Emotion is a key characteristic that differentiates humans from machines. It is intricate, encompassing a wide variety of emotional states, and is expressed through both verbal and non-verbal communication channels. Different modalities contribute in unique ways to the integrated expression of emotion. However, in most of the existing multimodal datasets, there is only one unified emotion label for the various modalities, ignoring the heterogeneity and complementarity of the different modalities. To bridge this gap, we introduce UniC, a novel multimodal emotion dataset featuring both integrated multimodal labels and independent unimodal labels. UniC is comprised of 965 emotion-rich video clips selected from YouTube, annotated in text, audio, silent video, and multimodal setups with both categorical and dimensional labels. We outline the steps taken to construct the dataset and analyze different modality perspectives in UniC. Our findings indicate that while in most cases the modality of text shares more emotional resemblance with the multimodal setup, other modalities can exhibit different, sometimes even opposite emotions that might contribute more to the overall emotion state. This dataset offers a modality-specific perspective on multimodal emotion analysis and has the potential to provide valuable insights for further research in human emotion understanding. | |
| dc.description.wosFundingText | This research received funding from the Flemish Government under the Research Program Artificial Intelligence (174K02325) and from the Research Foundation Flanders (FWO-Vlaanderen) with grant number 1S96322N. We would also like to thank the anonymous reviewers for their valuable and constructive feedback. | |
| dc.identifier.doi | 10.1007/s10579-025-09837-0 | |
| dc.identifier.issn | 1574-020X | |
| dc.identifier.uri | https://imec-publications.be/handle/20.500.12860/45707 | |
| dc.publisher | SPRINGER | |
| dc.source.beginpage | 2857 | |
| dc.source.endpage | 2892 | |
| dc.source.journal | LANGUAGE RESOURCES AND EVALUATION | |
| dc.source.numberofpages | 36 | |
| dc.source.volume | 2025 | |
| dc.subject.keywords | RECOGNITION | |
| dc.title | UniC: a dataset for emotion analysis of videos with multimodal and unimodal labels | |
| dc.type | Journal article | |
| dspace.entity.type | Publication | |
| Files | Original bundle
| |
| Publication available in collections: |