Publicação:
BEE-MER: Bimodal Embeddings Ensemble for Music Emotion Recognition

dc.contributor.authorLima Louro, Pedro Miguel
dc.contributor.authorRibeiro, Tiago F. R.
dc.contributor.authorMalheiro, Ricardo
dc.contributor.authorPanda, Renato
dc.contributor.authorPinto de Carvalho e Paiva, Rui Pedro
dc.date.accessioned2025-10-20T13:22:52Z
dc.date.available2025-10-20T13:22:52Z
dc.date.issued2025-07-07
dc.description.abstractStatic music emotion recognition systems typically focus on audio for classification, although some research has explored the potential of analyzing lyrics as well. Both approaches face challenges when it comes to accurately discerning emotions that have similar energy but differing valence, and vice versa, depending on the modality used. Previous studies have introduced bimodal audio-lyrics systems that outperform single-modality solutions by combining information from standalone systems and conducting joint classification. In this study, we propose and compare two bimodal approaches: one strictly based on embedding models (audio and word embeddings) and another one following a standard spectrogram-based deep learning method for the audio part. Additionally, we explore various information fusion strategies to leverage both modalities effectively. The main conclusions of this work are the following: i) the two approaches show comparable overall classification performance; ii) the embedding-only approach leads to a higher confusion between quadrants 3 and 4 of Russell’s circumplex model; iii) and this approach requires significantly less computational cost for training. We discuss the insights gained from the approaches we experimented with and highlight promising avenues for future research.eng
dc.identifier.source-work-idcv-prod-id-4810273
dc.identifier.urihttp://hdl.handle.net/10400.26/59278
dc.language.isoeng
dc.peerreviewedyes
dc.publisher7
dc.relationMERGE
dc.rights.urihttp://creativecommons.org/licenses/by/4.0/
dc.titleBEE-MER: Bimodal Embeddings Ensemble for Music Emotion Recognitioneng
dc.typeconference paper
dcterms.referenceshttps://zenodo.org/records/13939205
dcterms.referenceshttps://arxiv.org/abs/2407.06060
dspace.entity.typePublication
oaire.citation.conferenceDate2025-07-07
oaire.citation.conferencePlaceGraz, Austria
oaire.citation.title22nd Sound and Music Computing Conference – SMC 2025
oaire.versionhttp://purl.org/coar/version/c_ab4af688f83e57aa
relation.isAuthorOfPublication953b69db-f6a0-4a54-8618-93a912df6df6
relation.isAuthorOfPublicationd38dd344-0942-4fcb-b740-a4713fb170e7
relation.isAuthorOfPublication9cd470af-3968-45cc-ad6f-3b59e00ae823
relation.isAuthorOfPublication238ee6f8-61cd-49b4-9392-9e7763fd35f3
relation.isAuthorOfPublication.latestForDiscovery953b69db-f6a0-4a54-8618-93a912df6df6
relation.isProjectOfPublication32dd2667-ca4e-430b-a130-687ba6eee2e9
relation.isProjectOfPublication.latestForDiscovery32dd2667-ca4e-430b-a130-687ba6eee2e9

Ficheiros

Pacote original

A mostrar 1 - 1 de 1
A carregar...
Miniatura
Nome:
SMC_2025_Louro.pdf
Tamanho:
203.61 KB
Formato:
Adobe Portable Document Format

Licença do pacote

A mostrar 1 - 1 de 1
A carregar...
Miniatura
Nome:
license.txt
Tamanho:
1.85 KB
Formato:
Item-specific license agreed upon to submission
Descrição: