Pesquisa de referências

Multimodal emotion recognition for emphatic virtual agents in mental health interventions

<?xml version="1.0" encoding="UTF-8" standalone="no"?>
<rdf:RDF xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
<rdf:Description>
<dc:creator>Huerta Espinoza, Marcelo Alejandro</dc:creator>
<dc:creator>Rodriguez Gonzalez, Ansel Y.</dc:creator>
<dc:creator>Martinez Miranda, Juan</dc:creator>
<dc:creator>IBERAMIA, Sociedad Iberoamericana de Inteligencia Artificial
 </dc:creator>
<dc:date>2025-12-08</dc:date>
<dc:description xml:lang="es">Sumario: Depression and anxiety disorders affect millions of individuals globally and are commonly addressed through psychological interventions. A growing technological approach to support such treatments involves the use of embodied conversational agents that employ motivational interviewing, a method that promotes behavioral change through empathic engagement. Despite its critical role in therapeutic efficacy, empathy remains a significant challenge for virtual agents to emulate. Emotion Recognition (ER) technologies offer a potential solution by enabling agents to perceive and respond appropriately to users' emotional states. Given the inherently multimodal nature of human emotion, unimodal ER approaches often fall short in accurately interpreting affective cues. In this work, we propose a multimodal emotion recognition model that integrates verbal and non-verbal signals (text and video) using a Cross-Modal Attention fusion strategy. Trained and evaluated on the IEMOCAP dataset, our approach leverages Ekman's taxonomy of basic emotions and demonstrates superior performance over unimodal baselines across key metrics such as accuracy and F1-score. By prioritizing text as the main modality and dynamically incorporating complementary visual cues, the model proves effective in complex emotion classification tasks. The proposed model is designed for integration into an existing conversational agent aimed at supporting individuals experiencing emotional and psychological distress. Future work will involve embedding the model in the conversational agent platform for emotionally distressed users, aiming to assess its real-world impact on engagement, user experience, and perceived empathy</dc:description>
<dc:identifier>https://documentacion.fundacionmapfre.org/documentacion/publico/es/bib/189503.do</dc:identifier>
<dc:language>eng</dc:language>
<dc:rights xml:lang="es">InC - http://rightsstatements.org/vocab/InC/1.0/</dc:rights>
<dc:subject xml:lang="es">Inteligencia artificial</dc:subject>
<dc:subject xml:lang="es">Salud mental</dc:subject>
<dc:subject xml:lang="es">Medicina bioelectrónica</dc:subject>
<dc:subject xml:lang="es">Depresión</dc:subject>
<dc:subject xml:lang="es">Ansiedad</dc:subject>
<dc:type xml:lang="es">Artículos y capítulos</dc:type>
<dc:title xml:lang="es">Multimodal emotion recognition for emphatic virtual agents in mental health interventions</dc:title>
<dc:relation xml:lang="es">En: Revista Iberoamericana de Inteligencia Artificial. -  : IBERAMIA, Sociedad Iberoamericana de Inteligencia Artificial , 2018- = ISSN 1988-3064. - 08/12/2025 Volume 28 Number  76 - December 2025 , p. 28 - 39</dc:relation>
</rdf:Description>
</rdf:RDF>