A talk will present two projects related to the use of NLP for estimating a client’s depression severity and well-being. The first project examines emotional coherence between the subjective experience of emotions and emotion expression in therapy using transformer-based emotion recognition models. The second project proposes a semantic pipeline to study depression severity in individuals based on their social media posts by exploring different aggregation methods to answer one of four Beck Depression Inventory (BDI) options per symptom. Why it matters: This research explores how NLP techniques can be applied to mental health assessment, potentially offering new tools for diagnosis and treatment monitoring.
This research introduces a novel method using the Lateral Accretive Hybrid Network (LEARNet) to capture and analyze micro-expressions for mental health applications. The method refines both broad and subtle facial cues to detect mental health conditions like anxiety or depression. The authors also propose a neural architecture search (NAS) strategy to design a compact CNN for micro-expression recognition, improving performance and resource use. Why it matters: By integrating micro-emotion recognition with mental health estimation, the approach enables more accurate and early detection of emotional and mental health issues, potentially leading to improved well-being.
Justice Connect, an Australian charity, collaborated with MBZUAI's Prof. Timothy Baldwin to improve their legal intake tool using NLP. The tool helps route legal requests, but users struggled to identify the relevant area of law, leading to delays and frustration. By applying NLP, the collaboration aims to help users more easily navigate the tool and access appropriate legal resources. Why it matters: This project demonstrates how NLP can be applied to improve access to justice and address unmet legal needs, particularly for those unfamiliar with legal terminology.
The InterText project, funded by the European Research Council, aims to advance NLP by developing a framework for modeling fine-grained relationships between texts. This approach enables tracing the origin and evolution of texts and ideas. Iryna Gurevych from the Technical University of Darmstadt presented the intertextual approach to NLP, covering data modeling, representation learning, and practical applications. Why it matters: This research could enable a new generation of AI applications for text work and critical reading, with potential applications in collaborative knowledge construction and document revision assistance.
This study investigates the ability of six large language models, including Jais, Mistral, and GPT-4o, to mimic human emotional expression in English and personality markers in Arabic. The researchers evaluated whether machine classifiers could distinguish between human-authored and AI-generated texts and assessed the emotional/personality traits exhibited by the LLMs. Results indicate that AI-generated texts are distinguishable from human-authored ones, with classification performance impacted by paraphrasing, and that LLMs encode affective signals differently than humans. Why it matters: The findings have implications for authorship attribution, affective computing, and the responsible deployment of AI, especially in under-resourced languages like Arabic.