Researchers created a cross-cultural corpus of annotated verbal and nonverbal behaviors in receptionist interactions. The corpus includes native speakers of American English and Arabic role-playing scenarios at university reception desks in Doha, Qatar, and Pittsburgh, USA. The manually annotated nonverbal behaviors include gaze direction, hand gestures, torso positions, and facial expressions. Why it matters: This resource can be valuable for the human-robot interaction community, especially for building culturally aware AI systems.
This article previews a talk by Gül Varol from Ecole des Ponts ParisTech on bridging natural language and 3D human motions. The talk will cover text-to-motion synthesis using generative models and text-to-motion retrieval models based on the ACTOR, TEMOS, TMR, TEACH, and SINC papers. Varol's research interests include video representation learning, human motion synthesis, and sign languages. Why it matters: Research in this area could enable more intuitive human-computer interaction and new applications in areas like virtual reality and robotics.
This research introduces a novel method using the Lateral Accretive Hybrid Network (LEARNet) to capture and analyze micro-expressions for mental health applications. The method refines both broad and subtle facial cues to detect mental health conditions like anxiety or depression. The authors also propose a neural architecture search (NAS) strategy to design a compact CNN for micro-expression recognition, improving performance and resource use. Why it matters: By integrating micro-emotion recognition with mental health estimation, the approach enables more accurate and early detection of emotional and mental health issues, potentially leading to improved well-being.
Christian Montag from Ulm University gave a talk about assessing attitudes towards AI, covering the IMPACT framework (Modality, Person, Area, Country/Culture, and Transparency). He discussed how factors like age, gender, personality, and culture relate to attitudes toward AI, and how those attitudes link to trust in automation and specific AI models like ChatGPT and Ernie Bot. Montag's research explores the intersection of psychology, neuroscience, behavioral economics, and computer science, focusing on the impact of AI on the human mind. Why it matters: Understanding public perception of AI is crucial for responsible development and deployment, especially in the Arab world where cultural and demographic factors can significantly shape attitudes.
A presentation discusses the evolution of Vision-and-Language Navigation (VLN) from benchmarks like Room-to-Room (R2R). It highlights the role of Large Language Models (LLMs) such as GPT-4 in enabling more natural human-machine interactions. The presentation showcases work using LLMs to decode navigational instructions and improve robotic navigation. Why it matters: This research demonstrates the potential of merging vision, language, and robotics for advanced AI applications in navigation and human-computer interaction.
Ekaterina Vylomova from the University of Melbourne gave a talk on using NLP models to advance research in linguistic morphology, typology, and social psychology. The talk covered using models to study morphology, phonetic changes in words over time, and diachronic changes in language semantics. Vylomova presented the UniMorph project, a cross-lingual annotation schema and database with morphological paradigms for over 150 languages. Why it matters: This research demonstrates the potential of NLP to contribute to a deeper understanding of language evolution and structure, with applications in linguistic research and the study of social and cultural changes.
A public talk announcement features Professor Anil K. Jain from Michigan State University discussing biometric recognition. The talk will cover automated recognition of individuals based on biological and behavioral traits. It will also address challenges, research opportunities, and ongoing projects in Jain's lab related to biometrics. Why it matters: As biometric technologies become increasingly integrated into daily life across the Middle East, understanding their limitations and ethical implications is crucial for responsible development and deployment.
The paper introduces MIRAGE, a framework for evaluating LLMs' ability to simulate human behaviors in murder mystery games. MIRAGE uses four methods: TII, CIC, ICI and SCI to assess the LLMs' role-playing proficiency. Experiments show that even GPT-4 struggles with the complexities of the MIRAGE framework.