Sunday 23 March 2025
A new study has shed light on the ability of large language models (LLMs) to understand and interpret facial expressions, a crucial aspect of human communication. Researchers have long sought to develop machines that can accurately recognize emotions, and LLMs have been touted as a promising solution.
The study used a dataset of images featuring various facial expressions, each assigned a score on the valence-arousal (VA) scale, which measures the intensity of emotional responses. The researchers then fed this data into three different LLMs – BERT, Word2Vec and Transformer – to see how well they could infer emotions from the VA values.
The results were mixed. While the LLMs performed reasonably well in generating semantic descriptions of facial expressions, they struggled when it came to categorizing emotions into discrete categories. This was particularly true for more complex emotions, such as anger or surprise.
One of the key findings was that the LLMs’ performance varied depending on their underlying architecture and training data. For example, BERT – which is designed specifically for language tasks – performed better than Word2Vec and Transformer when it came to generating descriptive text about facial expressions. However, the latter two models were more effective at identifying patterns in the VA scores.
The study’s authors suggest that LLMs are better suited to capturing general affective meanings from facial expressions rather than making precise categorical judgments. This is because human emotions are often nuanced and context-dependent, making it difficult for machines to accurately categorize them.
The findings have significant implications for the development of artificial intelligence (AI) systems that interact with humans. Currently, many AI-powered chatbots and virtual assistants rely on simple emotional recognition algorithms, which can lead to misunderstandings and miscommunications.
By acknowledging the limitations of LLMs in interpreting facial expressions, researchers can work towards developing more sophisticated AI systems that better understand human emotions. This could involve integrating multimodal data – such as audio and video cues – into AI models or using alternative approaches, like machine learning algorithms specifically designed for emotion recognition.
Ultimately, the study highlights the importance of understanding the strengths and weaknesses of LLMs in order to create more effective and empathetic AI systems. By acknowledging the limitations of these models, researchers can work towards developing machines that better understand and respond to human emotions – a crucial aspect of building trust and rapport between humans and machines.
Cite this article: “Large Language Models Struggle to Accurately Interpret Facial Expressions”, The Science Archive, 2025.
Large Language Models, Facial Expressions, Emotion Recognition, Artificial Intelligence, Machine Learning, Natural Language Processing, Sentiment Analysis, Affective Computing, Human-Computer Interaction, Deep Learning.







