Emotion AI aims to help bots understand human emotions, but scepticism and regulatory hurdles challenge its effectiveness
As businesses rush to embed artificial intelligence into every corner of their operations, companies are turning to AI to help their bots better understand human emotions.
This area, known as "emotion AI," is on the rise, according to PitchBook’s latest Enterprise SaaS Emerging Tech Research report.
The premise is simple: if businesses are going to deploy AI assistants as front-line customer service reps or salespeople, these bots must understand the difference between an excited "Can you tell me more about that?" and a skeptical "Can you tell me more about that?"
Seems an impossibility, right?
Regardless, emotion AI is portrayed as an advanced successor to sentiment analysis, which has traditionally sought to extract emotional content from written communications, especially on social platforms.
While sentiment analysis primarily examines text, emotion AI adopts a more comprehensive strategy, utilising visual cues, audio signals, and additional sensory data to identify human emotional states.
The adoption of emotion AI raises concerns about user acceptance. Users may hesitate to grant AI access to cameras and microphones, viewing it as a privacy invasion.
Psychologically, many might feel uneasy about AI decoding their emotions, seeing it as intrusive. Balancing potential benefits with these privacy and psychological concerns will be crucial for emotion AI's ethical implementation and public acceptance.
Are folks ready for sentiment analysis on steroids?
Major AI cloud providers have already hopped on the emotion AI bandwagon. Microsoft Azure’s Emotion API and Amazon’s Rekognition service are just two examples.
According to Pitchbook, the sudden rise of AI bots in the workforce, suggests that emotion AI’s future in business is more assured than ever.
As Derek Hernandez, a senior analyst in emerging technology at PitchBook, points out, “With the proliferation of AI assistants and fully automated human-machine interactions, emotion AI promises to enable more human-like interpretations and responses.”
Startups are, of course, racing to capitalise on this trend. But for all the promise, is this technology really the answer to creating better human-machine interactions?
Here’s where the optimism meets a wall of skepticism. In 2019, researchers published a meta-review that questioned the very premise of emotion AI.
Their conclusion was blunt: human emotion cannot be reliably determined by facial movements alone. This revelation throws a wrench into the idea that AI can be trained to detect human emotions by mimicking the way humans interpret facial expressions, body language, and tone of voice.
As Paul Ekman, American psychologist and professor said, "Emotions are not simply linear phenomena but complex states of mind and body, involving multiple components - physiological, cognitive, and behavioural - that are difficult to completely separate."
The risk with emotion AI is that it could oversimplify these nuances, leading to misinterpretations that could cause more harm than good.
Imagine a customer service bot misjudging a frustrated customer as angry and responding defensively—hardly the recipe for stellar customer experience. Or worse still, a therapy bot that misses the mark entirely.
Adding to the skepticism is the regulatory landscape that is starting to take shape.
The European Union’s AI Act, which aims to regulate the use of AI technologies, explicitly bans computer-vision emotion detection for certain applications, like education. In the U.S., laws such as Illinois’ Biometric Information Privacy Act (BIPA) prohibit the collection of biometric data without explicit consent.
These regulations reflect growing concerns about privacy and ethical use of AI, which could significantly limit the deployment of emotion AI in business settings.
So, what next?
On one hand, the rise of emotion AI could usher in a future where AI interactions feel more natural, more human. Bots that can read and respond to our emotions could revolutionise customer service, sales, and even internal communications.
On the other hand, there’s a real risk that these AI bots could become little more than high-tech gimmicks—machines that feign understanding without truly grasping the complexities of human emotion.
Ultimately, the success of emotion AI will hinge on its ability to genuinely enhance human-machine interactions without overstepping ethical boundaries.
In the meantime, we might just have to put up with bots that misunderstand us, at least until the tech catches up with the promise. After all, who hasn’t had their fair share of misunderstandings with a chatbot that’s as empathetic as a brick wall?
Whether emotion AI can break that wall remains to be seen.
Emotion AI seeks to enhance AI-human interactions by interpreting emotions through visual, audio, and other sensory inputs for better responses
Scepticism exists because human emotions are complex and context-dependent, raising doubts about AI's ability to interpret them accurately
Regulatory hurdles, like the EU’s AI Act, aim to restrict emotion detection technology, reflecting concerns about privacy and ethical use