Scientists Create Emojify To Show The Risks Of AI Emotion Recognition

Scientists Create Emojify To Show The Risks Of AI Emotion Recognition

Emojify to help people to understand how computers can be used to scan facial expressions to detect emotion.

It is a technology that has been frowned upon by ethicists: now researchers are hoping to unmask the reality of emotion recognition systems in an effort to boost public debate.

Technology designed to identify human emotions using machine learning algorithms is a huge industry, with claims it could prove valuable in myriad situations, from road safety to market research. But critics say the technology not only raises privacy concerns but is inaccurate and racially biased.

A team of researchers have created a website – – where the public can try out emotion recognition systems through their own computer cameras. One game focuses on pulling faces to trick the technology, while another explores how such systems can struggle to read facial expressions in context.

Their hope, the researchers say, is to raise awareness of the technology and promote conversations about its use.

‘It is a form of facial recognition, but it goes farther because rather than just identifying people, it claims to read our emotions, our inner feelings from our faces’, said Dr Alexa Hagerty, project lead and researcher at the University of Cambridge Leverhulme Centre for the Future of Intelligence and the Centre for the Study of Existential Risk.

Facial recognition technology often used to identify people, has come under intense scrutiny in recent years. Last year the Equality and Human Rights Commission said its use for mass screening should be halted, saying it could increase police discrimination and harm freedom of expression.

But Hagerty said many people were not aware of how common emotion recognition systems were, noting they were employed in situations ranging from job hiring to customer insight work, airport security, and even education to see if students are engaged or doing their homework.

Such technology, she said, was in use all over the world, from Europe to the US and China. Taigusys, a company that specialises in emotion recognition systems and whose main office is in Shenzhen, says it has used them in settings ranging from care homes to prisons, while according to reports earlier this year, the Indian city of Lucknow is planning to use the technology to spot distress in women as a result of harassment – a move that has met with criticism, including from digital rights organisations.

While Hagerty said emotion recognition technology might have some potential benefits these must be weighed against concerns around accuracy, racial bias, as well as whether the technology was even the right tool for a particular job.

The developers, who are developing this technology, claim that it is for reading emotions but in reality, the system is reading facial movement and then combining the emotions with the assumption. And those movements are linked to emotions. 

‘There is lots of really solid science that says that is too simple; it doesn’t work quite like that’, said Hagerty, adding that even just human experience showed it was possible to fake a smile. ‘That is what that game was: to show you didn’t change your inner state of feeling rapidly six times, you just changed the way you looked [on your] face,’ she said.