ICO warns against emotional analysis using biometrics

26/10/2022 | ICO

The Information Commissioner's Office (ICO) has warned that it will investigate organisations using emotion analysis technologies irresponsibly, urging risk assessments before implementing systems. The ICO said emotion analysis technologies rely on "collecting, storing and processing a range of personal data, including subconscious behavioural or emotional responses, and in some cases, special category data." This kind of data use is "far more risky than traditional biometric technologies" and brings a risk of systemic bias, inaccuracy and discrimination. Deputy Commissioner Stephen Bonner (CIPP/E, CIPM) said, "Developments in the biometrics and emotion AI market are immature. They may not work yet, or indeed ever. While there are opportunities present, the risks are currently greater." 

Additional commentary in The Guardian.

Read Full Story
Emotion Analysis, biometrics

What is this page?

You are reading a summary article on the Privacy Newsfeed, a free resource for DPOs and other professionals with privacy or data protection responsibilities helping them stay informed of industry news all in one place. The information here is a brief snippet relating to a single piece of original content or several articles about a common topic or thread. The main contributor is listed in the top left-hand corner, just beneath the article title.

The Privacy Newsfeed monitors over 300 global publications, of which more than 4,350 summary articles have been posted to the online archive dating back to the beginning of 2020. A weekly roundup is available by email every Friday.

Freevacy has been shortlisted in the Best Educator category.
The PICCASO Privacy Awards recognise the people making an outstanding contribution to this dynamic and fast-growing sector.