31 May 2016

Opinion: The ugliest side of facial recognition technology

By Evan Selinger, Columnist Woodrow Hartzog, Contributor 
May 27, 2016 

The emergence of technologies that falsely promise to predict someone's behavior based on their facial features and expressions is a deeply troubling development.

It's no mystery that big data presents a challenge to privacy. But perhaps more alarming is the emergence of technology that combines facial recognition and data analytics to create a powerful surveillance tool.

It's a disturbing development that combines the most worrisome aspects of algorithmic and big data technology with the chilling and dangerous threats inherent in facial recognition.

A Chicago tech company is advertising its "predictive video" to anticipate behavior "based on the emotional state and personality style of any person in a video." In Russia, the app FindFace gives users "the power to identify total strangers on the street," according to The Atlantic.

It's not just the tech fringe, either. Google's new chat app Allo has a "smart reply" feature that apparently analyzes photos from contacts and offers​ suggested​ responses to them​.

But most troubling is the Israeli startup Faception. It offers a product that combines machine learning with facial recognition to "identify everything from great poker players to extroverts, pedophiles, geniuses, and white collar criminals." A Department of Homeland Security contractor has hired the firm to "help identify terrorists."

That's a problem. The government should not use people's faces as a way of tagging them with life-altering labels. The technology isn't even accurate. Faception’s own estimate for certain traits is a 20 percent error rate. Even if those optimistic numbers hold, that means for every 100 people, the best-case scenario is that 20 get wrongly branded as a terrorist.

And yet, according to the company, powerful profiling is possible due to two alleged facts: personalities are "affected by genes" and our faces are a "reflection of our DNA."

Get Monitor cybersecurity news and analysis delivered straight to your inbox.

Follow Passcode

The first premise doesn’t inspire confidence. It presumes nature affects our personalities more than nurture – a conclusion that experts constantly debate. For the sake of argument, though, let’s say this is true. Even then, we’re not dealing with a robust causal claim. Saying that personalities are "affected" by genes is a much weaker assertion than maintaining our genes determine them.

As to the face being a reflection of DNA, the folks at Faception admit that the evidence that supports that claim comes from animal studies, not psychological inquiry into human beings. Conveniently, they dismiss differences by accepting an unnamed researcher's conjectural claim that "the human face was likely to develop in the same way."

These assumptions completely ignore many established psychological theories such as "situationism," which is when environmental features dispose all of us to behave in new ways – ways that can lead a person who habitually does good things to commit evil and atrocious acts.

Physical images certainly have some revelatory power. A snapshot of body language, for instance, can reveal confidence or nervousness. But it would be serious mistake to view the face alone as a portal into deep character traits and future behavior.

Advocates of this kind of data analysis might argue that algorithms will get better as the data science advances and computers can make more decisions more quickly. But false correlations are already plaguing big data. Economist Ronald Coase famously told us years ago that if you torture the data long enough, it will confess anything.

But unfortunately people tend to place far too much confidence in anything a computer spits out. This phenomenon known as "automation bias" looms large with the implementation of predictive facial recognition technologies. A prime example of this was documented in a recent Pro Publica investigation that exposed racial bias embedded in predictive criminal risk assessments software.

Even if technology can guess correctly, do we really want to live in a society in which machines try to suss out deep truths based on our facial features? If that were the case, our "faceprints" would serve as beacons for unwanted attention, threatening our obscurity – the idea that when information about us is hard to find, it's safer. And you can't ditch the beacon unless you want to wear a mask in public for the rest of your life.

Our faces are indeed exceptional. But predictive facial recognition technology and companies like Faception exacerbate the most dangerous aspects of both big data and facial recognition.

Evan Selinger is a professor of philosophy at Rochester Institute of Technology. Follow him on Twitter @EvanSelinger.

Woodrow Hartzog is ​the ​​W. Stancil Starnes Professor of Law ​at S

No comments: