In science fiction003 Archives facial recognition technology is a hallmark of a dystopian society. The truth of how it was created, and how it's used today, is just as freaky.
In a new study, researchers conduct a historical survey of over 100 data sets used to train facial recognition systems compiled over the last 43 years. The broadest revelation is that, as the need for more data (i.e. photos) increased, researchers stopped bothering to ask for the consent of the people in the photos they used as data.
Researchers Deborah Raji of Mozilla and Genevieve Fried of AI Now published the study on Cornell University's free distribution service, arXiv.org. The MIT Technology Review published its analysis of the paper Friday, describing it as "the largest ever study of facial-recognition data" that "shows how much the rise of deep learning has fueled a loss of privacy."
Within the study's charting of the evolution of facial recognition datasets, there are moments in history and facts about this technology's development that are revealing. They show how the nature of facial recognition is that it's a flawed technology when applied to real-world scenarios, created with the express purpose of expanding the surveillance state, with the effect of degrading our privacy.
Here are 9 scary and surprising takeaways from 43 years of facial recognition research.
One of the reasons the researchers give for undertaking their study is to understand why facial recognition systems that perform at near 100 percent accuracy in testing are deeply flawed when they're applied in the real world. For example, they say, New York City's MTA halted a facial recognition pilot after it had a 100 percent error rate. Facial recognition, which has been proven to be less accurate on black and brown faces, recently led to the arrest of three Black men in who were incorrectly identified by the tech.
Though efforts to develop facial recognition began in academic settings, it took off in 1996 when the DoD and National Institute of Standards and Technology (NIST) allocated $6.5 million to create the largest dataset to date. The government got interested in this area because of its potential for surveillance that did not require people to actively participate, unlike fingerprinting.
It seems almost quaint, but before the mid-2000s, the way researchers amassed databases was by having people sit for portrait settings. Because some of the foundational facial recognition tech today came from these datasets, the flaws of the portrait technique resonate. Namely, a non-diverse set of participants, and staged settings that don't accurately reflect real-world conditions.
Yep, when researchers wanted to expand datasets beyond portraits, this is literally what happened. A 2007 dataset called Labeled Faces in the Wild scraped Google, Flickr, YouTube, and other online repositories of photos. That included photos of children. While this led to a greater variety of photos, it also discarded the privacy rights of the subjects.
"In exchange for more realistic and diverse datasets, there was also a loss of control, as it became unmanageable to obtain subject consent, record demographic distributions, maintain dataset quality and standardize attributes such as image resolution across Internet-sourced datasets," the paper reads.
The researchers cite a turning point in facial recognition when Facebook revealed the creation of its DeepFace database in 2014. Facebook showed how the collection of millions of photos could create neural networks that were far better at facial recognition tasks than previous systems, making deep learning a cornerstone of modern facial recognition.
Facebook has since been fined by the FTC and paid a settlement to the state of Illinois for using the photos users uploaded to Facebook to enable its facial recognition without getting users' affirmative consent. The way DeepFace manifested was through "Tag Suggestions," a feature that was able to suggest the person in your photo you might want to tag. Accepting or rejecting tags in turn made Facebook's systems smarter. Tag Suggestions were opt-out, which meant participating in this technology was the default.
In reality, we don't know the number or identity of people whose photos made them unwitting participants in the development of facial recognition tech.
Facial recognition systems have evolved beyond identifying a face or a person. They can also label people and their attributes in offensive ways.
"These labels include the problematic and potentially insulting labels regarding size - 'chubby', 'double chin' - or inappropriate racial characteristics such as 'Pale skin,' 'Pointy nose,' 'Narrow eyes' for Asian subjects and 'Big nose' and 'Big lips' for many Black subjects," the paper reads. "Additionally there is the bizarre inclusion of concepts, such as 'bags under eyes,' '5 o’clock shadow' and objectively impossible labels to consistently define, such as 'attractive.'"
Faces considered "western" became the default in training sets. And other datasets expressly created to increase diversity were problematic themselves: One such system's purpose was to “train unbiased and discrimination-aware face recognition algorithms,” but the researchers point out it only "divide[d] human ethnic origins into only three categories."
These faults go beyond just being offensive. Research has shown that discrimination in AI can reinforce discrimination in the real-world.
Facial recognition has both stayed true to its roots and expanded beyond what its creators in the 1970s could possibly imagine.
"We can see from the historical context that the government promoted and supported this technology from the start for the purpose of enabling criminal investigation and surveillance," the authors write. For example, Amazon has already sold its problematic Rekognition tech to an untold number of police departments.
On the other end of the spectrum, some training sets promise that it can help develop systems to analyze sentiment of shoppers and better track and understand potential customers.
Which is more dystopian: The surveillance state or an all-knowing capitalist advertising machine? You decide.
Topics Artificial Intelligence Cybersecurity Facebook Facial Recognition Privacy
New 'Dunkirk' footage is a heartThe Galaxy S8's iris scanner fixes the one thing that made it suck on the Note 7Best free speaker deal: Free Soundcore Select 4 Go speaker with AeroFit earbudsFirst he became a rock, now French artist wants to be a chickenBest Anker deal: Save $16 on Anker 6Kid won't take no for an answer: Builds a Nintendo Switch out of cardboardApparently the average person takes 25,000 selfies in a lifetime. Seems about right.SelfScore makes credit cards for the international students Trump wants to keep outHow a year of utterly insane sports finishes lit up TwitterA simple reason why you should stay out of floodwaters in Australia: SharksKid won't take no for an answer: Builds a Nintendo Switch out of cardboardMeet Black Lightning, DC and CW's first AfricanThree tornado chasers were killed Tuesday night, raising new safety questionsSelfScore makes credit cards for the international students Trump wants to keep outUnderrated horror MMO 'The Secret World' is going freeEd Sheeran reacts to baby that looks exactly like Ed SheeranThe U.S. version of the 'Love Actually' sequel will have one huge differenceTesla's Gigafactory looks like a town in the middle of the Nevada desertFirst he became a rock, now French artist wants to be a chicken10 YA books we want to see adapted as TV shows 'Fight Crab' is a bizarre, joyous experiment in crustacean combat Boys told not to wear shorts to school, so they showed up in skirts When this woman received an unsolicited dick pic, she took it straight to the dude's mom Kelly Osbourne peed her pants and is blaming Starbucks Facebook boasts that ad boycott won't make a dent Here's how to watch the NBA playoffs restart without cable Listen up everyone, this new, beautiful gorilla meme has something to say Twitter finally bans white supremacist David Duke World's Ugliest Dog 2017: Meet the floppy winner Banksy's identity 'revealed' by DJ Goldie Wearables are just fine, market expected to double by 2021 Man discovers 36 A sleepy Bruno Mars made an appearance at the BET awards The Queen reported to police for not wearing seat belt Justice Dept. charged three people in connection with the Twitter hack Gorilla dancing to 'Flashdance' has some serious moves Angry cow crashes cricket match, probably just mad it couldn't play 'How do I get rid of that annoying echo on my Zoom calls?' and more WFH help Woman can make her massive dog instantly less intimidating and it involves menswear Facebook CEO corrects confused congressman over Trump Jr. Twitter ban
2.9154s , 8254.3359375 kb
Copyright © 2025 Powered by 【2003 Archives】,Charm Information Network