DeepFace

DeepFace is a deep learning facial recognition system created by a research group at Facebook. It identifies human faces in digital images. The program employs a nine-layer neural network with over 120 million connection weights and was trained on four million images uploaded by Facebook users.[1][2] The Facebook Research team has stated that the DeepFace method reaches an accuracy of 97.35% ± 0.25% on Labeled Faces in the Wild (LFW) data set where human beings have 97.53%.[3] This means that DeepFace is sometimes more successful than human beings. As a result of growing societal concerns Meta announced[4] that it plans to shut down Facebook facial recognition system, deleting the face scan data of more than one billion users.[5] This change will represent one of the largest shifts in facial recognition usage in the technology's history. Facebook planned to delete by December 2021 more than one billion facial recognition templates, which are digital scans of facial features. However, it did not plan to eliminate DeepFace which is the software that powers the facial recognition system. The company has also not ruled out incorporating facial recognition technology into future products, according to Meta spokesperson.[5]

Commercial rollout

[edit]

Origin

[edit]

DeepFace was produced by a collection of scientists from Facebook's artificial intelligence research team. The team includes Yainiv Taigman and a Facebook research scientist Ming Yang. They were also joined by Lior Wolf, a faculty member from Tel Aviv University. Yaniv Taigman, came to Facebook when Facebook acquired Face.com in 2012.

Facebook started rolling out DeepFace to its users in early 2015, and have continuously expanding DeepFace's use and software,.[6] DeepFace, according to the director of Facebook's artificial intelligence research, is not intended to invade individual privacy. Instead, DeepFace alerts individuals when their face appears in any photo posted on Facebook. When they receive this notification, they have the option of removing their face from the photo.[6]

European Union

[edit]

When the DeepFace technology was initially deployed, users had the option to turn DeepFace off. However, they were not notified that it was on.[7] Because of this, DeepFace was not released in the European Union. A data privacy law in the EU argued that Facebook's facial recognition did not comply with EU data protection laws. Because users do not consent to all the uses of their biometric data, it does not comply.[8]

Accuracy

[edit]

DeepFace systems can identify faces with 97% accuracy, almost the same accuracy as a human in the same position. Facebook's facial recognition is more effective than the FBI's technology, which has 85% accuracy.[9] Google's technology, FaceNet is more successful than DeepFace using the same data sets. FaceNet set a record for accuracy, 99.63%. Google's FaceNet incorporates data from Google Photos.[10]

Applications

[edit]

Facebook uses individual facial recognition templates to find photos that an individual is in so they can review, engage, or share the content. DeepFace protects individuals from impersonation or identity theft. Take, for example, an instance where an individual used someone's profile photo as their own. Through DeepFace,  Facebook can identify and alert the person whose information is being misused.[11] To ensure that individuals have control over their facial recognition, Facebook does not share facial templates. Additionally, Facebook will remove images from facial recognition templates if someone has deleted their account or untagged themself from a photo. Individuals also have the ability to turn their facial recognition off on Facebook. If the feature is turned off, Facebook will cease facial recognition for that individual.  

Following the release of DeepFace in 2015, its uses have remained fairly stagnant. Because more individuals have uploaded images to Facebook, the algorithm has gotten more accurate. Facebook's DeepFace is the largest facial recognition dataset that currently exists. Because of this, some individuals argue that Facebook's facial ID database could be distributed to government agencies.[12] These uses, however, would be prohibited by most data privacy laws. In response to privacy concerns, Facebook removed their automatic facial recognition feature – allowing individuals to opt in to tagging through DeepFace. This change was implemented in 2019.

Architecture

[edit]

The DeepFace system consists of four modules: 2D alignment, 3D alignment, frontalization, and neural network. An image of a face is passed through them in sequence, resulting in a 4096-dimensional feature vector representing the face. The feature vector can then be further processed for many different tasks. For example, to identify the face, one can compare it against a list of feature vectors of known faces, and identify the face with the most similar feature vector.

DeepFace uses fiducial point detectors based on existing databases to direct the alignment of faces. The facial alignment begins with a 2D alignment, and then continues with 3D alignment and frontalization. That is, DeepFace's process is two steps. First, it corrects the angles of an image so that the face in the photo is looking forward. To accomplish this, it uses a 3-D model of a face.[13]

2D alignment

[edit]

The 2D alignment module detects 6 fiducial points on the detected face — the center of the eyes, tip of the nose and mouth location. These points are translated onto a warped image to help detect the face. However, 2D transformation fails to compensate for rotations that are out of place.

3D alignment

[edit]

In order to align faces, DeepFace uses a generic 3D model wherein 2D images are cropped as 3D versions. The 3D image has 67 fiducial points. After the image has been warped, there are 67 anchor points manually placed on the image to match the 67 fiducial points. A 3D-to-2D camera is then fitted that minimizes losses. Because 3D detected points on the contour of the face can be inaccurate, this step is important.

Frontalization

[edit]

Because full perspective projections are not modeled, the fitted camera is only an approximation of the individual's actual face. To reduce errors, DeepFace aims to warp the 2D images with smaller distortions. Also, thee camera P is capable of replacing parts of the image and blending them with their symmetrical counterparts.

Neural network

[edit]

The neural network is a sequence of layers, arranged as follows: convolutional layer - max pooling - convolutional layer - 3 locally connected layers - fully connected layer.

The input is an RGB image of the face, scaled to resolution , and the output is a real vector of dimension 4096, being the feature vector of the face image.

In the 2014 paper,[13] an additional fully connected layer is added at the end to classify the face image into one of 4030 possible persons that the network had seen during training time.

Reactions

[edit]

Industry

[edit]

AI researcher Ben Goertzel said Facebook had "pretty convincingly solved face recognition" with the project, but said it would be incorrect to conclude that deep learning is the entire solution to AI.

Neeraj Kumar, a researcher at the University of Washington said that Facebook's DeepFace shows how large sets of outside data can result in a "higher capacity" model. Because of Facebook's wide access to images of individuals, their facial recognition software can perform better than other software with much smaller data sets.[14][15]

Media

[edit]

A Huffington Post piece called the technology "creepy", citing data privacy concerns, noted that some European governments had already required Facebook to delete facial-recognition data.[16] According to Broadcasting & Cable, both Facebook and Google had been invited by the Center for Digital Democracy to attend a 2014 National Telecommunications and Information Administration "stakeholder meeting" to help develop a consumer privacy Bill of Rights, but they both declined. Broadcasting & Cable also noted that Facebook had not released any press announcements concerning DeepFace, although their research paper had been published earlier in the month. Slate said that DeepFace was not being publicized by Facebook because it is wary of another round of headlines decrying DeepFace's creepiness.

Users

[edit]

Many individuals fear facial recognition technology.[17][18] The technology's nearly perfect accuracy allows social media companies to create digital profiles of millions of Americans.[19] However, an individual's fear of facial recognition and other privacy concerns does not correspond to a decrease in social media use. Instead, attitudes towards privacy and privacy settings do not have a large impact on an individual's intention to use Facebook apps.[20][21][22] Because Facebook is a social media site, individual fears about privacy get over ruled by a desire to participate in social media.[23]

Privacy concerns

[edit]

BIPA lawsuit

[edit]

Facebook users raised a class action lawsuit against Facebook under Illinois Biometric Information Privacy Act (BIPA).[24] Illinois has the most comprehensive biometric privacy legislation, regulating the collection of biometric information by commercial entities.[25] Illinois' BIPA requires a corporation that obtains a person's biometric information to obtain a written release, provide them notice that their information is being collected, and state the duration the information will be collected. The lawsuit raised against DeepFace alleges that Facebook's collection of facial identification information for the purpose of the tag suggestion tool violates BIPA.[26] Because Facebook does not give notice or consent to individuals when they use this tool, Facebook users argue that it violates BIPA.[27]  The Ninth Circuit denied Facebook's motion to dismiss the case and ultimately certified the case. Facebook sought to appeal to the certification of the Ninth Circuit decision which was ultimately granted. Facebook claims that the case should not have been verified because Plaintiffs have no alleged any harm beyond Facebook's violation of BIPA. Facebook removed their automatic facial recognition tagging feature in 2019, in response to the concerns raised in the lawsuit.[28] Facebook proposed a $550 million settlement to the case, which was rejected. When Facebook increased the settlement to $650 million, the court accepted it. Facebook was ordered to pay their $650 million  settlement in early March 2021. 1.6 million residents of Illinois will receive at least $345.[29]

In July 2020, Facebook announced that it is building teams that will look into racism in its algorithms.[30] Facebook's teams will work with Facebook's Responsible AI team to study bias in their systems. The implementation of these programs is recent, and it is still unclear what reforms will be made.[31]

Ten-year challenge

[edit]

In 2019, a Facebook challenge went viral asking users to post a photo from 10 years ago and one from 2019. The challenge was coined the "10 Year challenge." More than 5 million people participated in the challenge, including many celebrities. Worry arose that Facebook's 10 year challenge was designed to train Facebook's facial recognition database. Kate O'Neill, a writer for Wired, wrote an op-ed that echoed this possibility.[32] Facebook denied that they played a role in generating the challenge.[33] However, individuals have argued that the concerns that underscore theories around the 10 year challenge are echoed by broader concerns about Facebook and the right to privacy.[34]

Racism in facial identification technology

[edit]

Facial recognition algorithms are not universally successful.[35] While the algorithms are capable of classifying faces with over 90% accuracy in some cases, accuracy is lower when the algorithms are applied to women, black individuals, and young people.[36] The systems falsely identify black and Asian faces 10 to 100 times more than they do with white faces.[37] Because algorithms are primarily trained with white men, systems like DeepFace have a more difficult time identifying them.[38] It is projected that once facial recognition data bases are trained to identify people of color — exposing them to more diverse faces — they will be more successful at identification.[39]

See also

[edit]

References

[edit]
  1. ^ "Facebook creates software that matches faces almost as well as you do", Technology Review, Massachusetts Institute of Technology, March 17, 2014
  2. ^ "Facebook's DeepFace shows serious facial recognition skills", CBS News, March 19, 2014
  3. ^ "DeepFace: Closing the Gap to Human-Level Performance in Face Verification". Facebook Research. Retrieved 2019-07-25.
  4. ^ Metz, Rachel (2 November 2021). "Facebook is shutting down its facial recognition software". CNN. Retrieved 2021-11-05.
  5. ^ a b Hill, Kashmir; Mac, Ryan (2021-11-02). "Facebook, Citing Societal Concerns, Plans to Shut Down Facial Recognition System". The New York Times. ISSN 0362-4331. Retrieved 2021-11-05.
  6. ^ a b Chowdhry, Amit. "Facebook's DeepFace Software Can Match Faces With 97.25% Accuracy". Forbes. Retrieved 2021-04-09.
  7. ^ "Facebook settles facial recognition dispute". BBC News. 2020-01-30. Retrieved 2021-04-08.
  8. ^ "Vol 23.1 – Winter 2017 | Journal of Science & Technology Law". www.bu.edu. Retrieved 2021-04-24.
  9. ^ "Facial Recognition Technology: Ensuring Transparency in Government Use — FBI". www.fbi.gov. Retrieved 2021-04-09.
  10. ^ "Facial recognition: top 7 trends (tech, vendors, markets, use cases & latest news)". Thales Group. Retrieved 2021-04-09.
  11. ^ "What is the face recognition setting on Facebook and how does it work? | Facebook Help Center". www.facebook.com. Retrieved 2021-04-22.
  12. ^ Glaser, April (2019-07-09). "Facebook's Face-ID Database Could Be the Biggest in the World. Yes, It Should Worry Us". Slate Magazine. Retrieved 2021-04-22.
  13. ^ a b Taigman, Yaniv; Yang, Ming; Ranzato, Marc'Aurelio; Wolf, Lior (June 2014). "DeepFace: Closing the Gap to Human-Level Performance in Face Verification". 2014 IEEE Conference on Computer Vision and Pattern Recognition. IEEE. pp. 1701–1708. doi:10.1109/cvpr.2014.220. ISBN 978-1-4799-5118-5. S2CID 2814088.
  14. ^ "Facebook Creates Software That Matches Faces Almost as Well as You Do". MIT Technology Review. Retrieved 2021-04-22.
  15. ^ Rubinstein, Ira; Good, Nathan (2012). "Privacy by Design: A Counterfactual Analysis of Google and Facebook Privacy Incidents". SSRN Electronic Journal. doi:10.2139/ssrn.2128146. ISSN 1556-5068.
  16. ^ Grandoni, Dino (2014-03-18). "Facebook's New 'DeepFace' Program Is Just As Creepy As It Sounds". HuffPost. Retrieved 2021-04-22.
  17. ^ "Privacy and identity on Facebook", Discourse and Identity on Facebook, Bloomsbury Academic, 2017, doi:10.5040/9781474289153.0014, ISBN 978-1-4742-8912-2, retrieved 2021-04-24
  18. ^ Barrett, Lindsey (2020-07-24). "Ban Facial Recognition Technologies for Children—And for Everyone Else". Rochester, NY. SSRN 3660118. {{cite journal}}: Cite journal requires |journal= (help)
  19. ^ Huang, Michelle Yan. "Facial recognition is almost perfectly accurate — here's why that could be a problem". Business Insider. Retrieved 2021-04-22.
  20. ^ Van Der Schyff, Karl; Flowerday, Stephen; Lowry, Paul Benjamin (2020-08-01). "Information privacy behavior in the use of Facebook apps: A personality-based vulnerability assessment". Heliyon. 6 (8): e04714. Bibcode:2020Heliy...604714V. doi:10.1016/j.heliyon.2020.e04714. ISSN 2405-8440. PMC 7452521. PMID 32904276.
  21. ^ Mathiyalakan, Sathasivam; Heilman, George; Ho, Kevin; Law, Wai (2018-01-01). "An Examination of the Impact of Gender and Culture on Facebook Privacy and Trust in Guam". Journal of International Technology and Information Management. 27 (1): 29–56. doi:10.58729/1941-6679.1363. ISSN 1941-6679. S2CID 159011924.
  22. ^ "Facebook face recognition hits privacy protests". Biometric Technology Today. 2011 (7): 1. July 2011. doi:10.1016/s0969-4765(11)70120-5. ISSN 0969-4765.
  23. ^ Rosenthal, Sonny; Wasenden, Ole-Christian; Gronnevet, Gorm-Andreas; Ling, Rich (2020-11-01). "A tripartite model of trust in Facebook: acceptance of information personalization, privacy concern, and privacy literacy". Media Psychology. 23 (6): 840–864. doi:10.1080/15213269.2019.1648218. hdl:10356/145658. ISSN 1521-3269. S2CID 201372342.
  24. ^ "Power, Pervasiveness and Potential: The Brave New World of Facial Recognition Through a Criminal Law Lens (and Beyond)". nycbar.org. Retrieved 2021-03-31.
  25. ^ "The rise and regulation of thermal facial recognition technology during the COVID-19 pandemic" - Google Search". www.google.com. Retrieved 2021-04-22.
  26. ^ Center, Electronic Privacy Information. "EPIC - Patel v. Facebook". epic.org. Retrieved 2021-04-22. {{cite web}}: |first= has generic name (help)
  27. ^ "Social Network or Social Nightmare: How California Courts Can Prevent Facebook's Frightening Foray Into Facial Recognition Technology From Haunting Consumer Privacy Rights Forever". vLex. Retrieved 2021-04-24.
  28. ^ "An Update About Face Recognition on Facebook". About Facebook. 2019-09-03. Retrieved 2021-04-22.
  29. ^ "Facebook will pay $650 million to settle class action suit centered on Illinois privacy law". TechCrunch. Retrieved 2021-04-22.
  30. ^ Heilweil, Rebecca (2020-07-22). "Facebook is taking a hard look at racial bias in its algorithms". Vox. Retrieved 2021-04-23.
  31. ^ Trautman, Lawrence J. (2020-03-27). "Governance of the Facebook Privacy Crisis". Pittsburgh Journal of Technology Law & Policy. 20 (1). doi:10.5195/tlp.2020.234. ISSN 2164-800X.
  32. ^ "Facebook's '10 Year Challenge' Is Just a Harmless Meme—Right?". Wired. ISSN 1059-1028. Retrieved 2021-04-22.
  33. ^ Twitter https://twitter.com/facebook/status/1085675097766031360. Retrieved 2021-04-22. {{cite web}}: Missing or empty |title= (help)
  34. ^ Slobom, Michael (2020-01-01). "Consent, Appropriation by Manipulation, and the 10-Year Challenge: How an Internet Meme Complicated Biometric Information Privacy". Mitchell Hamline Law Review. 46 (5).
  35. ^ Becerra-Riera, Fabiola; Morales-González, Annette; Méndez-Vázquez, Heydi (2019-08-01). "A survey on facial soft biometrics for video surveillance and forensic applications". Artificial Intelligence Review. 52 (2): 1155–1187. doi:10.1007/s10462-019-09689-5. ISSN 0269-2821. S2CID 186207594.
  36. ^ Buolamwini, Joy; Gebru, Timnit (2018-01-21). "Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification". Conference on Fairness, Accountability and Transparency. PMLR: 77–91.
  37. ^ "govinfo". www.govinfo.gov. Retrieved 2021-04-23.
  38. ^ "Bias in, Bias out: Why Legislation Placing Requirements on the Procurement of Commercialized Facial Recognition Technology Must Be Passed to Protect People of Color". www.americanbar.org. Retrieved 2021-04-23.
  39. ^ Kane, Kane; Young, Amber; Majchrzak, Ann; Ransbotham, Sam (2021-03-01). "Avoiding an Oppressive Future of Machine Learning: A Design Theory for Emancipatory Assistants". Management Information Systems Quarterly. 45 (1): 371–396. doi:10.25300/MISQ/2021/1578. ISSN 0276-7783. S2CID 232369411.

Further reading

[edit]