{"637504":{"#nid":"637504","#data":{"type":"news","title":"Facial Recognition Software Needs Human Subject Experiments","body":[{"value":"\u003Cp\u003EFacial recognition software is becoming the go-to security measure for businesses, but it can be inaccurate and racially biased. Although many companies have proposed adding human intervention to mitigate this, a Georgia Tech researcher says human-subject experiments must be a priority before human intervention is considered a one-size-fits-all solution.\u003C\/p\u003E\r\n\r\n\u003Cp\u003E\u0026ldquo;Humans are biased themselves, so how can you resolve an issue of bias with a human?\u0026rdquo; School of Computer Science Ph.D. alumna \u003Ca href=\u0022https:\/\/sites.google.com\/site\/ssamadi\/\u0022\u003E\u003Cstrong\u003ESamira Samadi\u003C\/strong\u003E\u003C\/a\u003E said. \u0026ldquo;It might even make it worse.\u0026rdquo;\u003C\/p\u003E\r\n\r\n\u003Cp\u003E\u003Cstrong\u003EThe limits of facial recognition\u003C\/strong\u003E\u003C\/p\u003E\r\n\r\n\u003Cp\u003EFacial recognition software supposedly automates building security. The software takes photos as people enter a building, which it \u0026nbsp;then cross-references with an employee database. If the software finds a match, a person can enter the building.\u003C\/p\u003E\r\n\r\n\u003Cp\u003EDespite many advances in image recognition and artificial intelligence, systems are often more accurate for men with lighter skin tones and less for women with darker skin tones. Companies have proposed adding a human evaluator to compensate for the software\u0026rsquo;s limitations.\u003C\/p\u003E\r\n\r\n\u003Cp\u003EYet Samadi, who researches algorithmic fairness, immediately recognized the potential for more bias. She wanted to know whether adding a human evaluator to the process increases fairness or bias.\u003C\/p\u003E\r\n\r\n\u003Cp\u003E\u003Cstrong\u003EExperimental design\u003C\/strong\u003E\u003C\/p\u003E\r\n\r\n\u003Cp\u003EYet designing such a human\/user study is challenging as Samadi and colleagues at Microsoft Research realized. Working with actual security guards or receptionists would be ideal, but was not feasible in practice.\u003C\/p\u003E\r\n\r\n\u003Cp\u003ESamadi turned to recruiting people through Mechanical Turk as she had done in the past. These users would offer her volume, but they were not trained in recognizing faces. First, she studied how to compare faces. Then she learned how to teach Mechanical Turk users about facial recognition systems, how to make decisions about the accuracy of the system, and how to be confident in that decision.\u003C\/p\u003E\r\n\r\n\u003Cp\u003EAfter research, Samadi developed a user study and did some trials with friends to ensure the study was clear and understandable. Then she ran the study on 300 users on Mechanical Turk.\u003C\/p\u003E\r\n\r\n\u003Cp\u003EEach user was trained on how to distinguish faces and evaluate the software. Next, the user saw two images and how they were scored by the software. Samadi expected the human evaluator would show bias between two lighter versus dark-skinned people, but the results were much different.\u003C\/p\u003E\r\n\r\n\u003Cp\u003E\u003Cstrong\u003EFuture studies\u003C\/strong\u003E\u003C\/p\u003E\r\n\r\n\u003Cp\u003E\u0026ldquo;We really tried to imitate a real world scenario, but that actually made it more complicated for the users,\u0026rdquo; Samadi said.\u003C\/p\u003E\r\n\r\n\u003Cp\u003EThe researchers were unsure whether the problem with the study \u0026nbsp;was because users didn\u0026rsquo;t understand the study or biased behavior, but they ultimately decided not to publish the research. However, Samadi did publish a position paper, \u003Ca href=\u0022https:\/\/drive.google.com\/file\/d\/1iH_KJXP5xD1ieZzrXZ2gKwFyFmdv93Bf\/view\u0022\u003E\u003Cem\u003EA Human in the Loop is Not Enough: The Need for Human-Subject Experiments in Facial Recognition\u003C\/em\u003E\u003C\/a\u003E, with Microsoft Research\u0026rsquo;s \u003Cstrong\u003EFarough Poursabzi-Sangdeh\u003C\/strong\u003E, \u003Cstrong\u003EJennifer Wortman Vaughan\u003C\/strong\u003E, and \u003Cstrong\u003EHanna Wallach\u003C\/strong\u003E. Samadi presented the work at the \u003Ca href=\u0022https:\/\/chi2020.acm.org\/\u0022\u003EConference on Human Factors in Computer Systems (CHI)\u003C\/a\u003E in April.\u003C\/p\u003E\r\n\r\n\u003Cp\u003EThe paper argued about both the necessity and issue with studies like these. There are four main challenges about both the efficacy and generalizability with a human-subject study like the one they conducted:\u003C\/p\u003E\r\n\r\n\u003Cp\u003E-\u003Cstrong\u003EDatasets: \u003C\/strong\u003EFinding an appropriate dataset is difficult for a number of factors: Sourcing images ethically is challenging because past research has relied on celebrity or politician images who are easily recognizable and thus bias the study. Many datasets are also already biased and contain more lighter-skinned faces than darker. Also, many datasets are higher quality than what would be found in camera footage and not an effective real world comparison.\u003C\/p\u003E\r\n\r\n\u003Cp\u003E-\u003Cstrong\u003EParticipants:\u003C\/strong\u003E Many available participants for studies like these are students or Mechanical Turk workers who are inexperienced in facial recognition.\u003C\/p\u003E\r\n\r\n\u003Cp\u003E-\u003Cstrong\u003EContext:\u003C\/strong\u003E Recognizing faces in an experiment is not comparable to on the job duties when an unfamiliar person may be a threat.\u003C\/p\u003E\r\n\r\n\u003Cp\u003E-\u003Cstrong\u003EUser Interface:\u003C\/strong\u003E Companies do not release their user interfaces for facial recognition software, leaving it up to researchers to design something that may not reflect what is used in real world software.\u003C\/p\u003E\r\n\r\n\u003Cp\u003E\u0026ldquo;If someone wants to attack this problem in the future, they should know the challenges they have ahead of them,\u0026rdquo; Samadi said.\u003C\/p\u003E\r\n","summary":null,"format":"limited_html"}],"field_subtitle":"","field_summary":"","field_summary_sentence":[{"value":" A Georgia Tech researcher says human-subject experiments must be a priority before human intervention is considered a one-size-fits-all solution."}],"uid":"34541","created_gmt":"2020-08-04 18:57:31","changed_gmt":"2020-08-04 19:02:47","author":"Tess Malone","boilerplate_text":"","field_publication":"","field_article_url":"","dateline":{"date":"2020-08-04T00:00:00-04:00","iso_date":"2020-08-04T00:00:00-04:00","tz":"America\/New_York"},"extras":[],"hg_media":{"637505":{"id":"637505","type":"image","title":"Facial Recognition ","body":null,"created":"1596567725","gmt_created":"2020-08-04 19:02:05","changed":"1596567744","gmt_changed":"2020-08-04 19:02:24","alt":"Facial recognition graphic","file":{"fid":"242495","name":"Face_Recognition_3252983.png","image_path":"\/sites\/default\/files\/images\/Face_Recognition_3252983.png","image_full_path":"http:\/\/www.tlwarc.hg.gatech.edu\/\/sites\/default\/files\/images\/Face_Recognition_3252983.png","mime":"image\/png","size":126445,"path_740":"http:\/\/www.tlwarc.hg.gatech.edu\/sites\/default\/files\/styles\/740xx_scale\/public\/images\/Face_Recognition_3252983.png?itok=bJSWQjqU"}}},"media_ids":["637505"],"groups":[{"id":"47223","name":"College of Computing"},{"id":"50875","name":"School of Computer Science"}],"categories":[],"keywords":[],"core_research_areas":[],"news_room_topics":[],"event_categories":[],"invited_audience":[],"affiliations":[],"classification":[],"areas_of_expertise":[],"news_and_recent_appearances":[],"phone":[],"contact":[{"value":"\u003Cp\u003ETess Malone, Communications Officer\u003C\/p\u003E\r\n\r\n\u003Cp\u003E\u003Ca href=\u0022mailto:tess.malone@cc.gatech.edu\u0022\u003Etess.malone@cc.gatech.edu\u003C\/a\u003E\u003C\/p\u003E\r\n\r\n\u003Cp\u003E\u0026nbsp;\u003C\/p\u003E\r\n","format":"limited_html"}],"email":[],"slides":[],"orientation":[],"userdata":""}}}