{"628671":{"#nid":"628671","#data":{"type":"news","title":"FairVis is Helping Data Scientists Discover Societal Biases in their Machine Learning Models ","body":[{"value":"\u003Cp\u003EResearchers at Georgia Tech, Carnegie Mellon University, and University of Washington have developed a data visualization system that can help data scientists discover bias in machine learning algorithms.\u0026nbsp;\u003C\/p\u003E\r\n\r\n\u003Cp\u003E\u003Ca href=\u0022https:\/\/arxiv.org\/pdf\/1904.05419.pdf\u0022\u003EFairVis\u003C\/a\u003E, presented at\u0026nbsp;\u003Ca href=\u0022http:\/\/ieeevis.org\/year\/2019\/welcome\u0022\u003EIEEE Vis 2019\u003C\/a\u003E\u0026nbsp;in Vancouver, is the first system to integrate a novel technique that allows users to audit the fairness of machine learning models by identifying and comparing different populations in their data sets.\u0026nbsp;\u003C\/p\u003E\r\n\r\n\u003Cp\u003EAccording to School of Computational Science and Engineering (CSE) Professor and co-investigator\u0026nbsp;\u003Ca href=\u0022https:\/\/poloclub.github.io\/polochau\/\u0022\u003E\u003Cstrong\u003EPolo Chau\u003C\/strong\u003E\u003C\/a\u003E\u003Cstrong\u003E,\u0026nbsp;\u003C\/strong\u003Ethis feat has never been accomplished by any platform before, and is a major contribution of FairVis to the data science and machine learning communities.\u003C\/p\u003E\r\n\r\n\u003Cp\u003E\u0026ldquo;Computers are never going to be perfect. So, the question is how to help people prioritize where to look in their data, and then, in a scalable way, enable them to compare these areas to other similar or dissimilar groups in the data. By enabling comparison of groups in a data set,\u0026nbsp;FairVis allows data to become very scannable,\u0026rdquo; he said.\u003C\/p\u003E\r\n\r\n\u003Cp\u003EIn order to do accomplish this, FairVis uses two novel techniques to find subgroups that are statistically similar.\u0026nbsp;\u003C\/p\u003E\r\n\r\n\u003Cp\u003EThe first technique groups similar items together in the training data set, calculates various performance metrics like accuracy, and then shows users which groups of people the algorithm may be biased against. The second technique uses statistical divergence to measure the distance between subgroups to allow users to compare similar groups and find larger patterns of bias.\u003C\/p\u003E\r\n\r\n\u003Cp\u003EThese outputs are then viewed and analyzed through FairVis\u0026rsquo; visual analytics system, which is designed to specifically discover and show intersectional bias.\u0026nbsp;\u003C\/p\u003E\r\n\r\n\u003Cp\u003EIntersectional bias, or bias that is found when looking at populations defined by multiple features, is a mounting challenge for scientists to tackle in an increasingly diverse world.\u003C\/p\u003E\r\n\r\n\u003Cp\u003E\u0026ldquo;While a machine learning algorithm may work very well in general, there may be certain groups for which it fails. For example, various face detection algorithms were found to be 30 percent less accurate for darker skinned women than for lighter skinned men. When you look at more specific groups of sex, race, nationality, and more, there can be hundreds or thousands of groups to audit,\u0026rdquo; said\u0026nbsp;Carnegie Mellon University\u0026nbsp;Ph.D. student\u0026nbsp;\u003Ca href=\u0022https:\/\/cabreraalex.com\/\u0022\u003E\u003Cstrong\u003EAlex Cabrera\u003C\/strong\u003E\u003C\/a\u003E.\u003C\/p\u003E\r\n\r\n\u003Cp\u003ECabrera is the primary investigator of FairVis and has been pursuing this problem since he was an undergraduate student at Georgia Tech.\u003C\/p\u003E\r\n\r\n\u003Cp\u003E\u0026ldquo;During the summer of my junior year I had been researching various topics in machine learning, and discovered some recent work showing how machine learning models can encode and worsen societal biases. I quickly realized that not only was this a significant issue, with examples of biased algorithms in everything from hiring systems to self-driving cars, but that my own work during my internship had the possibility to be biased against lower socioeconomic groups.\u0026rdquo;\u003C\/p\u003E\r\n\r\n\u003Cp\u003EThis is when Cabrera reached out to Chau who then recruited the help of CSE alumni\u0026nbsp;\u003Ca href=\u0022https:\/\/minsuk.com\/\u0022\u003E\u003Cstrong\u003EMinsuk Kahng\u003C\/strong\u003E\u003C\/a\u003E, CSE Ph.D.\u0026nbsp;\u003Ca href=\u0022https:\/\/fredhohman.com\/\u0022\u003E\u003Cstrong\u003EFred Hohman\u003C\/strong\u003E\u003C\/a\u003E\u003Cstrong\u003E,\u0026nbsp;\u003C\/strong\u003ECollege of Computing undergraduate student\u0026nbsp;\u003Ca href=\u0022http:\/\/www.willepperson.com\/\u0022\u003E\u003Cstrong\u003EWill Epperson\u003C\/strong\u003E\u003C\/a\u003E\u003Cstrong\u003E,\u0026nbsp;\u003C\/strong\u003Eand University of Washington Assistant Professor\u0026nbsp;\u003Ca href=\u0022http:\/\/jamiemorgenstern.com\/\u0022\u003E\u003Cstrong\u003EJamie Morgenstern\u003C\/strong\u003E\u003C\/a\u003E\u003Cstrong\u003E.\u003C\/strong\u003E\u003C\/p\u003E\r\n\r\n\u003Cp\u003EMorgenstern is the lead researcher for a number of projects related to fairness in machine learning, including the study Cabrera mentioned about self-driving cars. This particular study shows the potentially\u0026nbsp;\u003Ca href=\u0022https:\/\/www.scs.gatech.edu\/news\/620309\/research-reveals-possibly-fatal-consequences-algorithmic-bias\u0022\u003Efatal consequences of algorithmic bias\u003C\/a\u003E\u0026nbsp;which highlights the severity of software created without fairness embedded into its core.\u0026nbsp;\u003C\/p\u003E\r\n\r\n\u003Cp\u003EFairVis is one of the first systems that helps us achieve a dramatic step towards understanding and addressing the problem of fairness in machine learning, and prevents similar headlines from making their way to reality in the future.\u0026nbsp;\u0026nbsp;\u003C\/p\u003E\r\n\r\n\u003Cp\u003EHowever, Cabrera stressed that the solution does not simply end with better data practices.\u003C\/p\u003E\r\n\r\n\u003Cp\u003E\u0026ldquo;Fairness is an extremely difficult problem, a so-called \u0026lsquo;wicked problem\u0026rsquo;, that will not be solved by technology alone,\u0026rdquo; he said.\u0026nbsp;\u003C\/p\u003E\r\n\r\n\u003Cp\u003E\u0026ldquo;Social scientists, policy makers, and engineers need to work together to make inroads and ensure that our algorithms are equitable for all people. We hope FairVis is a step in this direction and helps people start the conversation about how to tackle and address these issues.\u0026rdquo;\u003C\/p\u003E\r\n","summary":null,"format":"limited_html"}],"field_subtitle":"","field_summary":"","field_summary_sentence":[{"value":"Researchers present FairVis -  a visual analytics system that enables discovery of user subgroups to discover bias in machine learning models."}],"uid":"34540","created_gmt":"2019-11-06 18:04:14","changed_gmt":"2019-12-06 14:44:50","author":"Kristen Perez","boilerplate_text":"","field_publication":"","field_article_url":"","dateline":{"date":"2019-11-06T00:00:00-05:00","iso_date":"2019-11-06T00:00:00-05:00","tz":"America\/New_York"},"extras":[],"hg_media":{"628667":{"id":"628667","type":"image","title":"FairVis","body":null,"created":"1573063180","gmt_created":"2019-11-06 17:59:40","changed":"1573063180","gmt_changed":"2019-11-06 17:59:40","alt":"A screenshot of a\u00a0visual analytics system that enables discovery of user subgroups to discover bias in machine learning models","file":{"fid":"239426","name":"FairVis.jpg","image_path":"\/sites\/default\/files\/images\/FairVis.jpg","image_full_path":"http:\/\/www.tlwarc.hg.gatech.edu\/\/sites\/default\/files\/images\/FairVis.jpg","mime":"image\/jpeg","size":29572,"path_740":"http:\/\/www.tlwarc.hg.gatech.edu\/sites\/default\/files\/styles\/740xx_scale\/public\/images\/FairVis.jpg?itok=FqHIbwJ2"}}},"media_ids":["628667"],"groups":[{"id":"47223","name":"College of Computing"},{"id":"431631","name":"OMS"},{"id":"50877","name":"School of Computational Science and Engineering"},{"id":"50875","name":"School of Computer Science"},{"id":"1299","name":"GVU Center"}],"categories":[],"keywords":[{"id":"4305","name":"cse"},{"id":"83261","name":"Polo Chau"},{"id":"181315","name":"cse-dse"},{"id":"181220","name":"cse-ml"},{"id":"182995","name":"FairVis"},{"id":"1496","name":"Ethics"},{"id":"9167","name":"machine learning"},{"id":"307","name":"fairness"},{"id":"182996","name":"Alex Cabrera"}],"core_research_areas":[],"news_room_topics":[],"event_categories":[],"invited_audience":[],"affiliations":[],"classification":[],"areas_of_expertise":[],"news_and_recent_appearances":[],"phone":[],"contact":[{"value":"\u003Cp\u003E\u003Ca href=\u0022mailto:kristen.perez@cc.gatech.edu\u0022\u003EKristen Perez\u003C\/a\u003E\u003C\/p\u003E\r\n\r\n\u003Cp\u003ECommunications Officer\u003C\/p\u003E\r\n","format":"limited_html"}],"email":["kristen.perez@cc.gatech.edu"],"slides":[],"orientation":[],"userdata":""}}}