{"id":494648,"date":"2018-07-17T10:46:26","date_gmt":"2018-07-17T17:46:26","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/?p=494648"},"modified":"2018-08-03T10:55:39","modified_gmt":"2018-08-03T17:55:39","slug":"machine-learning-for-fair-decisions","status":"publish","type":"post","link":"https:\/\/www.microsoft.com\/en-us\/research\/blog\/machine-learning-for-fair-decisions\/","title":{"rendered":"Machine Learning for fair decisions"},"content":{"rendered":"<h3><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter size-full wp-image-494657\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2018\/07\/FairClassificaction_Header_07_18_1000x400.png\" alt=\"\" width=\"2000\" height=\"800\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2018\/07\/FairClassificaction_Header_07_18_1000x400.png 2000w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2018\/07\/FairClassificaction_Header_07_18_1000x400-300x120.png 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2018\/07\/FairClassificaction_Header_07_18_1000x400-768x307.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2018\/07\/FairClassificaction_Header_07_18_1000x400-1024x410.png 1024w\" sizes=\"auto, (max-width: 2000px) 100vw, 2000px\" \/><\/h3>\n<p>Over the past decade, machine learning systems have begun to play a key role in many high-stakes decisions: Who is interviewed for a job? Who is approved for a bank loan? Who receives parole? Who is admitted to a school?<\/p>\n<p>Human decision makers are susceptible to many forms of prejudice and bias, such as those rooted in gender and racial stereotypes. One might hope that machines would be able to make decisions more fairly than humans. However, news stories and numerous research studies have found that machine learning systems can inadvertently discriminate against minorities, historically disadvantaged populations and other groups.<\/p>\n<p>In essence, this is because machine learning systems are trained to replicate decisions present in the data with which they are trained and these decisions reflect society&#8217;s historical biases.<\/p>\n<p>Naturally, researchers want to mitigate these biases, but there are several challenges. For example, there are many different definitions of fairness. Should the same number of men and women be interviewed for a job or should the number of men and women interviewed reflect the proportions of men and women in the applicant pool? What about nonbinary applicants? Should machine learning systems even be used in hiring contexts? Answers to questions like these are non-trivial and often depend on societal context. On top of that, re-engineering existing machine learning pipelines to incorporate fairness considerations can be hard. How can you train a boosted-decision-tree classifier to respect specific gender proportions? What about other fairness definitions? What about training a two-layer neural network? Or a residual network? Each of these questions can require many months of research and engineering.<\/p>\n<p>Our work, outlined in a paper titled, \u201c<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/arxiv.org\/abs\/1803.02453\">A Reductions Approach to Fair Classification<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>,\u201d presented this month at the 35th International Conference on Machine Learning (<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/icml.cc\/\">ICML 2018<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>) in Stockholm, Sweden, focuses on some of these challenges, providing a provably and empirically sound method for turning any common classifier into a \u201cfair\u201d classifier according to any of a wide range of fairness definitions.<\/p>\n<p>To understand <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/github.com\/Microsoft\/fairlearn\">our method<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, consider the process of choosing applicants to interview for a job where it is desirable to have an interview pool that is balanced with respect to gender and race\u2014a fairness definition known as demographic parity. Our method can turn a classifier that predicts who should be interviewed based on previous (potentially biased) hiring decisions into a classifier that predicts who should be interviewed while also respecting demographic parity (or another fairness definition).<\/p>\n<p>Such techniques are called \u201creductions\u201d because they reduce the problem we wish to solve into a different problem\u2014typically a more standard problem for which many algorithms already exist. Our fair learning reduction is, in fact, a special case of a more general reduction for imposing constraints on classifiers.<\/p>\n<p>Our method operates as a game between the classification algorithm and a &#8220;fairness enforcer.&#8221; The classification algorithm tries to come up with the most accurate classification rule on possibly-reweighted data, while the fairness enforcer checks the chosen fairness definition. The training data is reweighted based on the output of the fairness enforcer and passed back to the classification algorithm. For instance, applicants of a certain gender or race might be upweighted or downweighted, so that the classification algorithm is better able to find a classification rule that is fair with respect to the desired gender or racial proportions. Eventually, this process yields a classification rule that is fair, according to the fairness definition. Moreover, it is also the most accurate among all fair rules, provided the classification algorithm consistently tries to minimize the error on the reweighted data.<\/p>\n<div id=\"attachment_494681\" style=\"width: 1131px\" class=\"wp-caption aligncenter\"><img loading=\"lazy\" decoding=\"async\" aria-describedby=\"caption-attachment-494681\" class=\"wp-image-494681 size-full\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2018\/07\/figure-all-for-mean-simplified.png\" alt=\"\" width=\"1121\" height=\"663\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2018\/07\/figure-all-for-mean-simplified.png 1121w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2018\/07\/figure-all-for-mean-simplified-300x177.png 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2018\/07\/figure-all-for-mean-simplified-768x454.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2018\/07\/figure-all-for-mean-simplified-1024x606.png 1024w\" sizes=\"auto, (max-width: 1121px) 100vw, 1121px\" \/><p id=\"caption-attachment-494681\" class=\"wp-caption-text\">Our algorithm (exp. grad. reduction) matches or outperforms baseline approaches across many data sets, classifier types and fairness measures.<\/p><\/div>\n<p>In practice, this process takes about five iterations to yield classification rules that are as good as or better than many previous methods tailored to specific fairness definitions. Our method works with many different definitions of fairness and only needs access to protected attributes, such as gender or race, during training, not when the classifier is deployed in an application. Because our method works as a &#8220;wrapper&#8221; around any existing classifier, it is easy to incorporate into existing machine learning systems.<\/p>\n<p>Our results contribute to the ongoing conversation about developing \u201cfair\u201d machine learning systems. However, there are many important questions that remain to be addressed. For example, we might not have access to protected attributes at training time or might not want to commit to a single definition of fairness. More fundamentally, in order to effectively mitigate impacts of historical bias, it might be insufficient to impose a quantitative definition of fairness; it might be necessary to systematically collect additional data and monitor the effects of decisions over time, or even avoid the use of machine learning systems in some domains.<\/p>\n<h3>Related Links:<\/h3>\n<ul>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/github.com\/Microsoft\/fairlearn\">GitHub:\u00a0Reductions for Fair Machine Learning<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/a-reductions-approach-to-fair-classification\/\">Paper: A Reductions Approach to Fair Classification<\/a><\/li>\n<\/ul>\n","protected":false},"excerpt":{"rendered":"<p>Over the past decade, machine learning systems have begun to play a key role in many high-stakes decisions: Who is interviewed for a job? Who is approved for a bank loan? Who receives parole? Who is admitted to a school? Human decision makers are susceptible to many forms of prejudice and bias, such as those [&hellip;]<\/p>\n","protected":false},"author":37074,"featured_media":494696,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"msr-url-field":"","msr-podcast-episode":"","msrModifiedDate":"","msrModifiedDateEnabled":false,"ep_exclude_from_search":false,"_classifai_error":"","msr-author-ordering":[{"type":"user_nicename","value":"Miro Dud\u00edk","user_id":"32867"},{"type":"user_nicename","value":"John Langford","user_id":"32204"},{"type":"user_nicename","value":"Hanna Wallach","user_id":"34779"},{"type":"user_nicename","value":"Alekh Agarwal","user_id":"30928"}],"msr_hide_image_in_river":0,"footnotes":""},"categories":[194455],"tags":[],"research-area":[13556],"msr-region":[],"msr-event-type":[],"msr-locale":[268875],"msr-post-option":[],"msr-impact-theme":[],"msr-promo-type":[],"msr-podcast-series":[],"class_list":["post-494648","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-machine-learning","msr-research-area-artificial-intelligence","msr-locale-en_us"],"msr_event_details":{"start":"","end":"","location":""},"podcast_url":"","podcast_episode":"","msr_research_lab":[],"msr_impact_theme":[],"related-publications":[],"related-downloads":[],"related-videos":[],"related-academic-programs":[],"related-groups":[144902,372368,395930],"related-projects":[],"related-events":[492515],"related-researchers":[{"type":"user_nicename","value":"Miro Dud\u00edk","user_id":32867,"display_name":"Miro Dud\u00edk","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/mdudik\/\" aria-label=\"Visit the profile page for Miro Dud\u00edk\">Miro Dud\u00edk<\/a>","is_active":false,"last_first":"Dud\u00edk, Miro","people_section":0,"alias":"mdudik"},{"type":"user_nicename","value":"John Langford","user_id":32204,"display_name":"John Langford","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/jcl\/\" aria-label=\"Visit the profile page for John Langford\">John Langford<\/a>","is_active":false,"last_first":"Langford, John","people_section":0,"alias":"jcl"},{"type":"user_nicename","value":"Hanna Wallach","user_id":34779,"display_name":"Hanna Wallach","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/wallach\/\" aria-label=\"Visit the profile page for Hanna Wallach\">Hanna Wallach<\/a>","is_active":false,"last_first":"Wallach, Hanna","people_section":0,"alias":"wallach"}],"msr_type":"Post","featured_image_thumbnail":"<img width=\"926\" height=\"540\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2018\/07\/FairClassificaction_Carousel_07_18_480x280.png\" class=\"img-object-cover\" alt=\"Fair classification\" decoding=\"async\" loading=\"lazy\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2018\/07\/FairClassificaction_Carousel_07_18_480x280.png 960w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2018\/07\/FairClassificaction_Carousel_07_18_480x280-300x175.png 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2018\/07\/FairClassificaction_Carousel_07_18_480x280-768x448.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2018\/07\/FairClassificaction_Carousel_07_18_480x280-480x280.png 480w\" sizes=\"auto, (max-width: 926px) 100vw, 926px\" \/>","byline":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/mdudik\/\" title=\"Go to researcher profile for Miro Dud\u00edk\" aria-label=\"Go to researcher profile for Miro Dud\u00edk\" data-bi-type=\"byline author\" data-bi-cN=\"Miro Dud\u00edk\">Miro Dud\u00edk<\/a>, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/jcl\/\" title=\"Go to researcher profile for John Langford\" aria-label=\"Go to researcher profile for John Langford\" data-bi-type=\"byline author\" data-bi-cN=\"John Langford\">John Langford<\/a>, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/wallach\/\" title=\"Go to researcher profile for Hanna Wallach\" aria-label=\"Go to researcher profile for Hanna Wallach\" data-bi-type=\"byline author\" data-bi-cN=\"Hanna Wallach\">Hanna Wallach<\/a>, and Alekh Agarwal","formattedDate":"July 17, 2018","formattedExcerpt":"Over the past decade, machine learning systems have begun to play a key role in many high-stakes decisions: Who is interviewed for a job? Who is approved for a bank loan? Who receives parole? Who is admitted to a school? Human decision makers are susceptible&hellip;","locale":{"slug":"en_us","name":"English","native":"","english":"English"},"_links":{"self":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts\/494648","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/users\/37074"}],"replies":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/comments?post=494648"}],"version-history":[{"count":18,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts\/494648\/revisions"}],"predecessor-version":[{"id":495107,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts\/494648\/revisions\/495107"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media\/494696"}],"wp:attachment":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media?parent=494648"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/categories?post=494648"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/tags?post=494648"},{"taxonomy":"msr-research-area","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/research-area?post=494648"},{"taxonomy":"msr-region","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-region?post=494648"},{"taxonomy":"msr-event-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event-type?post=494648"},{"taxonomy":"msr-locale","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-locale?post=494648"},{"taxonomy":"msr-post-option","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-post-option?post=494648"},{"taxonomy":"msr-impact-theme","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-impact-theme?post=494648"},{"taxonomy":"msr-promo-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-promo-type?post=494648"},{"taxonomy":"msr-podcast-series","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-podcast-series?post=494648"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}