{"id":769108,"date":"2021-08-18T19:13:51","date_gmt":"2021-08-19T02:13:51","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/?post_type=msr-research-item&#038;p=769108"},"modified":"2021-11-24T05:29:14","modified_gmt":"2021-11-24T13:29:14","slug":"the-new-jim-code-reimagining-the-default-settings-of-technology-society","status":"publish","type":"msr-video","link":"https:\/\/www.microsoft.com\/en-us\/research\/video\/the-new-jim-code-reimagining-the-default-settings-of-technology-society\/","title":{"rendered":"The New Jim Code: Reimagining the Default Settings of Technology & Society"},"content":{"rendered":"<p>From everyday apps to complex algorithms, technology has the potential to hide, speed, and deepen discrimination, while appearing neutral and even benevolent when compared to racist practices of a previous era. In this talk, Ruha Benjamin presents the concept of the \u201cNew Jim Code\u201d to explore a range of discriminatory designs that encode inequity: by explicitly amplifying racial hierarchies, by ignoring but thereby replicating social divisions, or by aiming to fix racial bias but ultimately doing quite the opposite. This presentation takes us into the world of biased bots, altruistic algorithms, and their many entanglements, and provides conceptual tools to decode tech promises with historical and sociological insight. Ruha will also consider how race itself is a tool designed to naturalize social hierarchies and, in doing so, she challenges us to question not only the technologies we are sold, but also the ones we manufacture ourselves.<\/p>\n<h4>Learning Materials<\/h4>\n<ul>\n<li>Publication: <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/doi.org\/10.1126\/science.aaz3873\" target=\"_blank\" rel=\"noopener noreferrer\">Assessing risk, automating racism<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, 2019<\/li>\n<li>Book: <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/www.degruyter.com\/document\/doi\/10.1515\/9781478004493\/html\" target=\"_blank\" rel=\"noopener noreferrer\"><em>Captivating Technology: Race, Carceral Technoscience, and Liberatory Imagination in Everyday Life<\/em><span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, 2019<\/li>\n<li>Book: <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/www.degruyter.com\/document\/doi\/10.1515\/9780804786737\/html\" target=\"_blank\" rel=\"noopener noreferrer\"><em>People\u2019s Science: Bodies and Rights on the Stem Cell Frontier<\/em><span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, 2013<\/li>\n<li>Book: <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/www.wiley.com\/en-us\/Race+After+Technology%3A+Abolitionist+Tools+for+the+New+Jim+Code-p-9781509526437\" target=\"_blank\" rel=\"noopener noreferrer\"><em>Race After Technology: Abolitionist Tools for the New Jim Code<\/em><span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, 2019<\/li>\n<li>Podcast appearance: <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/omny.fm\/shows\/factually-with-adam-conover\/technology-and-race-with-ruha-benjamin#:~:text=Technology%20and%20Race%20with%20Ruha%20Benjamin.%20Princeton%20University,on%20to%20solve%20what%20are%20ultimately%20social%20problems.\" target=\"_blank\" rel=\"noopener noreferrer\">Technology and Race with Ruha Benjamin | Factually! with Adam Conover<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, 2020<\/li>\n<li>Podcast appearance: <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/www.cnet.com\/news\/why-tech-made-racial-injustice-worse-and-how-to-fix-it\/\" target=\"_blank\" rel=\"noopener noreferrer\">Why tech made racial injustice worse, and how to fix it | CNET&#8217;s Now What Podcast<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, 2020<\/li>\n<\/ul>\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/video\/racist-tropes-and-labor-discipline-how-tech-inherits-and-reproduces-global-imaginaries-of-race-and-work\/\">Learn more about the Race and Technology Research Lecture Series ><\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>From everyday apps to complex algorithms, technology has the potential to hide, speed, and deepen discrimination, while appearing neutral and even benevolent when compared to racist practices of a previous era. In this talk, Ruha Benjamin presents the concept of the \u201cNew Jim Code\u201d to explore a range of discriminatory designs that encode inequity: by [&hellip;]<\/p>\n","protected":false},"featured_media":799609,"template":"","meta":{"msr-url-field":"","msr-podcast-episode":"","msrModifiedDate":"","msrModifiedDateEnabled":false,"ep_exclude_from_search":false,"_classifai_error":"","msr_hide_image_in_river":0,"footnotes":""},"research-area":[13561,13556,13559],"msr-video-type":[262225],"msr-locale":[268875],"msr-post-option":[],"msr-session-type":[],"msr-impact-theme":[],"msr-pillar":[],"msr-episode":[],"msr-research-theme":[],"class_list":["post-769108","msr-video","type-msr-video","status-publish","has-post-thumbnail","hentry","msr-research-area-algorithms","msr-research-area-artificial-intelligence","msr-research-area-social-sciences","msr-video-type-race-and-technology-a-research-lecture-series","msr-locale-en_us"],"msr_download_urls":"","msr_external_url":"https:\/\/youtu.be\/aMuD_lAy2zQ","msr_secondary_video_url":"","msr_video_file":"","_links":{"self":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-video\/769108","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-video"}],"about":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/types\/msr-video"}],"version-history":[{"count":3,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-video\/769108\/revisions"}],"predecessor-version":[{"id":799582,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-video\/769108\/revisions\/799582"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media\/799609"}],"wp:attachment":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media?parent=769108"}],"wp:term":[{"taxonomy":"msr-research-area","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/research-area?post=769108"},{"taxonomy":"msr-video-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-video-type?post=769108"},{"taxonomy":"msr-locale","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-locale?post=769108"},{"taxonomy":"msr-post-option","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-post-option?post=769108"},{"taxonomy":"msr-session-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-session-type?post=769108"},{"taxonomy":"msr-impact-theme","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-impact-theme?post=769108"},{"taxonomy":"msr-pillar","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-pillar?post=769108"},{"taxonomy":"msr-episode","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-episode?post=769108"},{"taxonomy":"msr-research-theme","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-research-theme?post=769108"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}