{"id":866259,"date":"2022-08-02T10:37:42","date_gmt":"2022-08-02T17:37:42","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/?post_type=msr-project&#038;p=866259"},"modified":"2022-08-04T09:16:44","modified_gmt":"2022-08-04T16:16:44","slug":"dp-transformers","status":"publish","type":"msr-project","link":"https:\/\/www.microsoft.com\/en-us\/research\/project\/dp-transformers\/","title":{"rendered":"dp-transformers"},"content":{"rendered":"<section class=\"mb-3 moray-highlight\">\n\t<div class=\"card-img-overlay mx-lg-0\">\n\t\t<div class=\"card-background bg-gray-200 has-background- card-background--full-bleed\">\n\t\t\t\t\t<\/div>\n\t\t<!-- Foreground -->\n\t\t<div class=\"card-foreground d-flex mt-md-n5 my-lg-5 px-g px-lg-0\">\n\t\t\t<!-- Container -->\n\t\t\t<div class=\"container d-flex mt-md-n5 my-lg-5 align-self-center\">\n\t\t\t\t<!-- Card wrapper -->\n\t\t\t\t<div class=\"w-100 w-lg-col-5\">\n\t\t\t\t\t<!-- Card -->\n\t\t\t\t\t<div class=\"card material-md-card py-5 px-md-5\">\n\t\t\t\t\t\t<div class=\"card-body \">\n\t\t\t\t\t\t\t\n\t\t\t\t\t\t\t\n\n<h1 id=\"dp-transformers\">dp-transformers<\/h1>\n\n\n\n<p>Training transformer models with differential privacy<\/p>\n\n\t\t\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t<\/div>\n\t\t<\/div>\n\t<\/div>\n<\/section>\n\n\n\n<div class=\"wp-block-group is-layout-flow wp-block-group-is-layout-flow\">\n\n\n<p>Transformer models have recently taken the field of Natural Language Processing (NLP) by storm as large language models based on the transformer architecture have shown impressive performance across a wide range of applications. However, when investigating these models in terms of Responsible AI, a valid concern remains that privacy-preserving techniques must be properly applied when these models are trained with private data.&nbsp;<\/p>\n\n\n\n<p>Differential Privacy (DP) has become a gold standard definition of privacy that offers rigorous privacy guarantees to individuals while enabling learning from a population. Among a vast set of applications, training machine learning models with DP in particular has the potential to extract great value from private data while protecting privacy of the participants.&nbsp;<\/p>\n\n\n\n<p>Motivated by our recent <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/openreview.net\/forum?id=Q42f0dfjECO\" target=\"_blank\" rel=\"noopener noreferrer\">work<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, we are releasing a repository for training transformer models with differential privacy. Our repository, available at <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/www.github.com\/microsoft\/dp-transformers\">https:\/\/www.github.com\/microsoft\/dp-transformers<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, is based on integrating the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/opacus.ai\/\" target=\"_blank\" rel=\"noopener noreferrer\">Opacus library<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> to the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/huggingface.co\/\" target=\"_blank\" rel=\"noopener noreferrer\">Hugging Face<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> platform. We aim to serve the privacy-preserving ML community in utilizing the state-of-the-art models while respecting the privacy of the individuals constituting what these models learn from.<\/p>\n\n\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>Training transformer models with differential privacy Transformer models have recently taken the field of Natural Language Processing (NLP) by storm as large language models based on the transformer architecture have shown impressive performance across a wide range of applications. However, when investigating these models in terms of Responsible AI, a valid concern remains that privacy-preserving [&hellip;]<\/p>\n","protected":false},"featured_media":0,"template":"","meta":{"msr-url-field":"","msr-podcast-episode":"","msrModifiedDate":"","msrModifiedDateEnabled":false,"ep_exclude_from_search":false,"_classifai_error":"","footnotes":""},"research-area":[13556,13558],"msr-locale":[268875],"msr-impact-theme":[],"msr-pillar":[],"class_list":["post-866259","msr-project","type-msr-project","status-publish","hentry","msr-research-area-artificial-intelligence","msr-research-area-security-privacy-cryptography","msr-locale-en_us","msr-archive-status-active"],"msr_project_start":"","related-publications":[864054],"related-downloads":[],"related-videos":[],"related-groups":[702211,756487,761911,793670,1054512],"related-events":[],"related-opportunities":[],"related-posts":[],"related-articles":[],"tab-content":[],"slides":[],"related-researchers":[{"type":"user_nicename","display_name":"Huseyin Atahan Inan","user_id":40426,"people_section":"People","alias":"huinan"},{"type":"user_nicename","display_name":"Daniel Jones","user_id":41030,"people_section":"People","alias":"jonesdaniel"},{"type":"user_nicename","display_name":"Lukas Wutschitz","user_id":38775,"people_section":"People","alias":"luwutsch"}],"msr_research_lab":[],"msr_impact_theme":[],"_links":{"self":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-project\/866259","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-project"}],"about":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/types\/msr-project"}],"version-history":[{"count":11,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-project\/866259\/revisions"}],"predecessor-version":[{"id":867150,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-project\/866259\/revisions\/867150"}],"wp:attachment":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media?parent=866259"}],"wp:term":[{"taxonomy":"msr-research-area","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/research-area?post=866259"},{"taxonomy":"msr-locale","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-locale?post=866259"},{"taxonomy":"msr-impact-theme","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-impact-theme?post=866259"},{"taxonomy":"msr-pillar","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-pillar?post=866259"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}