{"id":632247,"date":"2020-01-22T03:01:40","date_gmt":"2020-01-22T11:01:40","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/?post_type=msr-event&#038;p=632247"},"modified":"2025-08-06T11:53:32","modified_gmt":"2025-08-06T18:53:32","slug":"msr-cambridge-lab-lecture-deep-inter-active-learning-for-nlp-cure-all-or-catastrophe","status":"publish","type":"msr-event","link":"https:\/\/www.microsoft.com\/en-us\/research\/event\/msr-cambridge-lab-lecture-deep-inter-active-learning-for-nlp-cure-all-or-catastrophe\/","title":{"rendered":"MSR Cambridge Lab Lecture: Deep (Inter-)Active Learning for NLP: Cure-all or Catastrophe?"},"content":{"rendered":"\n\n<p>21 Station Road<br \/>\nCambridge<br \/>\nCB1 2FB<span id=\"label-external-link\" class=\"sr-only\" aria-hidden=\"true\">Opens in a new tab<\/span><\/p>\n<p>Zachary Chase Lipton, Carnegie Mellon University<span id=\"label-external-link\" class=\"sr-only\" aria-hidden=\"true\">Opens in a new tab<\/span><\/p>\n<p><span style=\"color: black\">While deep learning produces supervised models with unprecedented predictive performance on many tasks, under typical training procedures, advantages over classical methods emerge only with large datasets. The extreme data-dependence of reinforcement learners may be even more problematic. Millions of experiences sampled from video-games come cheaply, but human-interacting systems can\u2019t afford to waste so much labor. In this talk, I will discuss several efforts to increase the labor-efficiency of learning from human interactions. Specifically, I will cover work on learning dialogue policies, deep active learning for natural language processing, learning from noisy and singly-labeled data, and active learning with partial feedback. Finally, time permitting, I\u2019ll discuss a new approach for reducing the reliance of NLP models on spurious associations in the data that relies on a new mechanism for interacting with annotators. <\/span><\/p>\n<p>&nbsp;<span id=\"label-external-link\" class=\"sr-only\" aria-hidden=\"true\">Opens in a new tab<\/span><\/p>\n<p><span style=\"color: black\">Zachary Chase Lipton is an assistant professor at Carnegie Mellon University appointed in both the Machine Learning Department and Tepper School of Business. His research spans core machine learning methods and their social impact and addresses diverse application areas, including clinical medicine and natural language processing. Current research focuses include robustness under distribution shift, breast cancer screening, the effective and equitable allocation of organs, and the intersection of causal thinking with messy data. He is the founder of the Approximately Correct (<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"https:\/\/nam06.safelinks.protection.outlook.com\/?url=http%3A%2F%2Fapproximatelycorrect.com%2F&data=02%7C01%7CLyndsay.Upex%40microsoft.com%7C663db0ec00b64aa3106908d79f25fa41%7C72f988bf86f141af91ab2d7cd011db47%7C1%7C0%7C637152857835682983&sdata=qg3zUFUAFNBZjQTF6DY1DRQAb96euYJhgQugYGUTMIo%3D&reserved=0\">approximatelycorrect.com<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>) blog and the creator of Dive Into Deep Learning, an interactive open-source book drafted entirely through Jupyter notebooks. Find on Twitter (@zacharylipton) or GitHub (@zackchase).<\/span><\/p>\n<p>&nbsp;<\/p>\n<p>&nbsp;<span id=\"label-external-link\" class=\"sr-only\" aria-hidden=\"true\">Opens in a new tab<\/span><\/p>\n","protected":false},"excerpt":{"rendered":"<p>21 Station Road Cambridge CB1 2FBOpens in a new tab Zachary Chase Lipton, Carnegie Mellon UniversityOpens in a new tab While deep learning produces supervised models with unprecedented predictive performance on many tasks, under typical training procedures, advantages over classical methods emerge only with large datasets. The extreme data-dependence of reinforcement learners may be even [&hellip;]<\/p>\n","protected":false},"featured_media":0,"template":"","meta":{"msr-url-field":"","msr-podcast-episode":"","msrModifiedDate":"","msrModifiedDateEnabled":false,"ep_exclude_from_search":false,"_classifai_error":"","msr_startdate":"2020-01-24","msr_enddate":"2020-01-24","msr_location":"MSR Cambridge","msr_expirationdate":"","msr_event_recording_link":"","msr_event_link":"","msr_event_link_redirect":false,"msr_event_time":"15:00-16:00","msr_hide_region":false,"msr_private_event":false,"msr_hide_image_in_river":0,"footnotes":""},"research-area":[],"msr-region":[],"msr-event-type":[197944],"msr-video-type":[],"msr-locale":[268875],"msr-program-audience":[],"msr-post-option":[],"msr-impact-theme":[],"class_list":["post-632247","msr-event","type-msr-event","status-publish","hentry","msr-event-type-hosted-by-microsoft","msr-locale-en_us"],"msr_about":"<!-- wp:msr\/event-details {\"title\":\"MSR Cambridge Lab Lecture: Deep (Inter-)Active Learning for NLP: Cure-all or Catastrophe?\",\"backgroundColor\":\"grey\"} \/-->\n\n<!-- wp:msr\/content-tabs --><!-- wp:msr\/content-tab {\"title\":\"Speaker\"} --><!-- wp:freeform --><p>21 Station Road<br \/>\nCambridge<br \/>\nCB1 2FB<span id=\"label-external-link\" class=\"sr-only\" aria-hidden=\"true\">Opens in a new tab<\/span><\/p>\n<p>Zachary Chase Lipton, Carnegie Mellon University<span id=\"label-external-link\" class=\"sr-only\" aria-hidden=\"true\">Opens in a new tab<\/span><\/p>\n<!-- \/wp:freeform --><!-- \/wp:msr\/content-tab --><!-- wp:msr\/content-tab {\"title\":\"Abstract\"} --><!-- wp:freeform --><p><span style=\"color: black\">While deep learning produces supervised models with unprecedented predictive performance on many tasks, under typical training procedures, advantages over classical methods emerge only with large datasets. The extreme data-dependence of reinforcement learners may be even more problematic. Millions of experiences sampled from video-games come cheaply, but human-interacting systems can\u2019t afford to waste so much labor. In this talk, I will discuss several efforts to increase the labor-efficiency of learning from human interactions. Specifically, I will cover work on learning dialogue policies, deep active learning for natural language processing, learning from noisy and singly-labeled data, and active learning with partial feedback. Finally, time permitting, I\u2019ll discuss a new approach for reducing the reliance of NLP models on spurious associations in the data that relies on a new mechanism for interacting with annotators. <\/span><\/p>\n<p>&nbsp;<span id=\"label-external-link\" class=\"sr-only\" aria-hidden=\"true\">Opens in a new tab<\/span><\/p>\n<!-- \/wp:freeform --><!-- \/wp:msr\/content-tab --><!-- wp:msr\/content-tab {\"title\":\"Bio\"} --><!-- wp:freeform --><p><span style=\"color: black\">Zachary Chase Lipton is an assistant professor at Carnegie Mellon University appointed in both the Machine Learning Department and Tepper School of Business. His research spans core machine learning methods and their social impact and addresses diverse application areas, including clinical medicine and natural language processing. Current research focuses include robustness under distribution shift, breast cancer screening, the effective and equitable allocation of organs, and the intersection of causal thinking with messy data. He is the founder of the Approximately Correct (<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"https:\/\/nam06.safelinks.protection.outlook.com\/?url=http%3A%2F%2Fapproximatelycorrect.com%2F&amp;data=02%7C01%7CLyndsay.Upex%40microsoft.com%7C663db0ec00b64aa3106908d79f25fa41%7C72f988bf86f141af91ab2d7cd011db47%7C1%7C0%7C637152857835682983&amp;sdata=qg3zUFUAFNBZjQTF6DY1DRQAb96euYJhgQugYGUTMIo%3D&amp;reserved=0\">approximatelycorrect.com<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>) blog and the creator of Dive Into Deep Learning, an interactive open-source book drafted entirely through Jupyter notebooks. Find on Twitter (@zacharylipton) or GitHub (@zackchase).<\/span><\/p>\n<p>&nbsp;<\/p>\n<p>&nbsp;<span id=\"label-external-link\" class=\"sr-only\" aria-hidden=\"true\">Opens in a new tab<\/span><\/p>\n<!-- \/wp:freeform --><!-- \/wp:msr\/content-tab --><!-- \/wp:msr\/content-tabs -->","tab-content":[{"id":0,"name":"Speaker","content":"Zachary Chase Lipton, Carnegie Mellon University"},{"id":1,"name":"Abstract","content":"<span style=\"color: black\">While deep learning produces supervised models with unprecedented predictive performance on many tasks, under typical training procedures, advantages over classical methods emerge only with large datasets. The extreme data-dependence of reinforcement learners may be even more problematic. Millions of experiences sampled from video-games come cheaply, but human-interacting systems can\u2019t afford to waste so much labor. In this talk, I will discuss several efforts to increase the labor-efficiency of learning from human interactions. Specifically, I will cover work on learning dialogue policies, deep active learning for natural language processing, learning from noisy and singly-labeled data, and active learning with partial feedback. Finally, time permitting, I\u2019ll discuss a new approach for reducing the reliance of NLP models on spurious associations in the data that relies on a new mechanism for interacting with annotators. <\/span>\r\n\r\n&nbsp;"},{"id":2,"name":"Bio","content":"<span style=\"color: black\">Zachary Chase Lipton is an assistant professor at Carnegie Mellon University appointed in both the Machine Learning Department and Tepper School of Business. His research spans core machine learning methods and their social impact and addresses diverse application areas, including clinical medicine and natural language processing. Current research focuses include robustness under distribution shift, breast cancer screening, the effective and equitable allocation of organs, and the intersection of causal thinking with messy data. He is the founder of the Approximately Correct (<a href=\"https:\/\/nam06.safelinks.protection.outlook.com\/?url=http%3A%2F%2Fapproximatelycorrect.com%2F&amp;data=02%7C01%7CLyndsay.Upex%40microsoft.com%7C663db0ec00b64aa3106908d79f25fa41%7C72f988bf86f141af91ab2d7cd011db47%7C1%7C0%7C637152857835682983&amp;sdata=qg3zUFUAFNBZjQTF6DY1DRQAb96euYJhgQugYGUTMIo%3D&amp;reserved=0\">approximatelycorrect.com<\/a>) blog and the creator of Dive Into Deep Learning, an interactive open-source book drafted entirely through Jupyter notebooks. Find on Twitter (@zacharylipton) or GitHub (@zackchase).<\/span>\r\n\r\n&nbsp;\r\n\r\n&nbsp;"}],"msr_startdate":"2020-01-24","msr_enddate":"2020-01-24","msr_event_time":"15:00-16:00","msr_location":"MSR Cambridge","msr_event_link":"","msr_event_recording_link":"","msr_startdate_formatted":"January 24, 2020","msr_register_text":"Watch now","msr_cta_link":"","msr_cta_text":"","msr_cta_bi_name":"","featured_image_thumbnail":null,"event_excerpt":"Zachary Chase Lipton, Carnegie Mellon University","msr_research_lab":[199561],"related-researchers":[],"msr_impact_theme":[],"related-academic-programs":[],"related-groups":[],"related-projects":[],"related-opportunities":[],"related-publications":[],"related-videos":[],"related-posts":[],"_links":{"self":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event\/632247","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event"}],"about":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/types\/msr-event"}],"version-history":[{"count":2,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event\/632247\/revisions"}],"predecessor-version":[{"id":1147000,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event\/632247\/revisions\/1147000"}],"wp:attachment":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media?parent=632247"}],"wp:term":[{"taxonomy":"msr-research-area","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/research-area?post=632247"},{"taxonomy":"msr-region","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-region?post=632247"},{"taxonomy":"msr-event-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event-type?post=632247"},{"taxonomy":"msr-video-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-video-type?post=632247"},{"taxonomy":"msr-locale","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-locale?post=632247"},{"taxonomy":"msr-program-audience","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-program-audience?post=632247"},{"taxonomy":"msr-post-option","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-post-option?post=632247"},{"taxonomy":"msr-impact-theme","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-impact-theme?post=632247"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}