{"id":1109289,"date":"2024-12-09T10:37:27","date_gmt":"2024-12-09T18:37:27","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/?post_type=msr-story&#038;p=1109289"},"modified":"2025-08-28T20:47:46","modified_gmt":"2025-08-29T03:47:46","slug":"microsoft-at-neurips-2024-advancing-ai-research-across-domains","status":"publish","type":"msr-story","link":"https:\/\/www.microsoft.com\/en-us\/research\/story\/microsoft-at-neurips-2024-advancing-ai-research-across-domains\/","title":{"rendered":"Microsoft at NeurIPS 2024: Advancing AI research across domains"},"content":{"rendered":"\n<div class=\"wp-block-cover has-parallax is-style-default\" style=\"min-height:360px;aspect-ratio:unset;\"><span aria-hidden=\"true\" class=\"wp-block-cover__background has-black-background-color has-background-dim-40 has-background-dim\"><\/span><div role=\"img\" aria-label=\"abstract geometric pattern in purple, blue, green\" class=\"wp-block-cover__image-background wp-image-1110399 has-parallax\" style=\"background-position:50% 50%;background-image:url(https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/12\/NeurIPS-2024-Stories_Hero_Feature-2000x1333-1.jpg)\"><\/div><div class=\"wp-block-cover__inner-container is-layout-constrained wp-container-core-cover-is-layout-2cb6a229 wp-block-cover-is-layout-constrained\">\n<div class=\"wp-block-group is-content-justification-left is-layout-constrained wp-container-core-group-is-layout-719fd2c2 wp-block-group-is-layout-constrained\">\n<div style=\"height:100px\" aria-hidden=\"true\" class=\"wp-block-spacer d-none d-sm-block\"><\/div>\n\n\n\n<h1 class=\"wp-block-heading is-style-display\" id=\"find-my-things-new-teachable-ai-tool-helps-blind-and-low-vision-people-locate-lost-personal-items-1\">Microsoft at NeurIPS 2024: Advancing AI research across domains<\/h1>\n\n\n\n<div style=\"height:100px\" aria-hidden=\"true\" class=\"wp-block-spacer d-none d-sm-block\"><\/div>\n<\/div>\n<\/div><\/div>\n\n\n\n<article class=\"wp-block-group alignfull mt-0 is-layout-constrained wp-block-group-is-layout-constrained\">\n<div style=\"padding-bottom:0; padding-top:0\" class=\"wp-block-msr-immersive-section alignfull row has-background-gradient has-background-gradient-spectrum-3 wp-block-msr-immersive-section\">\n\t\n\t<div class=\"container\">\n\t\t<div class=\"wp-block-msr-immersive-section__wrapper\">\n\t\t\t<div style=\"height:40px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\t\t<\/div>\n\t<\/div>\n\n\t<\/div>\n\n\n\n<div class=\"wp-block-columns is-style-dark-mode p-4 z-20 container theme-dark is-layout-flex wp-container-core-columns-is-layout-9d6595d7 wp-block-columns-is-layout-flex\">\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\" style=\"flex-basis:22%\"><\/div>\n\n\n\n<div class=\"wp-block-column headings-large is-layout-flow wp-block-column-is-layout-flow\" style=\"flex-basis:56%\">\n<div style=\"height:40px\" aria-hidden=\"true\" class=\"wp-block-spacer is-style-default d-none d-md-block\"><\/div>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"microsoft-sponsor-of-neurips-2024\">Microsoft is proud to sponsor the 38th Conference on Neural Information Processing Systems (NeurIPS 2024), a leading global forum for machine learning and AI.<\/h3>\n\n\n\n<div class=\"wp-block-buttons is-layout-flex wp-block-buttons-is-layout-flex\">\n<div class=\"wp-block-button is-style-pill\"><a data-bi-type=\"button\" class=\"wp-block-button__link wp-element-button\" href=\"https:\/\/www.microsoft.com\/en-us\/research\/microsoft-at-neurips-2024\/\" target=\"_blank\" rel=\"noreferrer noopener\">Microsoft at NeurIPS AI experience<\/a><\/div>\n\n\n\n<div class=\"wp-block-button is-style-outline is-style-outline--1\"><a data-bi-type=\"button\" class=\"wp-block-button__link wp-element-button\" href=\"https:\/\/www.microsoft.com\/en-us\/research\/event\/neurips-2024\">Microsoft at NeurIPS 2024<\/a><\/div>\n<\/div>\n\n\n\n<p>The event gathers researchers, industry leaders, and practitioners to exchange ideas, address challenges, and advance innovations to shape the future of AI. <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/lidongz\/\">Lidong Zhou<\/a>, managing director of <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/lab\/microsoft-research-asia\/\">Microsoft Research Asia<\/a>, will be one of this year\u2019s keynote speakers.<\/p>\n\n\n\n<p>More than 100 papers by Microsoft researchers and collaborators have been accepted at NeurIPS 2024, including five oral presentations and 19 spotlight sessions. While these research projects cover a broad range of topics, a shared theme ties them together: advancing the efficiency, scalability, and robustness of machine learning models while addressing real-world challenges like human-centric interaction and cultural considerations.<\/p>\n\n\n\n<blockquote class=\"wp-block-quote is-style-spectrum--blue-green is-layout-flow wp-block-quote-is-layout-flow\">\n<p>Visit us at <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/event\/neurips-2024\/booth-schedule\/\">Booth #445<\/a><\/p>\n<\/blockquote>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\" style=\"flex-basis:22%\"><\/div>\n<\/div>\n\n\n<div class=\"heading-wrapper has-text-align-center\">\n<h2 class=\"wp-block-heading has-text-align-center is-style-spectrum-fill\" id=\"neurips-oral-presentations\">NeurIPS oral presentations<\/h2>\n<\/div>\n\n\n<div class=\"wp-block-group theme-dark spectrum-border spectrum-border--blue-green spectrum-border--w-60 spectrum-border--position-right mt-0 briefing-book-talks is-layout-flow wp-block-group-is-layout-flow\" style=\"padding-right:64px;padding-bottom:71px\">\n<div class=\"wp-block-columns are-vertically-aligned-top wp-block-columns--stack-tablet is-layout-flex wp-container-core-columns-is-layout-9d6595d7 wp-block-columns-is-layout-flex\">\n<div class=\"wp-block-column is-vertically-aligned-top mt-0 mt-sm-5 is-layout-flow wp-block-column-is-layout-flow\" style=\"flex-basis:66.66%\">\n<figure class=\"wp-block-image size-full\"><img loading=\"lazy\" decoding=\"async\" width=\"1401\" height=\"788\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/12\/NeurIPS_Podcast_Hero_Feature_No_Text_1400x788.jpg\" alt=\"illustration of MSR Podcast guests for NeurIPS 2024 | Back row (left to right): Steven Euijong Whang, Chris Bishop, Pranjal Chitale | Middle row (left to right): Lidong Zhou, Amber Tingle, Jindong Wang, Eliza Strickland | Front row (left to right): Dylan Foster, Gretchen Huizinga, Weizhu Chen\" class=\"wp-image-1110120\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/12\/NeurIPS_Podcast_Hero_Feature_No_Text_1400x788.jpg 1401w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/12\/NeurIPS_Podcast_Hero_Feature_No_Text_1400x788-300x169.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/12\/NeurIPS_Podcast_Hero_Feature_No_Text_1400x788-1024x576.jpg 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/12\/NeurIPS_Podcast_Hero_Feature_No_Text_1400x788-768x432.jpg 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/12\/NeurIPS_Podcast_Hero_Feature_No_Text_1400x788-1066x600.jpg 1066w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/12\/NeurIPS_Podcast_Hero_Feature_No_Text_1400x788-655x368.jpg 655w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/12\/NeurIPS_Podcast_Hero_Feature_No_Text_1400x788-240x135.jpg 240w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/12\/NeurIPS_Podcast_Hero_Feature_No_Text_1400x788-640x360.jpg 640w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/12\/NeurIPS_Podcast_Hero_Feature_No_Text_1400x788-960x540.jpg 960w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/12\/NeurIPS_Podcast_Hero_Feature_No_Text_1400x788-1280x720.jpg 1280w\" sizes=\"auto, (max-width: 1401px) 100vw, 1401px\" \/><\/figure>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-vertically-aligned-top is-layout-flow wp-block-column-is-layout-flow\" style=\"flex-basis:45%\">\n<h2 class=\"wp-block-heading pr-md-5 mr-md-5 mt-5 h1 font-weight-bold\" id=\"featured-forum-talks-1\">Not All Tokens Are What You Need for Pretraining<\/h2>\n\n\n\n<p class=\"pr-md-5 mr-md-5 mt-5\"><strong>Recipient of &#8220;Best Paper Runner Up Award&#8221;<\/strong><br><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/yegong\/\"><em>Yeyun Gong<\/em><\/a><em>, <\/em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/xiaoliu2\/\"><em>Xiao Liu<\/em><\/a><em>, Yelong Shen, Ruochen Xu, Jian Jiao, Nan Duan, <\/em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/wzchen\/\"><em>Weizhu Chen<\/em><\/a><br><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/wzchen\/\"><br><\/a>Rho-1 is a new language model that uses selective language modeling. Unlike traditional language models that predict every next token, Rho-1 selectively trains on tokens aligned with the desired distribution. This involves scoring pretraining tokens using a reference model and then training the language model with a focused loss on tokens with higher scores.<\/p>\n\n\n\n<div class=\"wp-block-buttons is-layout-flex wp-block-buttons-is-layout-flex\">\n<div class=\"wp-block-button is-style-outline mb-0 is-style-outline--2\"><a data-bi-type=\"button\" class=\"wp-block-button__link wp-element-button\" href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/not-all-tokens-are-what-you-need-for-pretraining\/\">Read the paper<\/a><\/div>\n\n\n\n<div class=\"wp-block-button is-style-outline mb-0 is-style-outline--3\"><a data-bi-type=\"button\" class=\"wp-block-button__link wp-element-button\" href=\"https:\/\/www.microsoft.com\/en-us\/research\/podcast\/abstracts-neurips-2024-with-weizhu-chen\/\">Listen to the podcast<\/a><\/div>\n<\/div>\n<\/div>\n<\/div>\n\n\n\n<div style=\"height:40px\" aria-hidden=\"true\" class=\"wp-block-spacer d-none d-sm-block\"><\/div>\n\n\n\n<div class=\"wp-block-columns are-vertically-aligned-top wp-block-columns--flexible is-layout-flex wp-container-core-columns-is-layout-9d6595d7 wp-block-columns-is-layout-flex\">\n<div class=\"wp-block-column is-vertically-aligned-top is-layout-flow wp-block-column-is-layout-flow\" style=\"flex-basis:30%\"><div class=\"heading-wrapper\">\n<h2 class=\"wp-block-heading is-style-spectrum-fill\" id=\"lightning-talks\"><\/h2>\n<\/div>\n\n\n<p><\/p>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-vertically-aligned-top is-layout-flow wp-block-column-is-layout-flow\" style=\"flex-basis:6.66%\"><\/div>\n\n\n\n<div class=\"wp-block-column is-vertically-aligned-top pr-md-5 is-layout-flow wp-block-column-is-layout-flow\" style=\"flex-basis:60%\">\n<div class=\"wp-block-columns justify-content-around is-layout-flex wp-container-core-columns-is-layout-9d6595d7 wp-block-columns-is-layout-flex\">\n<div class=\"wp-block-column is-vertically-aligned-top is-layout-flow wp-block-column-is-layout-flow\" style=\"flex-basis:90%\">\n<div class=\"wp-block-group mb-0 is-layout-constrained wp-block-group-is-layout-constrained\">\n<h4 class=\"wp-block-heading has-text-align-left\" id=\"lightning-talk-title\">Reinforcement Learning Under Latent Dynamics: Toward Statistical and Algorithmic Modularity<\/h4>\n\n\n\n<p class=\"has-text-align-left\"><em>Philip Amortila, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/dylanfoster\/\">Dylan J. Foster<\/a>, Nan Jiang, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/akshaykr\/\">Akshay Krishnamurthy<\/a>, Zakaria Mhammedi<br><\/em><br>This research investigates reinforcement learning under general latent dynamics, demonstrating that traditional function approximation becomes intractable with rich observations unless latent pushforward coverability is present. The authors also developed efficient reductions to adapt latent Markov decision process (MDP) algorithms for complex observations, providing a foundation for a unified statistical and algorithmic theory for reinforcement learning under latent dynamics.<\/p>\n\n\n\n<div class=\"wp-block-buttons mb-0 is-layout-flex wp-block-buttons-is-layout-flex\">\n<div class=\"wp-block-button is-style-outline is-style-outline--4\"><a data-bi-type=\"button\" class=\"wp-block-button__link wp-element-button\" href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/reinforcement-learning-under-latent-dynamics-toward-statistical-and-algorithmic-modularity\/\">Read the paper<\/a><\/div>\n\n\n\n<div class=\"wp-block-button is-style-outline is-style-outline--5\"><a data-bi-type=\"button\" class=\"wp-block-button__link wp-element-button\" href=\"https:\/\/www.microsoft.com\/en-us\/research\/podcast\/abstracts-neurips-2024-with-dylan-foster\/\">Listen to the podcast<\/a><\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n\n\n\n<div class=\"wp-block-columns justify-content-around is-layout-flex wp-container-core-columns-is-layout-9d6595d7 wp-block-columns-is-layout-flex\">\n<div class=\"wp-block-column is-vertically-aligned-top is-layout-flow wp-block-column-is-layout-flow\" style=\"flex-basis:90%\">\n<div class=\"wp-block-group mb-0 is-layout-constrained wp-block-group-is-layout-constrained\">\n<h4 class=\"wp-block-heading has-text-align-left\" id=\"lightning-talk-title\">CVQA: Culturally-diverse Multilingual Visual Question Answering Benchmark<\/h4>\n\n\n\n<p class=\"has-text-align-left\"><em>David Romero, Chenyang Lyu, Haryo Akbarianto Wibowo, Teresa Lynn, Injy Hamed, Aditya Nanda Kishore, Aishik Mandal, <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/www.linkedin.com\/in\/pranjalchitale\/?originalSubdomain=in\">Pranjal Chitale<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, et al.<\/em><br><br>CVQA is a culturally diverse, multilingual visual question-answering benchmark that involves native speakers and cultural experts in the data collection process. It includes culturally driven images and questions from 30 countries across four continents, covering 31 languages and 13 scripts, and provides a total of 10k questions. While it is a challenging benchmark for current state-of-the-art multimodal large language models (MLLMs), it is also a tool to assess cultural capabilities and biases in these models.<\/p>\n\n\n\n<div class=\"wp-block-buttons mb-0 is-layout-flex wp-block-buttons-is-layout-flex\">\n<div class=\"wp-block-button is-style-outline is-style-outline--6\"><a data-bi-type=\"button\" class=\"wp-block-button__link wp-element-button\" href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/cvqa-culturally-diverse-multilingual-visual-question-answering-benchmark\/\">Read the paper<\/a><\/div>\n\n\n\n<div class=\"wp-block-button is-style-outline is-style-outline--7\"><a data-bi-type=\"button\" class=\"wp-block-button__link wp-element-button\" href=\"https:\/\/www.microsoft.com\/en-us\/research\/podcast\/abstracts-neurips-2024-with-pranjal-chitale\/\">Listen to the podcast<\/a><\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n\n\n\n<div class=\"wp-block-columns justify-content-around is-layout-flex wp-container-core-columns-is-layout-9d6595d7 wp-block-columns-is-layout-flex\">\n<div class=\"wp-block-column is-vertically-aligned-top is-layout-flow wp-block-column-is-layout-flow\" style=\"flex-basis:90%\">\n<div class=\"wp-block-group mb-0 is-layout-constrained wp-block-group-is-layout-constrained\">\n<h4 class=\"wp-block-heading has-text-align-left\" id=\"lightning-talk-title\">VASA-1: Lifelike Audio-Driven Talking Faces Generated in Real Time<\/h4>\n\n\n\n<p class=\"has-text-align-left\"><em>Sicheng Xu, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/guoch\/\">Guojun Chen<\/a>, Yu-Xiao Guo, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/jiaoyan\/\">Jiaolong Yang<\/a>, Chong Li, Zhenyu Zang, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/yizzhan\/\">Yizhong Zhang<\/a>, Xin Tong, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/bainguo\/\">Baining Guo<\/a><\/em><br><br>VASA is a framework for generating lifelike talking faces with visual affective skills (VAS) from a static image and audio clip. The premiere model, VASA-1, synchronizes lip movements with speech while capturing facial nuances and natural head motions, enabled by a holistic facial dynamics and head movement generation model and an expressive face latent space built from video data.<\/p>\n\n\n\n<div class=\"wp-block-buttons mb-0 is-layout-flex wp-block-buttons-is-layout-flex\">\n<div class=\"wp-block-button is-style-outline is-style-outline--8\"><a data-bi-type=\"button\" class=\"wp-block-button__link wp-element-button\" href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/vasa-1-lifelike-audio-driven-talking-faces-generated-in-real-time\/\">Read the paper<\/a><\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n\n\n\n<div class=\"wp-block-columns justify-content-around is-layout-flex wp-container-core-columns-is-layout-9d6595d7 wp-block-columns-is-layout-flex\">\n<div class=\"wp-block-column is-vertically-aligned-top is-layout-flow wp-block-column-is-layout-flow\" style=\"flex-basis:90%\">\n<div class=\"wp-block-group mb-0 is-layout-constrained wp-block-group-is-layout-constrained\">\n<h4 class=\"wp-block-heading has-text-align-left\" id=\"lightning-talk-title\">You Only Cache Once: Decoder-Decoder Architectures for Language Models<\/h4>\n\n\n\n<p class=\"has-text-align-left\"><em>Yutao Sun, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/lidong1\/\">Li Dong<\/a>, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/yizhu1\/\">Yi Zhu<\/a>, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/shaohanh\/\">Shaohan Huang<\/a>, Wenhui Wang, Shuming Ma, Quanlu Zhang, Jianyong Wang, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/fuwei\/\">Furu Wei<\/a><\/em><br><br>You only cache once (YOCO) is a decoder-decoder architecture for LLMs that reduces GPU memory usage by caching key-value pairs only once, while retaining global attention. A self-decoder encodes key-value caches that are reused by a cross-decoder that leverages cross-attention. This enables YOCO to speed up the prefill stage through a computation flow that allows early exit without altering the final output.<\/p>\n\n\n\n<div class=\"wp-block-buttons mb-0 is-layout-flex wp-block-buttons-is-layout-flex\">\n<div class=\"wp-block-button is-style-outline is-style-outline--9\"><a data-bi-type=\"button\" class=\"wp-block-button__link wp-element-button\" href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/you-only-cache-once-decoder-decoder-architectures-for-language-models\/\">Read the paper<\/a><\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n\n\n\n<div style=\"height:30px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<div class=\"wp-block-group theme-dark is-style-default container is-layout-constrained wp-block-group-is-layout-constrained\">\n<div class=\"wp-block-columns is-style-dark-mode p-4 z-20 container theme-dark is-layout-flex wp-container-core-columns-is-layout-9d6595d7 wp-block-columns-is-layout-flex\">\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\" style=\"flex-basis:22%\"><\/div>\n\n\n\n<div class=\"wp-block-column headings-large is-layout-flow wp-block-column-is-layout-flow\" style=\"flex-basis:56%\">\n<div style=\"height:40px\" aria-hidden=\"true\" class=\"wp-block-spacer is-style-default d-none d-md-block\"><\/div>\n\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity\"\/>\n\n\n<div class=\"heading-wrapper\">\n<h3 class=\"wp-block-heading has-text-align-center is-style-spectrum-fill\" id=\"neurips-spotlight-sessions\">NeurIPS spotlight sessions<\/h3>\n<\/div>\n\n\n<p class=\"small text-neutral-300\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/erbench-an-entity-relationship-based-automatically-verifiable-hallucination-benchmark-for-large-language-models\/\"><strong>ERBench: An Entity-Relationship based Automatically Verifiable Hallucination Benchmark for Large Language Models<\/strong><\/a><br><em>Jio Oh, Soyeon Kim, Junseok Seo, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/jindwang\/\">Jindong Wang<\/a>, Ruochen Xu, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/xingx\/\">Xing Xie<\/a>, Steven Euijong Whang<\/em><br>To thoroughly analyze LLMs, the authors propose ERBench, which automatically converts any relational database into a benchmark based on the entity-relationship model.<\/p>\n\n\n\n<p class=\"small text-neutral-300\"><strong><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/a-study-of-plasticity-loss-in-on-policy-deep-reinforcement-learning\/\">A Study of Plasticity Loss in On-Policy Deep Reinforcement Learning<\/a><\/strong><br><em>Arthur Juliani, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/joash\/\">Jordan Ash<\/a><\/em><br>The authors to conduct extensive experiments on plasticity loss in on-policy deep reinforcement learning and various mitigation methods.<\/p>\n\n\n\n<p class=\"small text-neutral-300\"><strong><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/advancing-spiking-neural-networks-for-sequential-modeling-with-central-pattern-generators\/\">Advancing Spiking Neural Networks for Sequential Modeling through Central Pattern Generators<\/a><\/strong><br><em><em>Changze Lv, <\/em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/dongqihan\/\"><em>Dongqi Han<\/em><\/a><em>, <\/em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/yansenwang\/\"><em>Yansen Wang<\/em><\/a><em>, Xiaoqing Zheng, Xuanjing Huang, <\/em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/dongsli\/\"><em>Dongsheng Li<\/em><\/a><\/em><br>CPG-PE is a novel positional encoding (PE) technique for spiking neural networks inspired by central pattern generators in the human brain.<\/p>\n\n\n\n<p class=\"small text-neutral-300\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/beyond-assouad-fano-and-le-cam-toward-unified-lower-bounds-for-statistical-estimation-and-interactive-decision-making\/\"><strong>Assouad, Fano, and Le Cam with Interaction: A Unifying Lower Bound Framework and Characterization for Bandit Learnability<\/strong><br><\/a><em>Fan Chen, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/dylanfoster\/\">Dylan J. Foster<\/a>, Yanjun Han, Jian Qian, Alexander Rakhlin, Yunbei Xu<\/em><br>The authors develop a unified framework for lower bound methods in statistical estimation and interactive decision making. They also propose a unified view of these distinct methodologies.<\/p>\n\n\n\n<p class=\"small text-neutral-300\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/bpqp-a-differentiable-convex-optimization-framework-for-efficient-end-to-end-learning\/\"><strong>BPQP: A Differentiable Convex Optimization Framework for Efficient End-to-End Learning<\/strong><\/a><br><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/xiaoyang\/\"><em>Xiao Yang<\/em><\/a><em>, Xu Yang, <\/em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/weiqiliu\/\"><em>Weiqing Liu<\/em><\/a><em>, Lewen Wang, <\/em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/jiabia\/\"><em>Jiang Bian<\/em><\/a><br>To enhance efficiency, the authors reformulate the backward pass as a simplified and decoupled quadratic programming problem by leveraging the structural properties of the Karush\u2013Kuhn\u2013Tucker (KKT) matrix.<\/p>\n\n\n\n<p class=\"small text-neutral-300\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/compositional-generalization-across-distributional-shifts-with-sparse-tree-operations\/\"><strong>Compositional Generalization Across Distributional Shifts with Sparse Tree Operations<\/strong><\/a><br><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/psmo\/\"><em>Paul Smolensky<\/em><\/a><em>, <\/em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/jfgao\/\"><em>Jianfeng Gao<\/em><\/a><em>, <\/em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/rfernand\/\"><em>Roland Fernandez<\/em><\/a><br>This work investigates a unified neurosymbolic system where transformations in the network can be interpreted as both symbolic and neural computation simultaneously. It extends a unified neurosymbolic architecture.<\/p>\n\n\n\n<p class=\"small text-neutral-300\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/dataset-and-lessons-learned-from-the-2024-satml-llm-capture-the-flag-competition\/\"><strong>Dataset and Lessons Learned from the 2024 SaTML LLM Capture-the-Flag Competition<\/strong><\/a><br><em>Edoardo Debenedetti, Javier Rando, Daniel Paleka, Fineas Silaghi, Dragos Albastroiu, Niv Cohen, Yuval Lemberg, <\/em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/reshmighosh\/\"><em>Reshmi Ghosh<\/em><\/a><em>, Ahmed Salem, Rui Wen, <\/em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/gcherubin\/\"><em>Giovanni Cherubin<\/em><\/a><em>, <\/em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/santiago\/\"><em>Santiago Zanella-B\u00e9guelin<\/em><\/a><em>, Robin Schmid, Victor Klemm, Takahiro Miki, Chenhao Li, Stefan Kraft, Mario Fritz, Florian Tramer, Sahar Abdelnabi, Lea Sch\u00f6nherr<\/em><br>This report summarizes insights from a capture-the-flag competition at IEEE SaTML 2024, which highlighted the challenges in defending large language model systems against malicious message attacks.<\/p>\n\n\n\n<p class=\"small text-neutral-300\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/diffusion-for-world-modeling-visual-details-matter-in-atari\"><strong>Diffusion for World Modeling: Visual Details Matter in Atari<\/strong><\/a><br><em>Eloi Alonso, Adam Jelley, Vincent Micheli, Anssi Kanervisto, Amos Storkey, <\/em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/teapearce.github.io\/\"><em>Tim Pearce<\/em><span class=\"sr-only\"> (opens in new tab)<\/span><\/a><em>, Fran\u00e7ois Fleuret<\/em><br>This work presents DIAMOND (diffusion as a model of environment dreams), an open-source reinforcement learning agent trained in a diffusion world model.<\/p>\n\n\n\n<p class=\"small text-neutral-300\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/discoveryworld-a-virtual-environment-for-developing-and-evaluating-automated-scientific-discovery-agents\/\"><strong>DISCOVERYWORLD: A Virtual Environment for Developing and Evaluating Automated Scientific Discovery Agents<\/strong><\/a><br><em>Peter Alexander Jansen, <\/em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/macote\/\"><em>Marc-Alexandre <em>C\u00f4t\u00e9<\/em><\/em><\/a><em>, Tushar Khot, Erin Bransom, Bhavana Dalvi, Bodhisattwa Prasad Majumder, Oyvind Tafjord, Peter Clark<\/em><br>DISCOVERYWORLD is an open-source virtual environment for developing and benchmarking an agent&#8217;s ability to perform complete scientific discovery cycles, with 120 diverse tasks across diverse topics.<\/p>\n\n\n\n<p class=\"small text-neutral-300\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/efficient-adversarial-training-in-llms-with-continuous-attacks\/\"><strong>Efficient Adversarial Training in LLMs with Continuous Attacks<\/strong><br><\/a><em>Sophie Xhonneux, <\/em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/alsordon\/\"><em>Alessandro Sordoni<\/em><\/a><em>, Stephan G\u00fcnnemann, Gauthier Gidel, Leo Schwinn<\/em><br>This research introduces an efficient approach to adversarial attacks by calculating them in the LLM\u2019s continuous embedding space.<\/p>\n\n\n\n<p class=\"small text-neutral-300\"><strong><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/generalized-linear-bandits-with-limited-adaptivity\/\">Generalized Linear Bandits with Limited Adaptivity<\/a><\/strong><br><em>Ayush Sawarni, Nirjhar Das, Siddharth Barman, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/gauravsinha\/\">Gaurav Sinha<\/a><\/em><br>This paper studies the generalized linear contextual bandit problem under limited adaptivity and introduces two algorithms, B-GLinCB and RS-GLinCB, to address two prevalent settings.<\/p>\n\n\n\n<p class=\"small text-neutral-300\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/human-aware-vision-and-language-navigation-bridging-simulation-to-reality-with-dynamic-human-interactions\/\"><strong>Human-Aware Vision-and-Language Navigation: Bridging Simulation to Reality with Dynamic Human Interactions<\/strong><\/a>&nbsp;<br><em>Minghan Li, Heng Li, Zhi-Qi Cheng, Yifei Dong, Yuxuan Zhou, Jun-Yan He, <\/em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/qid\/\"><em>Qi Dai<\/em><\/a><em>, Teruko Mitamura, Alexander G. Hauptmann&nbsp;<\/em><br>This research introduces Human-Aware Vision-and-Language Navigation (HA-VLN), extending traditional VLN by incorporating dynamic human activities and relaxing key assumptions.<\/p>\n\n\n\n<p class=\"small text-neutral-300\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/identifying-equivalent-training-dynamics\/\"><strong>Identifying Equivalent Training Dynamics<\/strong><br><\/a><em>William T. Redman, <\/em><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/www.linkedin.com\/in\/juan-m-bello-rivas\/\"><em>Juan M. Bello-Rivas<\/em><span class=\"sr-only\"> (opens in new tab)<\/span><\/a><em>, M. Fonoberova, Ryan Mohr, I. Kevrekidis, Igor Mezi\u0107<\/em><br>Using advances in Koopman operator theory, the authors developed a framework for identifying conjugate and nonconjugate training dynamics.<\/p>\n\n\n\n<p class=\"small text-neutral-300\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/implicit-curriculum-in-procgen-made-explicit\/\"><strong>Implicit Curriculum in Procgen Made Explicit<\/strong><br><\/a><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/kaixin96.github.io\/\"><em>Kaixin Wang<\/em><span class=\"sr-only\"> (opens in new tab)<\/span><\/a><em>, Xinchao Wang<\/em><br>This work investigates the learning process itself under the multi-level training in Procgen, which exhibits a gradual shift from easy to hard contexts, suggesting an implicit curriculum in multi-level training.<\/p>\n\n\n\n<p class=\"small text-neutral-300\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/is-behavior-cloning-all-you-need-understanding-horizon-in-imitation-learning\/\"><strong>Is Behavior Cloning All You Need? Understanding Horizon in Imitation Learning<\/strong><br><\/a><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/dylanfoster\/\"><em>Dylan J. Foster<\/em><\/a><em>, <\/em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/blockadam\/\"><em>Adam Block<\/em><\/a><em>, Dipendra Misra<\/em><br>The authors show they can achieve horizon-independent sample complexity in offline imitation learning when the range of the cumulative payoffs and an appropriate notion of supervised learning complexity for the policy class are controlled.<\/p>\n\n\n\n<p class=\"small text-neutral-300\"><strong><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/minference-1-0-accelerating-pre-filling-for-long-context-llms-via-dynamic-sparse-attention\/\">MInference: Accelerating Pre-filling for Long-Context LLMs via Dynamic Sparse Attention<\/a><\/strong><br><em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/hjiang\/\">Huiqiang Jiang<\/a>, Yucheng Li, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/chengzhang\/\">Chengruidong Zhang<\/a>, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/qianhuiwu\/\">Qianhui Wu<\/a>, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/xufluo\/\">Xufang Luo<\/a>, Surin Ahn, Zhenhua Han, Amir H. Abdi, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/dongsli\/\">Dongsheng Li<\/a>, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/cyl\/\">Chin-Yew Lin<\/a>, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/yuqyang\/\">Yuqing Yang<\/a>, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/liliqiu\/\">Lili Qiu<\/a><\/em><br>MInference is sparse calculation method designed to accelerate pre-filling of long-sequence processing, identifying three unique patterns in long-context attention matrices that can be used for efficient sparse computation on GPUs.<\/p>\n\n\n\n<p class=\"small text-neutral-300\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/the-power-of-resets-in-online-reinforcement-learning\/\"><strong>The Power of Resets in Online Reinforcement Learning<\/strong><br><\/a><em>Zakaria Mhammedi, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/dylanfoster\/\">Dylan J. Foster<\/a>, Alexander Rakhlin<\/em><br>This study explores the potential of simulators through reinforcement learning with local simulator access, an RL protocol where the agent is allowed to reset to previously observed states and follow their dynamics during training.<\/p>\n\n\n\n<p class=\"small text-neutral-300\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/videogui-a-benchmark-for-gui-automation-from-instructional-videos\/\"><strong>VideoGUI: A Benchmark for GUI Automation from Instructional Videos<\/strong><\/a><br><em>Kevin Qinghong Lin, <\/em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/linjli\/\"><em>Linjie Li<\/em><\/a><em>, Difei Gao, Qinchen Wu, Mingyi Yan, Zhengyuan Yang, <\/em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/lijuanw\/\"><em>Lijuan Wang<\/em><\/a><em>, Mike Zheng Shou<\/em><br>This research introduces VideoGUI, a novel multi-modal benchmark designed to evaluate GUI assistants on visual-centric GUI tasks through a hierarchical process, allowing for identification of the specific levels at which they may fail.<\/p>\n\n\n\n<p class=\"small text-neutral-300\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/voila-a-aligning-vision-language-models-with-users-gaze-attention\/\"><strong>Voila-A: Aligning Vision-Language Models with User\u2019s Gaze Attention<\/strong><br><\/a><em>Kun Yan, <\/em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/leiji\/\"><em>Lei Ji<\/em><\/a><em>, Zeyu Wang, Yuntao Wang, Nan Duan, Shuai Ma<\/em><br>The authors introduce gaze information, feasibly collected by AR or VR devices, and propose a novel approach for gaze alignment to enhance the interpretability and effectiveness of these models in real-world applications.<\/p>\n\n\n\n<blockquote class=\"wp-block-quote is-style-spectrum--blue-green is-layout-flow wp-block-quote-is-layout-flow\">\n<p><strong><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/event\/neurips-2024\/publications\/\">Explore Microsoft&#8217;s 100+ accepted papers<\/a><\/strong><\/p>\n<\/blockquote>\n\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity\"\/>\n\n\n<div class=\"heading-wrapper\">\n<h3 class=\"wp-block-heading is-style-spectrum-fill\" id=\"ML4H\">Microsoft at ML4H 2024<\/h3>\n<\/div>\n\n\n<p>Co-located with NeurIPS is the <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/event\/neurips-2024\/ml4h\/\">AHLI Machine Learning for Health (ML4H) Symposium<\/a>, an event that unites machine learning researchers, clinicians, and healthcare data experts to advance AI applications in healthcare. Microsoft&#8217;s contribution of four papers to this symposium underscores its commitment to improving medical imaging and clinical workflows through AI, focusing on accuracy, efficiency, and interpretability.<\/p>\n\n\n\n<div class=\"wp-block-buttons is-layout-flex wp-block-buttons-is-layout-flex\">\n<div class=\"wp-block-button is-style-outline is-style-outline--10\"><a data-bi-type=\"button\" class=\"wp-block-button__link wp-element-button\" href=\"https:\/\/www.microsoft.com\/en-us\/research\/event\/neurips-2024\/ml4h\/\">Accepted papers<\/a><\/div>\n<\/div>\n\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity\"\/>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\" style=\"flex-basis:22%\"><\/div>\n<\/div>\n\n\n\n<div style=\"height:60px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<div class=\"wp-block-columns is-layout-flex wp-container-core-columns-is-layout-9d6595d7 wp-block-columns-is-layout-flex\">\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\" style=\"flex-basis:33.33%\">\n<h3 class=\"wp-block-heading is-style-default h2\" id=\"resources\">Other resources<\/h3>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\" style=\"flex-basis:50%\">\n<div class=\"wp-block-columns is-layout-flex wp-container-core-columns-is-layout-9d6595d7 wp-block-columns-is-layout-flex\">\n<div class=\"wp-block-column is-vertically-aligned-top is-layout-flow wp-block-column-is-layout-flow\">\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/event\/neurips-2024\/booth-schedule\/\">NeurIPS 2024 booth schedule<\/a><\/p>\n\n\n\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/event\/neurips-2024\/opportunities\/\">NeurIPS 2024 career opportunities<\/a><\/p>\n\n\n\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/event\/neurips-2024\/ml4h\/\">ML4H 2024 accepted papers<\/a><\/p>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-vertically-aligned-top is-layout-flow wp-block-column-is-layout-flow\">\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/podcast\/\">Microsoft Research Podcast<\/a><\/p>\n\n\n\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/blog\">Microsoft Research Blog<\/a><\/p>\n\n\n\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/story\/sep-2024-brief\/\">Microsoft Research Forum series<\/a><\/p>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n<\/div>\n\n\n\n<div style=\"height:60px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n<\/article>\n","protected":false},"excerpt":{"rendered":"<p>We\u2019re excited to be a part of #NeurIPS2024! Explore the future of AI with over 100 groundbreaking papers, including oral and spotlight sessions, on reinforcement learning, advanced language model training, and multilingual, culturally inclusive benchmarks.<\/p>\n","protected":false},"featured_media":1110402,"template":"","meta":{"msr-url-field":"","msr-podcast-episode":"","msrModifiedDate":"","msrModifiedDateEnabled":false,"ep_exclude_from_search":false,"_classifai_error":"","footnotes":""},"research-area":[13561,13556,13562,13551,13554],"msr-locale":[268875],"msr-post-option":[],"class_list":["post-1109289","msr-story","type-msr-story","status-publish","has-post-thumbnail","hentry","msr-research-area-algorithms","msr-research-area-artificial-intelligence","msr-research-area-computer-vision","msr-research-area-graphics-and-multimedia","msr-research-area-human-computer-interaction","msr-locale-en_us"],"related-researchers":[],"related-publications":[],"related-downloads":[],"related-videos":[],"related-projects":[],"related-groups":[],"related-events":[],"related-posts":[],"msr_impact_theme":[],"_links":{"self":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-story\/1109289","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-story"}],"about":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/types\/msr-story"}],"version-history":[{"count":89,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-story\/1109289\/revisions"}],"predecessor-version":[{"id":1111377,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-story\/1109289\/revisions\/1111377"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media\/1110402"}],"wp:attachment":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media?parent=1109289"}],"wp:term":[{"taxonomy":"msr-research-area","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/research-area?post=1109289"},{"taxonomy":"msr-locale","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-locale?post=1109289"},{"taxonomy":"msr-post-option","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-post-option?post=1109289"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}