{"id":1081755,"date":"2024-09-26T05:15:00","date_gmt":"2024-09-26T12:15:00","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/?p=1081755"},"modified":"2025-01-14T13:38:56","modified_gmt":"2025-01-14T21:38:56","slug":"microsoft-research-forum-episode-4-the-future-of-multimodal-models-a-new-small-language-model-and-other-ai-updates","status":"publish","type":"post","link":"https:\/\/www.microsoft.com\/en-us\/research\/blog\/microsoft-research-forum-episode-4-the-future-of-multimodal-models-a-new-small-language-model-and-other-ai-updates\/","title":{"rendered":"Microsoft Research Forum Episode 4: The future of multimodal models, a new \u201csmall\u201d language model, and other AI updates"},"content":{"rendered":"\n<p>Microsoft Research Forum is a continuous exchange of ideas about science and technology research in the era of general AI. In <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/story\/sep-2024-brief\/\">the latest episode<\/a>,\u202fresearchers discussed the latest multimodal AI models, advanced benchmarks for AI evaluation and model self-improvement, and an entirely new kind of computer for AI inference and hard optimization. Researchers at Microsoft are working to explore breakthrough technology that can help advance everything from weather prediction to materials design.\u00a0<\/p>\n\n\n\n<p>Below is a brief recap of the event, including select quotes from the presentations. <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/register.researchforum.microsoft.com\/\" target=\"_blank\" rel=\"noopener noreferrer\">Register<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> to join future Research Forum episodes and <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/researchforum.microsoft.com\/\" target=\"_blank\" rel=\"noopener noreferrer\">view previous sessions<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>. Transcripts and additional resources can be found in the Research Forum <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/story\/sep-2024-brief\/\">briefing book<\/a>.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"keynote-phi-3-vision-a-highly-capable-and-small-language-vision-model\">Keynote<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"phi-3-vision-a-highly-capable-and-small-language-vision-model\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/quarterly-brief\/sep-2024-brief\/articles\/keynote-phi-3-vision-a-highly-capable-and-small-language-vision-model\">Phi-3-Vision: A highly capable and \u201csmall\u201d language vision model<\/a><\/h3>\n\n\n\n<figure class=\"wp-block-image aligncenter size-full\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/quarterly-brief\/sep-2024-brief\/articles\/keynote-phi-3-vision-a-highly-capable-and-small-language-vision-model\"><img loading=\"lazy\" decoding=\"async\" width=\"1400\" height=\"788\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_Keynote_Jianfeng-Gao_1400x788.jpg\" alt=\"Research Forum | Episode 4 Keynote | Jianfeng Gao\" class=\"wp-image-1079760\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_Keynote_Jianfeng-Gao_1400x788.jpg 1400w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_Keynote_Jianfeng-Gao_1400x788-300x169.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_Keynote_Jianfeng-Gao_1400x788-1024x576.jpg 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_Keynote_Jianfeng-Gao_1400x788-768x432.jpg 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_Keynote_Jianfeng-Gao_1400x788-1066x600.jpg 1066w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_Keynote_Jianfeng-Gao_1400x788-655x368.jpg 655w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_Keynote_Jianfeng-Gao_1400x788-240x135.jpg 240w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_Keynote_Jianfeng-Gao_1400x788-640x360.jpg 640w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_Keynote_Jianfeng-Gao_1400x788-960x540.jpg 960w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_Keynote_Jianfeng-Gao_1400x788-1280x720.jpg 1280w\" sizes=\"auto, (max-width: 1400px) 100vw, 1400px\" \/><\/a><\/figure>\n\n\n\n<p>Jianfeng Gao introduced Phi-3-Vision, an advanced and economical open-source multimodal model. As a member of the Phi-3 model family, Phi-3-Vision enhances language models by integrating multisensory skills, seamlessly combining language and vision capabilities.<\/p>\n\n\n\n<blockquote class=\"wp-block-quote is-style-spectrum--blue-green is-layout-flow wp-block-quote-is-layout-flow\">\n<p>&#8220;Phi-3-Vision is the first multimodal model in the Phi small model family. It matches and sometimes exceeds some of the capabilities of much larger models \u2026 at a much lower cost. And to help everyone build more affordable and accessible AI systems, we have released the model weights into the open-source community.&#8221;<\/p>\n<cite><em>\u2014 <\/em><strong><em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/jfgao\/\">Jianfeng Gao<\/a><\/em><\/strong><em>, Distinguished Scientist and Vice President, Microsoft Research Redmond<\/em><\/cite><\/blockquote>\n\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity is-style-dots\"\/>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"panel-discussion-beyond-language-the-future-of-multimodal-models-in-healthcare-gaming-and-ai\">Panel Discussion<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"beyond-language-the-future-of-multimodal-models-in-healthcare-gaming-and-ai\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/quarterly-brief\/sep-2024-brief\/articles\/panel-discussion-beyond-language-the-future-of-multimodal-models-in-healthcare-gaming-and-ai\">Beyond language: The future of multimodal models in healthcare, gaming, and AI<\/a><\/h3>\n\n\n\n<figure class=\"wp-block-image size-large\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/quarterly-brief\/sep-2024-brief\/articles\/panel-discussion-beyond-language-the-future-of-multimodal-models-in-healthcare-gaming-and-ai\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"576\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_Panel_1400x788-1024x576.jpg\" alt=\"Research Forum | Episode 4 Panel | John Langford, Hoifung Poon, Katja Hofmann, Jianwei Yang\" class=\"wp-image-1079766\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_Panel_1400x788-1024x576.jpg 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_Panel_1400x788-300x169.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_Panel_1400x788-768x432.jpg 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_Panel_1400x788-1066x600.jpg 1066w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_Panel_1400x788-655x368.jpg 655w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_Panel_1400x788-240x135.jpg 240w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_Panel_1400x788-640x360.jpg 640w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_Panel_1400x788-960x540.jpg 960w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_Panel_1400x788-1280x720.jpg 1280w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_Panel_1400x788.jpg 1400w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/a><\/figure>\n\n\n\n<p>This discussion examined the transformative potential and core challenges of multimodal models across various domains, including precision health, game intelligence, and foundation models. Microsoft researchers <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/jcl\/\" target=\"_blank\" rel=\"noreferrer noopener\">John Langford<\/a>, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/hoifung\/\" target=\"_blank\" rel=\"noreferrer noopener\">Hoifung Poon<\/a>, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/kahofman\/\" target=\"_blank\" rel=\"noreferrer noopener\">Katja Hofmann<\/a>, and <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/jianwyan\/\" target=\"_blank\" rel=\"noreferrer noopener\">Jianwei Yang<\/a> shared their thoughts on future directions, bridging gaps, and fostering synergies within the field.\u202f<\/p>\n\n\n\n<p>\u201cOne of the really cutting-edge treatments for cancer these days is immunotherapy. That works by mobilizing the immune system to fight the cancer. And then one of the blockbuster drugs is a KEYTRUDA, that really can work miracles for some of the late- stage cancers &#8230; Unfortunately, only 20 to 30 percent of the patients actually respond. So that&#8217;s \u2026 a marquee example of what are the growth opportunity in precision health.\u201d<br><em><sub>\u2014 <\/sub><\/em><strong><sub><em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/hoifung\/\">Hoifung Poon<\/a><\/em><\/sub><\/strong><em><sub>, General Manager, Microsoft Research Health Futures<\/sub><\/em><\/p>\n\n\n\n<p>\u201cWe experience the world through vision, touch, and all our other senses before we start to make sense of any of the language that is spoken around us. So, it&#8217;s really, really interesting to think through the implications of that, and potentially, as we start to understand more about the different modalities that we can model and the different ways in which we combine them.\u201d<br><em><sub>\u2014 <\/sub><\/em><strong><sub><em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/kahofman\/\">Katja Hofmann<\/a><\/em><\/sub><\/strong><em><sub>, Senior Principal Researcher, Microsoft Research<\/sub><\/em><\/p>\n\n\n\n<p>\u201cTo really have a capable multimodal model, we need to encode different information from different modalities, for example, from vision, from language, from even audio, speech, etc. We need to develop a very capable encoder for each of these domains and then \u2026 tokenize each of these raw data.\u201d<br><em><sub>\u2014 <\/sub><\/em><strong><sub><em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/jianwyan\/\">Jianwei Yang<\/a><\/em><\/sub><\/strong><em><sub>, Principal Researcher, Microsoft Research Redmond<\/sub><\/em><\/p>\n\n\n\n<hr class=\"wp-block-separator has-text-color has-blue-color has-alpha-channel-opacity has-blue-background-color has-background is-style-dots\"\/>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"lightning-talks\">Lightning Talks<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"analog-optical-computing-for-sustainable-ai-and-beyond\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/quarterly-brief\/sep-2024-brief\/articles\/analog-optical-computing-for-sustainable-ai-and-beyond\">Analog optical computing for sustainable AI and beyond<\/a><\/h3>\n\n\n\n<figure class=\"wp-block-image size-large\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/quarterly-brief\/sep-2024-brief\/articles\/analog-optical-computing-for-sustainable-ai-and-beyond\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"576\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RF4_LT1_FrancescaP-JiaqiC-split_1400x788-1024x576.jpg\" alt=\"Research Forum | Episode 4 Talk 1 | Francesca Parmigiani and Jiaqi Chu\" class=\"wp-image-1081761\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RF4_LT1_FrancescaP-JiaqiC-split_1400x788-1024x576.jpg 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RF4_LT1_FrancescaP-JiaqiC-split_1400x788-300x169.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RF4_LT1_FrancescaP-JiaqiC-split_1400x788-768x432.jpg 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RF4_LT1_FrancescaP-JiaqiC-split_1400x788-1066x600.jpg 1066w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RF4_LT1_FrancescaP-JiaqiC-split_1400x788-655x368.jpg 655w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RF4_LT1_FrancescaP-JiaqiC-split_1400x788-240x135.jpg 240w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RF4_LT1_FrancescaP-JiaqiC-split_1400x788-640x360.jpg 640w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RF4_LT1_FrancescaP-JiaqiC-split_1400x788-960x540.jpg 960w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RF4_LT1_FrancescaP-JiaqiC-split_1400x788-1280x720.jpg 1280w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/09\/RF4_LT1_FrancescaP-JiaqiC-split_1400x788.jpg 1400w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/a><\/figure>\n\n\n\n<p>This talk presented a new kind of computer\u2014an analog optical computer\u2014that has the potential to accelerate AI inference and hard optimization workloads by 100x, leveraging hardware-software co-design to improve the efficiency and sustainability of real-world applications.\u202f<\/p>\n\n\n\n<p>\u201cMost likely, you or your loved ones have been inside an MRI scan <em>\u2014 <\/em>not really a great place to be in. Imagine if you can reduce that amount of time from 20 to 40 minutes to less than five minutes.\u201d<br><sub><em>\u2014 <\/em><strong><em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/frparmig\/\">Francesca Parmigiani<\/a><\/em><\/strong><em>, Principal Researcher, Microsoft Research Cambridge<\/em>\u202f<\/sub><\/p>\n\n\n\n<p>\u201cI&#8217;m really excited to share that we have just completed the second generation of [this] computer. It is much smaller in physical size, and this is a world first in that exactly the same computer is simultaneously solving hard optimization problems and accelerating machine learning inference. Looking ahead, we estimate that at scale, this computer can achieve around 450 tera operations per second per watt, which is a 100-times improvement as compared to state-of-the-art GPUs.\u201d<br><em><sub>\u2014 <\/sub><\/em><strong><sub><em><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/jiaqchu\/\">Jiaqi Chu<\/a><\/em><\/sub><\/strong><em><sub>, Principal Researcher, Microsoft Research Cambridge<\/sub><\/em><\/p>\n\n\n\n<hr class=\"wp-block-separator has-text-color has-blue-color has-alpha-channel-opacity has-blue-background-color has-background is-style-dots\"\/>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"analog-optical-computing-for-sustainable-ai-and-beyond\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/quarterly-brief\/sep-2024-brief\/articles\/direct-nash-optimization-teaching-language-models-to-self-improve-with-general-preferences\">Direct Nash Optimization: Teaching language models to self-improve with general preferences<\/a><\/h3>\n\n\n\n<figure class=\"wp-block-image size-large\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/quarterly-brief\/sep-2024-brief\/articles\/direct-nash-optimization-teaching-language-models-to-self-improve-with-general-preferences\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"576\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT2_Corby-Rosset_1400x788-1024x576.jpg\" alt=\"Research Forum | Episode 4 Talk 2 | Corby Rosset\" class=\"wp-image-1079778\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT2_Corby-Rosset_1400x788-1024x576.jpg 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT2_Corby-Rosset_1400x788-300x169.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT2_Corby-Rosset_1400x788-768x432.jpg 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT2_Corby-Rosset_1400x788-1066x600.jpg 1066w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT2_Corby-Rosset_1400x788-655x368.jpg 655w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT2_Corby-Rosset_1400x788-240x135.jpg 240w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT2_Corby-Rosset_1400x788-640x360.jpg 640w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT2_Corby-Rosset_1400x788-960x540.jpg 960w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT2_Corby-Rosset_1400x788-1280x720.jpg 1280w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT2_Corby-Rosset_1400x788.jpg 1400w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/a><\/figure>\n\n\n\n<p>This talk explored teaching language models to self-improve using AI preference feedback, challenging the model to play against itself and a powerful teacher until it arrives at a Nash equilibrium, resulting in state-of-the-art win rates against GPT-4 Turbo on benchmarks such as AlpacaEval and MT-Bench.\u202f<\/p>\n\n\n\n<p>\u201cThe traditional way to fine-tune an LLM for post-training \u2026 basically tells the model to emulate good behaviors, but it does not target or correct any mistakes or bad behaviors that it makes explicitly. \u2026 Self-improving post-training explicitly identifies and tries to correct bad behaviors or mistakes that the model makes.\u201d<br><sub>\u2014 <em><strong><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/corbyrosset\/\">Corby Rosset<\/a><\/strong>, Senior Researcher, Microsoft Research AI Frontiers<\/em><\/sub><\/p>\n\n\n\n<hr class=\"wp-block-separator has-text-color has-blue-color has-alpha-channel-opacity has-blue-background-color has-background is-style-dots\"\/>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"analog-optical-computing-for-sustainable-ai-and-beyond\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/quarterly-brief\/sep-2024-brief\/articles\/project-aurora-the-first-large-scale-foundation-model-of-the-atmosphere\">Project Aurora: The first large-scale foundation model of the atmosphere<\/a><\/h3>\n\n\n\n<figure class=\"wp-block-image size-large\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/quarterly-brief\/sep-2024-brief\/articles\/project-aurora-the-first-large-scale-foundation-model-of-the-atmosphere\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"576\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT3_Megan-Stanley_1400x788-1024x576.jpg\" alt=\"Research Forum | Episode 4 Talk 3 | Megan Stanley\" class=\"wp-image-1079784\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT3_Megan-Stanley_1400x788-1024x576.jpg 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT3_Megan-Stanley_1400x788-300x169.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT3_Megan-Stanley_1400x788-768x432.jpg 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT3_Megan-Stanley_1400x788-1066x600.jpg 1066w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT3_Megan-Stanley_1400x788-655x368.jpg 655w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT3_Megan-Stanley_1400x788-240x135.jpg 240w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT3_Megan-Stanley_1400x788-640x360.jpg 640w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT3_Megan-Stanley_1400x788-960x540.jpg 960w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT3_Megan-Stanley_1400x788-1280x720.jpg 1280w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT3_Megan-Stanley_1400x788.jpg 1400w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/a><\/figure>\n\n\n\n<p>This talk presented Aurora, a cutting-edge foundation model that offers a new approach to weather forecasting that could transform our ability to predict and mitigate the impacts of extreme events, air pollution, and the changing climate.<\/p>\n\n\n\n<p>\u201cIf we look at Aurora&#8217;s ability to predict pollutants such as nitrogen dioxide that are strongly related to emissions from human activity, we can see that the model has learned to make these predictions with no emissions data provided. It&#8217;s learned the implicit patterns that cause the gas concentrations, which is very impressive.\u201d<br><sub>\u2014 <em><strong><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/meganstanley\/\">Megan Stanley<\/a><\/strong>, Senior Researcher, Microsoft Research AI for Science<\/em><\/sub><\/p>\n\n\n\n<hr class=\"wp-block-separator has-text-color has-blue-color has-alpha-channel-opacity has-blue-background-color has-background is-style-dots\"\/>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"analog-optical-computing-for-sustainable-ai-and-beyond\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/quarterly-brief\/sep-2024-brief\/articles\/a-generative-model-of-biology-for-in-silico-experimentation-and-discovery\">A generative model of biology for in-silico experimentation and discovery<\/a><\/h3>\n\n\n\n<figure class=\"wp-block-image size-large\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/quarterly-brief\/sep-2024-brief\/articles\/a-generative-model-of-biology-for-in-silico-experimentation-and-discovery\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"576\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT4_Kevin-Yang_1400x788-1024x576.jpg\" alt=\"Research Forum | Episode 4 Talk 4 | Kevin Yang\" class=\"wp-image-1079790\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT4_Kevin-Yang_1400x788-1024x576.jpg 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT4_Kevin-Yang_1400x788-300x169.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT4_Kevin-Yang_1400x788-768x432.jpg 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT4_Kevin-Yang_1400x788-1066x600.jpg 1066w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT4_Kevin-Yang_1400x788-655x368.jpg 655w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT4_Kevin-Yang_1400x788-240x135.jpg 240w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT4_Kevin-Yang_1400x788-640x360.jpg 640w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT4_Kevin-Yang_1400x788-960x540.jpg 960w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT4_Kevin-Yang_1400x788-1280x720.jpg 1280w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT4_Kevin-Yang_1400x788.jpg 1400w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/a><\/figure>\n\n\n\n<p>This talk explored how deep learning enables generation of novel and useful biomolecules, allowing researchers and practitioners to better understand biology.\u202fThis includes EvoDiff, a general-purpose diffusion framework that combines evolutionary-scale data with the distinct conditioning capabilities of diffusion models to generate new proteins, given a protein sequence.<\/p>\n\n\n\n<p>\u201cOften, protein engineers want proteins that perform a similar function to a natural protein, or they want to produce a protein that performs the same function but has other desirable properties, such as stability. By conditioning EvoDiff with a family of related sequences, we can generate new proteins that are very different in sequence space to the natural proteins but are predicted to fold into similar three-dimensional structures. These may be good starting points for finding new functions or for discovering versions of a protein with desirable properties.\u201d<br><sub>\u2014 <em><strong><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/kevyan\/\">Kevin Yang<\/a><\/strong>, Senior Researcher, Microsoft Research New England<\/em><\/sub><\/p>\n\n\n\n<hr class=\"wp-block-separator has-text-color has-blue-color has-alpha-channel-opacity has-blue-background-color has-background is-style-dots\"\/>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"analog-optical-computing-for-sustainable-ai-and-beyond\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/quarterly-brief\/sep-2024-brief\/articles\/fostering-appropriate-reliance-on-ai\">Fostering appropriate reliance on AI<\/a><\/h3>\n\n\n\n<figure class=\"wp-block-image size-large\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/quarterly-brief\/sep-2024-brief\/articles\/fostering-appropriate-reliance-on-ai\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"576\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT5_Mihaela-Vorvoreanu_1400x788-1024x576.jpg\" alt=\"Research Forum | Episode 4 Talk 5 | Mihaela Vorvoreanu\" class=\"wp-image-1079799\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT5_Mihaela-Vorvoreanu_1400x788-1024x576.jpg 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT5_Mihaela-Vorvoreanu_1400x788-300x169.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT5_Mihaela-Vorvoreanu_1400x788-768x432.jpg 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT5_Mihaela-Vorvoreanu_1400x788-1066x600.jpg 1066w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT5_Mihaela-Vorvoreanu_1400x788-655x368.jpg 655w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT5_Mihaela-Vorvoreanu_1400x788-240x135.jpg 240w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT5_Mihaela-Vorvoreanu_1400x788-640x360.jpg 640w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT5_Mihaela-Vorvoreanu_1400x788-960x540.jpg 960w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT5_Mihaela-Vorvoreanu_1400x788-1280x720.jpg 1280w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/RF4_LT5_Mihaela-Vorvoreanu_1400x788.jpg 1400w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/a><\/figure>\n\n\n\n<p>Since AI systems are probabilistic, they can make mistakes. One of the main challenges in human-AI interaction is to avoid overreliance on AI and empower people to determine when to accept or not accept an AI system&#8217;s recommendation. This talk explores Microsoft\u2019s work in this area.<\/p>\n\n\n\n<p>\u201cThis is where I think it is our responsibility as people working in UX disciplines\u2014as people researching UX and human-computer interaction\u2014to really, really step up to the front and see how it is our moment to shine and to address this problem.\u201d<br><sub>\u2014 <em><strong><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/mivorvor\/\">Mihaela Vorvoreanu<\/a><\/strong>, Director UX Research and Responsible AI Education, Microsoft AI Ethics and Effects in Engineering and Research (Aether)<\/em><\/sub><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Explore multimodal & small language models, plus advanced benchmarks for AI evaluation. Microsoft researchers are working on breakthroughs in weather prediction, materials design, even a new kind of computer for AI inference and hard optimization problems.<\/p>\n","protected":false},"author":42735,"featured_media":1079448,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"msr-url-field":"","msr-podcast-episode":"","msrModifiedDate":"","msrModifiedDateEnabled":false,"ep_exclude_from_search":false,"_classifai_error":"","msr-author-ordering":[{"type":"user_nicename","value":"Jianfeng Gao","user_id":"32246"},{"type":"user_nicename","value":"John Langford","user_id":"32204"},{"type":"user_nicename","value":"Hoifung Poon","user_id":"32016"},{"type":"user_nicename","value":"Katja Hofmann","user_id":"32468"},{"type":"user_nicename","value":"Jianwei Yang","user_id":"40261"},{"type":"user_nicename","value":"Corby Rosset","user_id":"41997"},{"type":"user_nicename","value":"Megan Stanley","user_id":"41482"},{"type":"user_nicename","value":"Kevin Kaichuang Yang","user_id":"39093"},{"type":"user_nicename","value":"Mihaela Vorvoreanu","user_id":"36804"}],"msr_hide_image_in_river":null,"footnotes":""},"categories":[1],"tags":[],"research-area":[13556],"msr-region":[],"msr-event-type":[],"msr-locale":[268875],"msr-post-option":[269148,243984,269142],"msr-impact-theme":[],"msr-promo-type":[],"msr-podcast-series":[],"class_list":["post-1081755","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-research-blog","msr-research-area-artificial-intelligence","msr-locale-en_us","msr-post-option-approved-for-river","msr-post-option-blog-homepage-featured","msr-post-option-include-in-river"],"msr_event_details":{"start":"","end":"","location":""},"podcast_url":"","podcast_episode":"","msr_research_lab":[199561,199563,199565,849856,851467,992148],"msr_impact_theme":[],"related-publications":[],"related-downloads":[],"related-videos":[],"related-academic-programs":[],"related-groups":[],"related-projects":[],"related-events":[],"related-researchers":[{"type":"user_nicename","value":"Jianfeng Gao","user_id":32246,"display_name":"Jianfeng Gao","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/jfgao\/\" aria-label=\"Visit the profile page for Jianfeng Gao\">Jianfeng Gao<\/a>","is_active":false,"last_first":"Gao, Jianfeng","people_section":0,"alias":"jfgao"},{"type":"user_nicename","value":"John Langford","user_id":32204,"display_name":"John Langford","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/jcl\/\" aria-label=\"Visit the profile page for John Langford\">John Langford<\/a>","is_active":false,"last_first":"Langford, John","people_section":0,"alias":"jcl"},{"type":"user_nicename","value":"Hoifung Poon","user_id":32016,"display_name":"Hoifung Poon","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/hoifung\/\" aria-label=\"Visit the profile page for Hoifung Poon\">Hoifung Poon<\/a>","is_active":false,"last_first":"Poon, Hoifung","people_section":0,"alias":"hoifung"},{"type":"user_nicename","value":"Katja Hofmann","user_id":32468,"display_name":"Katja Hofmann","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/kahofman\/\" aria-label=\"Visit the profile page for Katja Hofmann\">Katja Hofmann<\/a>","is_active":false,"last_first":"Hofmann, Katja","people_section":0,"alias":"kahofman"},{"type":"user_nicename","value":"Corby Rosset","user_id":41997,"display_name":"Corby Rosset","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/corbyrosset\/\" aria-label=\"Visit the profile page for Corby Rosset\">Corby Rosset<\/a>","is_active":false,"last_first":"Rosset, Corby","people_section":0,"alias":"corbyrosset"},{"type":"user_nicename","value":"Kevin Kaichuang Yang","user_id":39093,"display_name":"Kevin Kaichuang Yang","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/kevyan\/\" aria-label=\"Visit the profile page for Kevin Kaichuang Yang\">Kevin Kaichuang Yang<\/a>","is_active":false,"last_first":"Yang, Kevin Kaichuang","people_section":0,"alias":"kevyan"},{"type":"user_nicename","value":"Mihaela Vorvoreanu","user_id":36804,"display_name":"Mihaela Vorvoreanu","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/mivorvor\/\" aria-label=\"Visit the profile page for Mihaela Vorvoreanu\">Mihaela Vorvoreanu<\/a>","is_active":false,"last_first":"Vorvoreanu, Mihaela","people_section":0,"alias":"mivorvor"}],"msr_type":"Post","featured_image_thumbnail":"<img width=\"960\" height=\"540\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/Forum-Ep4-cover_multimodality_holasoyka_1400x788-960x540.jpg\" class=\"img-object-cover\" alt=\"Research Forum | Episode 4 - abstract chalkboard background with colorful network nodes and circular icons\" decoding=\"async\" loading=\"lazy\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/Forum-Ep4-cover_multimodality_holasoyka_1400x788-960x540.jpg 960w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/Forum-Ep4-cover_multimodality_holasoyka_1400x788-300x169.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/Forum-Ep4-cover_multimodality_holasoyka_1400x788-1024x576.jpg 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/Forum-Ep4-cover_multimodality_holasoyka_1400x788-768x432.jpg 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/Forum-Ep4-cover_multimodality_holasoyka_1400x788-1066x600.jpg 1066w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/Forum-Ep4-cover_multimodality_holasoyka_1400x788-655x368.jpg 655w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/Forum-Ep4-cover_multimodality_holasoyka_1400x788-240x135.jpg 240w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/Forum-Ep4-cover_multimodality_holasoyka_1400x788-640x360.jpg 640w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/Forum-Ep4-cover_multimodality_holasoyka_1400x788-1280x720.jpg 1280w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/08\/Forum-Ep4-cover_multimodality_holasoyka_1400x788.jpg 1400w\" sizes=\"auto, (max-width: 960px) 100vw, 960px\" \/>","byline":"","formattedDate":"September 26, 2024","formattedExcerpt":"Explore multimodal &amp; small language models, plus advanced benchmarks for AI evaluation. Microsoft researchers are working on breakthroughs in weather prediction, materials design, even a new kind of computer for AI inference and hard optimization problems.","locale":{"slug":"en_us","name":"English","native":"","english":"English"},"_links":{"self":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts\/1081755","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/users\/42735"}],"replies":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/comments?post=1081755"}],"version-history":[{"count":50,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts\/1081755\/revisions"}],"predecessor-version":[{"id":1120479,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts\/1081755\/revisions\/1120479"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media\/1079448"}],"wp:attachment":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media?parent=1081755"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/categories?post=1081755"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/tags?post=1081755"},{"taxonomy":"msr-research-area","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/research-area?post=1081755"},{"taxonomy":"msr-region","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-region?post=1081755"},{"taxonomy":"msr-event-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event-type?post=1081755"},{"taxonomy":"msr-locale","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-locale?post=1081755"},{"taxonomy":"msr-post-option","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-post-option?post=1081755"},{"taxonomy":"msr-impact-theme","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-impact-theme?post=1081755"},{"taxonomy":"msr-promo-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-promo-type?post=1081755"},{"taxonomy":"msr-podcast-series","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-podcast-series?post=1081755"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}