{"id":644958,"date":"2020-03-25T03:00:18","date_gmt":"2020-03-25T10:00:18","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/?p=644958"},"modified":"2020-06-18T07:32:26","modified_gmt":"2020-06-18T14:32:26","slug":"microsofts-ai-transformation-project-turing-and-smarter-search-with-rangan-majumder","status":"publish","type":"post","link":"https:\/\/www.microsoft.com\/en-us\/research\/podcast\/microsofts-ai-transformation-project-turing-and-smarter-search-with-rangan-majumder\/","title":{"rendered":"Microsoft\u2019s AI Transformation, Project Turing and smarter search with Rangan Majumder"},"content":{"rendered":"<h3><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-large wp-image-644961\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/03\/Research_Podcast_Majumder_Site_1400x788-1024x576.png\" alt=\"Headshot of Rangan Majumder for the Microsoft Research Podcast\" width=\"1024\" height=\"576\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/03\/Research_Podcast_Majumder_Site_1400x788-1024x576.png 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/03\/Research_Podcast_Majumder_Site_1400x788-300x169.png 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/03\/Research_Podcast_Majumder_Site_1400x788-768x432.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/03\/Research_Podcast_Majumder_Site_1400x788-1066x600.png 1066w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/03\/Research_Podcast_Majumder_Site_1400x788-655x368.png 655w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/03\/Research_Podcast_Majumder_Site_1400x788-343x193.png 343w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/03\/Research_Podcast_Majumder_Site_1400x788-640x360.png 640w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/03\/Research_Podcast_Majumder_Site_1400x788-960x540.png 960w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/03\/Research_Podcast_Majumder_Site_1400x788-1280x720.png 1280w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/03\/Research_Podcast_Majumder_Site_1400x788.png 1400w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/h3>\n<h3>Episode 112 | March 25, 2020<\/h3>\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/ranganm\/\">Rangan Majumder<\/a> is the Partner Group Program Manager of Microsoft\u2019s Search and AI, and he has a simple goal: to make the world smarter and more productive. But nobody said simple was easy, so he and his team are working on better \u2013 and faster \u2013 ways to help you find the information you\u2019re looking for, anywhere you\u2019re looking for it.<\/p>\n<p>Today, Rangan talks about how three big trends have changed the way Microsoft is building \u2013 and sharing \u2013 AI stacks across product groups. He also tells us about <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/msturing.org\/\" target=\"_blank\" rel=\"noopener noreferrer\">Project Turing<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, an internal deep learning moonshot that aims to harness the resources of the web and bring the power of deep learning to a search box near you.<\/p>\n<h3>Related:<\/h3>\n<ul type=\"disc\">\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"https:\/\/www.bing.com\/visualsearch\/\" rel=\"noopener noreferrer\">Visual Search page<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"https:\/\/blogs.microsoft.com\/ai\/build-2018-project-brainwave\/\" rel=\"noopener noreferrer\">Brainwave blogpost<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"https:\/\/youtu.be\/V20DKWr0Ps0\" rel=\"noopener noreferrer\">Brainwave in Bing video<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"https:\/\/blogs.microsoft.com\/ai\/microsoft-creates-ai-can-read-document-answer-questions-well-person\/\" rel=\"noopener noreferrer\">Microsoft reaches human parity on SQuAD Challenge<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/blog\/machine-reading-systems-are-becoming-more-conversational\/\">Microsoft reaches human parity on Conversation QA<\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"https:\/\/docs.microsoft.com\/en-us\/archive\/blogs\/stevengu\/microsoft-achieves-human-performance-estimate-on-glue-benchmark\" rel=\"noopener noreferrer\">Microsoft achieves human parity on GLUE<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/blogs.microsoft.com\/ai\/bing-vector-search\/\">Open sourced vector search algorithm<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/blogs.microsoft.com\/ai\/mona-lisa-translation-research-products\/\">Multi-task learning for search and question answering<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" target=\"_blank\" href=\"https:\/\/cloudblogs.microsoft.com\/opensource\/2020\/01\/21\/microsoft-onnx-open-source-optimizations-transformer-inference-gpu-cpu\/\" rel=\"noopener noreferrer\">Fast Transformer inferencing in ONNX Runtime<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/blog\/turing-nlg-a-17-billion-parameter-language-model-by-microsoft\/\">Turing NLG Announcement<\/a><\/li>\n<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/blog\/zero-deepspeed-new-system-optimizations-enable-training-models-with-over-100-billion-parameters\/\">DeepSpeed Open Source Announcement<\/a><\/li>\n<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/podcast\">Microsoft Research Podcast<\/a>: View more podcasts on Microsoft.com<\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/itunes.apple.com\/us\/podcast\/microsoft-research-a-podcast\/id1318021537?mt=2\">iTunes<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>: Subscribe and listen to new podcasts each week on iTunes<\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/subscribebyemail.com\/www.blubrry.com\/feeds\/microsoftresearch.xml\">Email<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>: Subscribe and listen by email<\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/subscribeonandroid.com\/www.blubrry.com\/feeds\/microsoftresearch.xml\">Android<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>: Subscribe and listen on Android<\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/open.spotify.com\/show\/4ndjUXyL0hH1FXHgwIiTWU\">Spotify<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>: Listen on Spotify<\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/www.blubrry.com\/feeds\/microsoftresearch.xml\">RSS feed<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/note.microsoft.com\/ww-registration-microsoft-research-newsletter-s.html?wt.mc_id=S-webpage_podcast\">Microsoft Research Newsletter<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>: Sign up to receive the latest news from Microsoft Research<\/li>\n<\/ul>\n<hr \/>\n<h3>Transcript<\/h3>\n<p>Rangan\u00a0Majumder:\u00a0At the time, deep learning was really impressive in terms of these perception tasks like vision,\u00a0you know,\u00a0speech\u2026\u00a0so we were thinking,\u00a0like,\u00a0could it be really good at these other higher level tasks like language? So, that\u2019s when we started Project Turing,\u00a0and the idea was, what if we could do,\u00a0like,\u00a0end-to-end deep learning across the entire web to be able to answer these questions?<\/p>\n<p><b>Host:\u00a0<\/b><b>You\u2019re listening to the Microsoft Research Podcast, a show that brings you closer to the cutting-edge of technology research and the scientists behind it. I\u2019m your host, Gretchen Huizinga.<\/b><\/p>\n<p><b>Host:\u00a0<\/b><b>Rangan<\/b><b>\u00a0Majumder is the Partner Group Program Manager of Microsoft\u2019s Search and AI, and he has a simple goal: to make the world smarter and more productive. But nobody said simple was easy, so he and his team are working on better \u2013 and faster \u2013 ways to help you find the information you\u2019re looking for, anywhere you\u2019re looking for it.<\/b><\/p>\n<p><b>Today,\u00a0<\/b><b>Rangan<\/b><b>\u00a0talks about how three big trends have changed the way Microsoft is building \u2013 and sharing \u2013 AI stacks across product groups. He also tells us about Project Turing, an internal deep learning moonshot that aims to harness the resources of\u00a0<\/b><b>the\u00a0<\/b><b>web and bring the power of deep learning to a search box near you.<\/b><b>\u00a0<\/b><b>T<\/b><b>hat and much more on this episode of the Microsoft Research Podcast.<\/b><\/p>\n<p><b>Host:\u00a0<\/b><b>Rangan<\/b><b>\u00a0Majumder, welcome to the podcast.<\/b><\/p>\n<p>Rangan\u00a0Majumder: Thank you. It\u2019s great to be here.<\/p>\n<p><b>Host: So you\u2019re a different kind of guest here in the booth. You\u2019re a Partner Group Program Manager over in Search and AI at Microsof<\/b><b>t<\/b><b>. Let\u2019s start by situating your group and its work since you\u2019re not in Microsoft Research per se, but you do a lot of work with the folks here. How and where do you \u201croll up\u201d as they say?<\/b><\/p>\n<p>Rangan\u00a0Majumder: Yeah, great question. So, as you know, the broader organization is called Microsoft AI and Research, and Microsoft Research is one of the sub-groups there. So another sister team of Microsoft Research is the Bing Search team,\u00a0and my group is actually Search and AI, which is inside of Bing. So we\u2019re a sister team to Microsoft Research. And it\u2019s really great to be on this team because what we get to do is work closely with Microsoft researchers and then productionize some of their great research efforts\u2026<\/p>\n<p><b>Host: Yeah.<\/b><\/p>\n<p>Rangan\u00a0Majumder: \u2026put it into production, and then, once we get it to work at scale in Bing, we can actually go take that technology and place it elsewhere, like in Office and Dynamics and other parts of Microsoft.<\/p>\n<p><b>Host: Right. So I\u2019m getting the visual of those nesting dolls with each part going inside the other. So<\/b><b>\u00a0<\/b><b>top big doll is Microsoft AI and Research.<\/b><\/p>\n<p>Rangan\u00a0Majumder: Correct.<\/p>\n<p><b>Host: And then Microsoft Research is part of that.<\/b><\/p>\n<p>Rangan\u00a0Majumder: Right.<\/p>\n<p><b>Host: And Bing is part of that.<\/b><\/p>\n<p>Rangan\u00a0Majumder: That\u2019s right.<\/p>\n<p><b>Host: And then your group, Search and AI, is nested within the Bing group.<\/b><\/p>\n<p>Rangan\u00a0Majumder: That\u2019s correct.<\/p>\n<p><b>Host: Okay, and who do you all roll up to?<\/b><\/p>\n<p>Rangan\u00a0Majumder: Kevin Scott, who is our CTO\u2026<\/p>\n<p><b>Host: Okay.<\/b><\/p>\n<p>Rangan\u00a0Majumder: \u2026and also EVP.<\/p>\n<p><b>Host: Got it.\u00a0<\/b><b>Al<\/b><b>l\u00a0<\/b><b>right,\u00a0<\/b><b>well let\u2019s talk about what you all do in Search and AI and now that we\u2019ve situated you<\/b><b>.<\/b><b>\u00a0<\/b><b>Y<\/b><b>ou have a delightfully short, but incredibly ambitious mission statement. Tell us what it is and, if you can, what\u2019s your high-level strategy for making it real?<\/b><\/p>\n<p>Rangan\u00a0Majumder: Yeah, so\u00a0our mission statement is to make the world smarter and more productive. And you\u2019ll notice that our mission statement doesn\u2019t just talk about search because search is obviously the biggest thing that we do, but it\u2019s important to understand what is the underlying user need for why people are searching, and it\u2019s really to learn about something or to get something done, right? So people want to learn a lot about what\u2019s happening with the coronavirus today. So that\u2019s an example of how our technology helps make people smarter so they can know what\u2019s going on in the world. An example of where we\u2019re helping make people productive is something like when, you know, I got my sink clogged, right? So that\u2019s something I just want to learn really quickly like, how do I unclog my sink? So that\u2019s an example of\u00a0productivity. So, the reason you need to understand the underlying user need versus like how they do it today is, the solutions actually change over time.<\/p>\n<p><b>Host: Okay.<\/b><\/p>\n<p>Rangan\u00a0Majumder: So, we want to be really close to, what is the user need that people have?\u00a0And then our technology helps provide that and satisfy that need. As I said, the mission is about making the world smarter and more productive. If we just focused on the users on Bing, we can still have a lot of impact.\u00a0But if you look at the entire pie of customers from Microsoft, there\u2019s a lot more we can do. So that\u2019s where we\u2019ve been working a lot with Office, taking our AI technology and not just bringing it to Bing, but bringing it to Office. So\u00a0that\u2019s an example where we increase the number of people we can impact by, like, a billion.<\/p>\n<p><b>Host: Right.<\/b><\/p>\n<p>Rangan\u00a0Majumder: Because there\u2019s a lot more users using Word. And then, if you think about, like, Azure becoming, you know, the world\u2019s computer. So there\u2019s a lot more impact we could have by bringing our technology into Azure as well.<\/p>\n<p><b>Host: Well let\u2019s talk about what gets you up in the morning. In your role as a Partner Group Program Manager for Search and AI, do you have a personal mission, or a personal passion, is maybe a better way to put it?<\/b><\/p>\n<p>Rangan\u00a0Majumder: Yeah, well as any program manager, your goal is really to maximize product\/market fit, but my personal mission is basically the same as my team\u2019s mission, which is really around making the world smarter and more productive, and if you just look at what\u2019s happening today, which is, people are finding new ways to look for information, right? Like ten years ago was all about search. Like, people just kind of typed in words. But now people want to find stuff more naturally like smart assistants,\u00a0people just want to ask a question, you know, in an ambient room and get the answer.<\/p>\n<p><b>Host: Mm-hmm.<\/b><\/p>\n<p>Rangan\u00a0Majumder: People want to be able to take a picture of a flower and say, hey, what is this flower? How do I take care of this?\u00a0Then the amount of information is changing too, so people aren\u2019t just writing web pages like they were ten years ago. People are now taking pictures, uploading photos, uploading videos. So going back to my example of, you know, how do I unclog a sink? You don\u2019t just want a web page walking through the steps. Sometimes you want a video\u2026<\/p>\n<p><b>Host: Right.<\/b><\/p>\n<p>Rangan\u00a0Majumder: \u2026that just shows you, hey, here\u2019s how I unclog a sink. So I think there\u2019s just a lot to do in that mission, and something that I feel like we\u2019ll be doing for like easily a decade or more.<\/p>\n<p><b>Host: You know, as you brought up the flower and taking a picture of it, I\u2019m thinking of this music app that I use, Shazam,\u00a0<\/b><b>where you find out what a song<\/b><b>\u00a0is.\u00a0<\/b><b>I\u2019ve often said I want Shazam for all these different categories. What\u2019s that tree? I don\u2019t even know what it is, but if I take a picture of it could you tell me? Are you guys working on stuff like that too?<\/b><\/p>\n<p>Rangan\u00a0Majumder:\u00a0Uh, we\u2019ve actually shipped it already! So if you go install the Bing app you can actually go take a\u2026 I\u2019ve done this when I moved into my new house, like there were these flowers and I\u2019m like what are these flowers? They look really interesting. I could take a picture of it and it tells you what it is, and then you can find out more information. So plants, dogs, those kinds of things, the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/www.bing.com\/visualsearch\/\" target=\"_blank\" rel=\"noopener noreferrer\">Bing Visual Search app<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> does really well so, go install it today and try it out!<\/p>\n<p><b>Host: Well, much of what we call AI is still work in progress<\/b><b>,<\/b><b>\u00a0and there are some fundamentally different ways of doing AI within a company, especially one as large as Microsoft, so give us a snapshot of how product groups have traditionally approached building AI stacks and then tell us some of the big trends that you\u2019ve noted in the science of AI that have enabled a disruption in that approach.<\/b><\/p>\n<p>Rangan\u00a0Majumder: I think this is probably the most exciting thing happening at Microsoft today.\u00a0So the way we\u2019re doing AI is definitely transforming.\u00a0If you think about how we used to do AI, maybe five years ago,\u00a0we\u00a0would have multiple different product groups doing AI kind of independently, and for the most part didn\u2019t share anything. But there\u2019s three trends that have really been changing that. The first trend is really around\u00a0transfer learning, which is this concept that, as you train a model on one type of data and one set of tasks, you can actually reuse that model for other tasks and it does, sometimes, even better than it would if you just trained it on that task specifically.<\/p>\n<p><b>Host: Huh.<\/b><\/p>\n<p>Rangan\u00a0Majumder: The second one that\u2019s happening is this trend with\u00a0large pre-trained models.\u00a0I\u00a0think there\u2019s a couple of these out there, right, like you probably heard about Open AI\u2019s GPT, Google has BERT, Microsoft has\u00a0its\u00a0MT-DNN. So you can take these models and just train them on a bunch of data in a self-supervised way, it\u2019s called, make it very large, and then you can actually apply it on lots of other tasks and it just does\u00a0phenomenal. Just to give you an example, like, let\u2019s say the Search team was about a hundred people and they\u2019re working on various parts of\u00a0search all the time so what we did is take about ten folks and said, okay, I want you guys to look at these large transformer networks and see what kind of impact could you have. So in just, like, a few months they were able to ship an improvement so large that it was larger than all the other, like, ninety folks, all the work they did, combined. So we were just, like, shocked how important and how impactful this kind of work was.<\/p>\n<p><b>Host: Right.<\/b><\/p>\n<p>Rangan\u00a0Majumder: So much so that, at first, we thought, well, does that mean we don\u2019t need these other ninety folks? We can just work on these ten folks?\u00a0But instead we really embraced it and we said well, let\u2019s get all hundred folks working on these large transformer networks. And then, in the end, like we just had a wave of improvements over the last six months of just like improvement after improvement equally as impactful as the one we had before,\u00a0so this is a really big trend right now in these large pre-trained models.<\/p>\n<p><b>Host: Okay.<\/b><\/p>\n<p>Rangan\u00a0Majumder: The third trend is really around the culture of Microsoft and how it\u2019s changing. And this really started with Satya when he became CEO. He really\u00a0has been focused on changing the culture and making it a lot more\u00a0collaborative. In fact, he\u2019s changed incentive structure in the team, so when you\u2019re actually going through a performance review, it\u2019s not just about, you know, what did you do? But it\u2019s about, how did you use someone else\u2019s work or how did you contribute to someone else\u2019s work? The other, like,\u00a0person\u00a0who\u2019s really changed a lot is Kevin Scott, our CTO. So he did a bunch of AI reviews and realized like there\u2019s a lot of teams doing similar stuff, but some teams are a little bit better than others. So why don\u2019t we do this work in a coordinated way? So when you take those three trends together, what we\u2019re doing is, we\u2019re starting to build this coordinated AI stack across Microsoft where we have certain teams saying, look, we are going to build these really large NLP models for the company, not just ourselves, because the problem is, if each team tried to do that, it would be just way too costly, and then, through transfer learning, I can now reuse this model in other parts. So the stack is kind of looking like this: at the very top you have applications like Bing, the different Office apps, you know, Dynamics, Azure Cognitive Services. The layer underneath is a bunch of these pre-trained models. Like we have one called the Turing Neural Language Representation, we\u2019ve got Language Generation, we\u2019ve got these vision models\u2026 The layer underneath is these software systems, which can actually run these models really, really fast because the models are very big and they\u2019re very expensive.<\/p>\n<p><b>Host: Yeah.<\/b><\/p>\n<p>Rangan Majumder: So if you, if you run them in a na\u00efve way, it would just, like, take too long and you\u2019d hurt the customer experience so you need to actually do a lot of software optimizations. And then the final layer is around the hardware, so that\u2019s around like CPU, GPUs and we even have our own little effort on chips with FPGAs.<\/p>\n<p><b><i>(music plays)<\/i><\/b><\/p>\n<p><b>Host: I want to talk a little bit about four big areas you\u2019ve identified as important to progress and innovation in Search and AI and you\u2019ve sort of labeled them web search,\u00a0<\/b><b>question answering, multi-media, and platform. So why are each of these areas important, especially as it relates to the customer experience<\/b><b>,<\/b><b>\u00a0and what innovations are you exploring as you seek to improve that experience?<\/b><\/p>\n<p>Rangan\u00a0Majumder: Yeah, so I would say, about five years ago these things seemed pretty different. Like web search, question answering, multi-media and then the platform team would sort of support all those teams. I\u2019ve noticed, and now you\u2019ll see it more and more, that these experiences are very integrated. So if you go to Bing today and you search for, what do alligators eat?\u00a0You\u2019ll see, at the very top, an answer that says, you know, alligators eat things like fish, turtles, birds\u2026 but then you\u2019ll also see an image there, sort of fused in with that answer, because an image actually helps you really get the emotional part. So just reading it is one thing, but humans also need that emotional experience, so by showing that image right next to the answer, it just makes the answer come to life.<\/p>\n<p><b>Host: Right.<\/b><\/p>\n<p>Rangan\u00a0Majumder: So that\u2019s one way where these things are kind of related. Like the experience, putting them all together, makes it much better for the customer\u2026<\/p>\n<p><b>Host: Right.<\/b><\/p>\n<p>Rangan\u00a0Majumder: \u2026but also the technology stacks are becoming very similar too, especially with deep learning. So with deep learning, it\u2019s mostly operating on vectors. So the first step in all of these systems, whether it\u2019s question answering, web search and multi-media, is really taking this corpus of information and converting it to vectors using an encoder. So that part is pretty different for each one, but then, once you have this corpus of vectors, the rest of the stack is very similar. Like the first thing you do when a query comes in is, you do a vector search to say, all right, what are the most similar vectors here? And then you run a cascade of different deep learning models, and each one gets heavier and a little bit more costly, and that\u2019s what\u2019s been super interesting, where, before, each team had its own very different stack, but with deep learning and everything just betting on this\u00a0vectors,\u00a0there\u2019s just a few services I need to build really, really well. One is this inference service which is, you know, given some content, vectorize it really quick. The other one is this vector search service which is, given a set of vectors how do I search them extremely fast?<\/p>\n<p><b>Host: Your team has been involved in achieving several impressive milestones over the past five years. So take us on a little guided tour of that timeline and tell us about the challenges you face, along with the rewards that you reap when you try to bring research milestones into production.<\/b><\/p>\n<p>Rangan\u00a0Majumder: So first, I think a lot of the milestones,\u00a0I have to give most of the credit to Microsoft Research because they\u2019re the ones really leading the way on pushing the state-of-the-art on those benchmarks. Like our team doesn\u2019t really focus too much on the academic\u00a0benchmarks. So, ever since we went on this mission of, let\u2019s really push deep learning for NLP, the first academic data set that came out that was\u00a0really aligned with that mission was\u00a0by Stanford called the\u00a0Stanford Question Answering Dataset,\u00a0SQuAD.\u00a0So it came out around 2016 and Microsoft Research Asia was actually at the top of the leader board, like, throughout its existence. So for, like 2016, 2017, they kept building better and better models until around 2018 they actually achieved human parity, which is just a big milestone in general when you have these academic benchmarks. I think that was like one of the most exciting milestones around the natural language space, that we were able to <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/www.linkedin.com\/pulse\/microsofts-ai-now-good-humans-squad-reading-test-rangan-majumder?articleId=6359118161786347520#comments-6359118161786347520&trk=public_profile_article_view\" target=\"_blank\" rel=\"noopener noreferrer\">achieve human parity on this\u00a0SQuAD\u00a0data set<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>\u2026<\/p>\n<p><b>Host: Right.<\/b><\/p>\n<p>Rangan\u00a0Majumder: \u2026within two years. And then, I think around 2018, another dataset came out, which is\u00a0Conversational Question Answering.\u00a0A year later, 2019, once again Microsoft Research Asia, along with some other folks in, I think, XD\u2019s Speech Team was able to\u2026<\/p>\n<p><b>Host: Yeah.<\/b><\/p>\n<p>Rangan\u00a0Majumder: \u2026achieve human parity on that. Around that same time there was this GLUE benchmark, which was also very interesting.<\/p>\n<p><b>Host: And GLUE stands for?<\/b><\/p>\n<p>Rangan\u00a0Majumder: General Language Understanding benchmark. So they had, I think, ten very different natural language tasks. So they thought, well this\u00a0one\u2019s going to be very hard. If we can build one model that can do well on all ten of these, that\u2019s going to be pretty impressive and once again, in a year, Microsoft Research was able to do that.<\/p>\n<p><b>Host: Unbelievable.<\/b><\/p>\n<p>Rangan\u00a0Majumder: So that\u2019s where they came up with this MT-DNN model.<\/p>\n<p><b>Host: Which stands for?<\/b><\/p>\n<p>Rangan\u00a0Majumder: Multi-Task Deep Neural Network.<\/p>\n<p><b>Host:\u00a0<\/b><b>Right<\/b><b>.<\/b><\/p>\n<p>Rangan\u00a0Majumder: Yeah. So basically, like, in language, Microsoft Research has been doing a really awesome job.<\/p>\n<p><b>Host: Yeah.<\/b><\/p>\n<p>Rangan\u00a0Majumder: And while they\u2019re doing that, our team is just taking those models and productionizing them. And what\u2019s interesting is, just because you do well on academic tasks doesn\u2019t necessarily mean it\u2019s really ready to be shipped into production. And the first big learning was with the\u00a0SQuAD\u00a0dataset\u2026<\/p>\n<p><b>Host: Yeah.<\/b><\/p>\n<p>Rangan\u00a0Majumder: \u2026which I talked about back in 2016,\u00a02018. So the model they used there was called Reading Net or R-Net.<\/p>\n<p><b>Host: Mm-hmm.<\/b><\/p>\n<p>Rangan\u00a0Majumder: And we realized that data set they had was a little bit biased because every \u2013 like the way this data set works is,\u00a0you have a question and you have, like, a passage and you\u2019re basically trying to answer the question, but their entire dataset was guaranteed that every question has an answer.<\/p>\n<p><b>Host: Hmm.<\/b><\/p>\n<p>Rangan\u00a0Majumder: But in a production context, when people are asking questions to the search engine, not every question has an answer. And in fact, some questions shouldn\u2019t be answered at all, right?\u00a0So we need to actually also add unanswerable questions.<\/p>\n<p><b>Host: Well, I want to talk about a project that you\u2019ve been involved in called Project Turing, named after the eponymous Alan Turing, and you call it your internal deep learning moonshot, which I love! What was the motivation and inspiration behind that project and what are some of the cool products, or product features, that have come out of that work?<\/b><\/p>\n<p>Rangan\u00a0Majumder: Yeah, so Project Turing was started about 2016. The motivation for it was, we were doing a bunch of analysis on, basically, the types of queries we were getting. And there was one segment that really stood out because it was the fastest growing segment of queries. It was question queries. So people were no longer just typing in key words, they were asking questions to a search engine. So, like, instead of\u00a0people\u00a0typing in, you know, fishing license, they would say like,\u00a0fishing age in Washington, right?\u00a0What is the fishing age in Washington when I could go fish? So we looked at that and we thought well, people just don\u2019t want to click on a web page, they just want you to find the answer for them.<\/p>\n<p><b>Host: Right.<\/b><\/p>\n<p>Rangan\u00a0Majumder: And then, many times, the words that were in the question and the words that were actually in the answer were very different.<\/p>\n<p><b>Host: Right.<\/b><\/p>\n<p>Rangan\u00a0Majumder:\u00a0So the previous approach, which\u00a0was,\u00a0like,\u00a0let\u2019s just do key word matching,\u00a0was not going to work.\u00a0We had to match at a different level, at the semantic level. So,\u00a0at the time, deep learning was really impressive in terms of these perception tasks like vision,\u00a0you know,\u00a0speech\u2026\u00a0so we were thinking like could it be really good at these other higher level tasks like language? So, that\u2019s when we started Project Turing,\u00a0and the idea was, what if we could do,\u00a0like,\u00a0end-to-end deep learning across the entire web to be able to answer these questions?\u00a0And it basically completely changed our search\u00a0architecture\u00a0to be able to do this kind of thing. And that\u2019s why it was a moonshot. So today, every time you issue a query,\u00a0we\u2019re running deep learning across,\u00a0basically,\u00a0the entire web to get that answer. And if we didn\u2019t use deep learning, we wouldn\u2019t be able to answer a lot of these questions\u2026<\/p>\n<p><b>Host: Right.<\/b><\/p>\n<p>Rangan\u00a0Majumder: \u2026because key word matching just wouldn\u2019t work.<\/p>\n<p><b>Host: So that actually is happening now?<\/b><\/p>\n<p>Rangan\u00a0Majumder: Yes, that\u2019s correct. That is happening and as we did it, there are all sorts of new innovations that came out of it that we realized are reusable for other parts of the company. And as we kept pushing the system, we noticed users kept asking harder and harder questions so then we just had to build better and better models. So, there are a lot of interesting things that came out of Project Turing. So first was, we\u2019ve got this deep learning search stack, deep learning question answering system, but then we started to build these Turing Neural Language Representation. And then, just recently we announced the Turing NLG, or Natural Language Generation. So we realized, many times, the passage itself can be kind of long, that comes from a web page, so sometimes we need to rewrite it and shorten it for people, so that\u2019s why we started to look into this generation task. We were able to train one of the largest deep learning\u00a0language models and that\u2019s called Turing NLG and we announced that I think last month.<\/p>\n<p><b>Host: Right, so it\u2019s very new.<\/b><\/p>\n<p>Rangan\u00a0Majumder: Yes, very new. It\u2019s seventeen billion parameters, it was\u00a0like impressing\u2026<\/p>\n<p><b>Host: Wait, wait, wait. Seventeen billion?<\/b><\/p>\n<p>Rangan\u00a0Majumder: Yes, seventeen billion parameters.<\/p>\n<p><b>Host: Oh, my gosh.<\/b><\/p>\n<p>Rangan\u00a0Majumder: Yeah, and just like three years ago, our biggest model was probably ten million parameters. So\u00a0it\u00a0just shows you how quickly the space is growing.<\/p>\n<p><b>Host: Okay, so with that kind of context, where\u2019s the next number of parameters? Are we going to hit a trillion? I mean, is, is this scalable to that level?<\/b><\/p>\n<p>Rangan\u00a0Majumder: Yeah, that\u2019s a good question. So definitely we\u2019re going to keep pushing it because every time we get an order of magnitude, we notice it could just do better, so we\u2019re not seeing it slowing down. So as long as you get improvements that could ship to customers, we\u2019re going to keep pushing the boundaries. But at the same time, we need to be more and more efficient with our computation and also\u00a0just not chase something for vanity\u2019s sake, right?<\/p>\n<p><b>Host: Right.<\/b><\/p>\n<p>Rangan\u00a0Majumder: Like just because we can get to a hundred billion parameters, which we want to be able to do, we also need to make sure we\u2019re really maximizing\u00a0the value that the model is actually getting with all those parameters too.<\/p>\n<p><b>Host: I guess I should have said a hundred billion before jumping to a trillion\u2026 It\u2019s like a \u201ctriple dog dare\u201d right after the \u201cdare you.\u201d<\/b><\/p>\n<p><b><i>(music plays)<\/i><\/b><\/p>\n<p><b>Host: So drilling in a little bit on these different manifestations of your technology, I know that there\u2019s one called Brainwave that is part of the search experience now, and you had talked a little bit about the fact that Project Turing and Brainwave were co-developed, or concurrently developed, because they each had gaps that they needed to fill. Tell our listeners how Turing and Brainwave came about together and how it speaks to the importance of collaboration, which you\u2019ve already referred to earlier on, across research and product boundaries.<\/b><\/p>\n<p>Rangan Majumder: Yeah, so these really large deep learning models are very expensive. So they really actually push both the software and the hardware to its limits. So while we\u2019re trying to train these really big models, or even ship them to customers, we need to push the software and push the hardware. So Brainwave, the idea was, they could actually take deep learning models and <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/blogs.microsoft.com\/ai\/build-2018-project-brainwave\/\" target=\"_blank\" rel=\"noopener noreferrer\">accelerate them really fast<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, but they really needed models that were worthy of that kind of hardware, right? They spent a lot of time building this Brainwave compiler and we got all these FPGAs in our data center and when our models were kind of small, like ten million parameters, sure, you can use Brainwave, but it was just making something that was already possible just a little bit faster.<\/p>\n<p><b>Host: It\u2019s like taking a thoroughbred to a kid\u2019s party\u2026<\/b><\/p>\n<p>Rangan\u00a0Majumder: That\u2019s right. But it wasn\u2019t until we got to these really large models, like the Turing NLR model, which was, you know,\u00a0three hundred million parameters, or even six\u00a0hundred million parameters, and it was so big, if we tried to run it without any kind of optimizations, it would probably take about six hundred milliseconds.\u00a0And we\u00a0would have to run this multiple times for every search. So imagine, you know, you type in a query, hit enter and it took you like five seconds to load the page, right? So this\u00a0is\u00a0something that was unacceptable. But we were getting these huge improvements from it. Like I said before, it was the biggest improvements we were getting.<\/p>\n<p><b>Host: Right.<\/b><\/p>\n<p>Rangan\u00a0Majumder: So imagine that we\u2019ve got this thing, which we knew was excellent for customers, but we had no way to ship it. And that was the problem we had on our modeling side. And then, on my Brainwave team, they\u2019re like, I\u2019ve got this awesome hardware, but I have no really, like, big models pushing us. So that\u2019s how these two were kind of co-developed. So they needed something to push their platform, and the modeling side of my team needed hardware that can actually run these models. So, what ended up happening is, these models, which would take six hundred milliseconds, unoptimized, we got it down to five milliseconds, which is blazing fast. So the way to think about five milliseconds is, the blink of an eye is, like, you know, three hundred milliseconds. So every time you blink, you know, we\u2019re running about\u00a0like\u00a0fifty inferences. I think Brainwave was just one part of that hardware story. The other thing we\u2019d done is we partnered with Nvidia to be able to build faster and faster ways to run inference on GPUs. So we actually open sourced that in ONNX, the ONNX Runtime, so if people want to reuse our work, they can just go download the ONNX Runtime, and the other thing we\u2019ve been able to do, this was also part of our announcement in February is, to train that seventeen billion parameter model, we had to do all sorts of things that weren\u2019t done before because you can\u2019t fit this model into GPUs, right? So we open sourced this library called\u00a0DeepSpeed. It\u2019s very easy to use, and it\u2019s just a great way to train really large models super-fast.<\/p>\n<p><b>Host: Talk about what you\u2019ve called the most interesting story here, something you call the network effect. What do you mean by that, and how does the network effect make everything better?<\/b><\/p>\n<p>Rangan\u00a0Majumder: It\u2019s super-interesting just the type of collaboration we\u2019re getting. So we train a model once for a scenario in Bing, and that same model is reused for lots of scenarios in Bing, lots of scenarios in Office\u2026 So the economies of scale, which is, you know, each team can just easily get huge impact by just reusing something somebody else did, is really transformative. The second type of network effect we\u2019re seeing is, basically by open sourcing this code, like the ONNX Runtime and\u00a0DeepSpeed, and we also open sourced our vector search code called SPTAG,\u00a0so by doing that, other people can now reuse the work, but also contribute to the work.<\/p>\n<p><b>Host: Wow.<\/b><\/p>\n<p>Rangan\u00a0Majumder: So it just keeps getting better and better. So that\u2019s something, you know, our team really believes in. Like, if you open source something that we think is state-of-the-art, by other people contributing to it, it can continue being the state-of-the-art.<\/p>\n<p><b>Host: Right, are you seeing this across the industry? That other companies are open sourcing these really powerful technologies and code?<\/b><\/p>\n<p>Rangan\u00a0Majumder: Yeah, absolutely. That\u2019s one of the exciting things. You know, Google open sourced TensorFlow, they open sourced their BERT model\u2026 Facebook\u2019s open sourcing a lot of their models. They have\u00a0PyTorch, which is open source\u2026<\/p>\n<p><b>Host: Right.<\/b><\/p>\n<p>Rangan\u00a0Majumder: \u2026so it\u2019s been really great that all these AI companies and leaders are actually open sourcing their technology.<\/p>\n<p><b>Host: Is anyone keeping their cards close to their chest on any particular things?<\/b><\/p>\n<p>Rangan\u00a0Majumder: Well, I mean, you can never know for sure.<\/p>\n<p><b>Host: Right.<\/b><\/p>\n<p>Rangan\u00a0Majumder: But the general consensus is, researchers want to work in an open way, you know. The old way of working in an open way was just publishing papers.<\/p>\n<p><b>Host: Right.<\/b><\/p>\n<p>Rangan\u00a0Majumder: But now it\u2019s about open sourcing.<\/p>\n<p><b>Host: Right.<\/b><\/p>\n<p>Rangan\u00a0Majumder: So that\u2019s\u2026 I think open sourcing is the new, like, publishing papers. So they really want to share their achievement, and that\u2019s one way of just proving\u2026 like, you can write a paper, but is that reproducible? Many times it\u2019s not. But if you open source it, then people can really test it and you know it really works.<\/p>\n<p><b>Host: Right. Well, as much as I love asking questions about the upside of technological innovation, I always have to ask about the downside. So now I\u2019ll ask you,\u00a0<\/b><b>Rangan<\/b><b>, is there anything about the work you\u2019re doing that keeps you up at night, metaphorically, and if so, what are you doing at the outset to help mitigate it?<\/b><\/p>\n<p>Rangan\u00a0Majumder: Yeah, definitely the thing that worries me the most is around AI and ethics. So if you think about these really large models, they\u2019re trained on existing data in a self-supervised way. So they take the data, all this text, and you know all this text that humans write actually have biases.<\/p>\n<p><b>Host: Right.<\/b><\/p>\n<p>Rangan\u00a0Majumder: All the existing data out there has a bias. And I think Microsoft Research even showed this in one of their papers so if you look at the word embedding for nurse,\u00a0it tends to be closer to female word embedding than the male word embedding, right? So these models, if they\u2019re just trained on this already biased data, they\u2019re going to learn that kind of bias and that kind of stuff definitely worries me. In fact, when we launched Turing NLG, we had a demo page for it and we wanted to share the demo with everybody, but right before we did that, I gave it to a couple of folks on my team who were like hackers and I said, hey, why don\u2019t you try to break this? And within a couple of hours they came back and were like, you know, they could manipulate the model to say, like, offensive things.\u00a0I just said, well, if I just gave this to everybody, they could easily show examples where this model was saying some things that are inappropriate.<\/p>\n<p><b>Host: Right.<\/b><\/p>\n<p>Rangan\u00a0Majumder: And then, like, just one or two examples of it doing inappropriate things would just wash away all the good things that it could do.<\/p>\n<p><b>Host: Right.<\/b><\/p>\n<p>Rangan\u00a0Majumder: So that\u2019s\u00a0why\u00a0we decided to release it in a controlled way. But I think that\u2019s a really important problem for the entire AI community to solve, like how do we solve the bias that we have in our data?<\/p>\n<p><b>Host: Right.<\/b><\/p>\n<p>Rangan\u00a0Majumder: Especially when you\u2019re using these AI models to make decisions that could affect people\u2019s lives.<\/p>\n<p><b>Host: All right, so many, many people that have been in this booth have said the same thing in terms of identifying the problem, and this is something we need to think about and something we need to talk about. Is anybody, and maybe the product side is the closest to, you know, the reification of all of this, is anybody thinking about\u00a0<\/b><b><i>how<\/i><\/b><b>\u00a0you do that? How you let this out in a controlled way and\/or keep the bad actors from doing their best work?<\/b><\/p>\n<p>Rangan\u00a0Majumder: Yeah, obviously we have to because we can\u2019t just let these models out and do inappropriate things, especially when they show up in products like Bing and Office and so on.<\/p>\n<p><b>Host: Right.<\/b><\/p>\n<p>Rangan\u00a0Majumder: So the first thing we have to do is measure the problem. We have a lot of metrics to just make sure, like, okay, the question answering experience is not saying offensive things, right? And it\u2019s actually kind of tough because, as the models get smarter, they get better at finding answers, and anywhere on the web, there\u2019s, like, somebody who\u2019s written some garbage, right?<\/p>\n<p><b>Host: Right.<\/b><\/p>\n<p>Rangan\u00a0Majumder: So you could ask, basically, fill in the blank,\u00a0like,\u00a0is so and so a bad person?\u00a0And there will be somebody out there\u2026<\/p>\n<p><b>Host: Absolutely.<\/b><\/p>\n<p>Rangan\u00a0Majumder: \u2026who has written that, right? So we actually first have to measure the thing you don\u2019t want to accidentally do, and I think that\u2019s probably the first thing around this space you have to do. Like come up with some metrics around bias.<\/p>\n<p><b>Host: Right.<\/b><\/p>\n<p>Rangan\u00a0Majumder: And then,\u00a0once you have good metrics, the thing I\u2019ve seen is, teams are very good at optimizing for that. But it\u2019s still a very hard problem to do. Like how do you even measure bias?\u00a0How do you make sure that you\u2019re measuring all sorts of bias?<\/p>\n<p><b>Host: Yeah, this is going all the way up to the C-Suite. Brad Smith is talking about it in his book and it\u2019s a big deal even in academic works is how, you know, do you put out \u201cparental controls\u201d on a product?\u00a0<\/b><b>I use parental controls in<\/b><b>\u2026<\/b><\/p>\n<p>Rangan\u00a0Majumder: Yeah.<\/p>\n<p><b>Host:<\/b><b>\u00a0<\/b><b>\u2026<\/b><b>quotation marks, but you know?<\/b><\/p>\n<p>Rangan\u00a0Majumder: That\u2019s right and he has this\u00a0Aether\u00a0committee that our team is actually involved in.<\/p>\n<p><b>Host: Right.<\/b><\/p>\n<p>Rangan\u00a0Majumder: Around just making sure, like, this AI we are building, make sure it can\u2019t be harmful, it\u2019s used in, you know, like a responsible way, and so on.<\/p>\n<p><b>Host: Yeah.<\/b><\/p>\n<p>Rangan\u00a0Majumder: So I think that\u2019s one AI ethical angle that I\u2019m worried about. The second one is really around inclusivity.<\/p>\n<p><b>Host: Mm-hmm.<\/b><\/p>\n<p>Rangan\u00a0Majumder: So if you look at all the AI breakthroughs, they\u2019re mostly coming from a few companies.\u00a0They\u2019re,\u00a0you know, Microsoft, Google, Facebook, some of the Chinese companies like Alibaba, and that\u2019s because these really large models take a lot of compute to go build, so I am worried that it\u2019s just going to be a few companies that are just doing all the AI breakthroughs. So we really need to think about, how do we make it more inclusive? And I think the good news is people are open sourcing a lot of their technologies so others can use it, but when it comes to compute and things like that, like there is only a few companies that can afford that kind of stuff.<\/p>\n<p><b>Host: Right.<\/b><\/p>\n<p>Rangan\u00a0Majumder: So we need to also think about,\u00a0like,\u00a0how do we make sure this AI transformation is inclusive for as many people as possible.<\/p>\n<p><b>Host: Well,\u00a0<\/b><b>and\u00a0<\/b><b>I think you\u2019re starting to see that in some of the AI for Good efforts that Microsoft is doing\u2026<\/b><\/p>\n<p>Rangan\u00a0Majumder: Absolutely.<\/p>\n<p><b>Host: \u2026with, you know, these grants that aren\u2019t just money, but they\u2019re compute resources, right? You can use this and use Azure for free.<\/b><\/p>\n<p>Rangan\u00a0Majumder: That\u2019s absolutely right, yeah. So that\u2019s one way to do that.<\/p>\n<p><b>Host: All right. Tell us a bit about yourself. Where did the high-tech life begin for you and how did you end up at Microsoft?<\/b><\/p>\n<p>Rangan\u00a0Majumder: I guess the high-tech life started for me when I went to Carnegie Mellon and I studied computer science and computer engineering. And while I was there, I took some machine learning courses. And this was\u00a0early\u00a02000s, so machine learning wasn\u2019t nearly as impressive as it is today, but at the time it was extremely fascinating because I\u2019ve always been interested in these open-ended questions, like the meaning of life.\u00a0How do people think, is also one of those open ended questions that I was very fascinated with, so when I was learning machine learning it\u2019s like well, one way to learn how people think is to kind of rebuild it in machines. And then, I came to Microsoft. I started as a developer for about four years. Then I switched to program management,\u00a0and\u00a0I also switched to the Bing team because, at that point, while they were building up the Search team, I realized this is the best place to apply machine learning, right?\u00a0So if I want to really be at the cutting edge of machine learning,\u00a0like this is the place to be,\u00a0so\u2026\u00a0and I\u2019ve been there for the last ten years.<\/p>\n<p><b>Host: Yeah.<\/b><\/p>\n<p>Rangan\u00a0Majumder: Just applying machine learning to solve customer problems.<\/p>\n<p><b>Host: Has anyone ever tried to say, come back to academia and get an advanced degree?<\/b><\/p>\n<p>Rangan\u00a0Majumder: Yeah, definitely. I\u2019d say my parents are the ones who are saying that, because they, you know, they think, if you get a PhD it just means so much to you and the family and so on\u2026 I\u2019m like well,\u00a0I could, but I\u2019m having so much fun here!<\/p>\n<p><b>Host: What\u2019s something interesting that people might not know about you? And maybe it\u2019s a life event that impacted your career or maybe it\u2019s a personality trait that made you who you are today?\u00a0<\/b><b>Or maybe it has no connection to any of that and it\u2019s just an interesting data point that we couldn\u2019t find out about you if we typed your name into the Bing box?<\/b><\/p>\n<p>Rangan\u00a0Majumder: The thing that probably most people would be surprised at first,\u00a0that you can\u2019t find in the web\u00a0is, when I was younger, I got diagnosed with ADHD.\u00a0I was getting in trouble in school all the time. I wasn\u2019t,\u00a0like,\u00a0doing well,\u00a0and at one point,\u00a0the teacher and the principal, like,\u00a0brought my parents in and they said, hey, like,\u00a0you have to do something about him otherwise,\u00a0like,\u00a0he won\u2019t be able to return to school.\u00a0So my parents took me to a therapist and then they diagnosed me with ADHD, and then they gave me some drugs and\u00a0it completely changed my life because\u00a0I started to, like, get the highest grades in the class. Like, I was no longer getting into trouble. It was so strange, I remember,\u00a0my parents noticed that my behavior was different, so they took me off it for a little while and almost immediately I started getting in trouble again. But one thing that was different this time was, I noticed it and I realized well, I don\u2019t like getting in trouble. Like, this is not fun for me! So then I actually made a conscious effort to try and do well at school and not get in trouble. And then I was able to kind of make up for it.\u00a0So I really made an effort to sort of control and change my behavior.<\/p>\n<p><b>Host: As we close, I\u2019d like to circle back to the beginning, and sort of tie things together. If the big goal is to make us smarter and more productive, and we\u2019re not there yet, what are the big open problems in the fields that, if solved, would get us closer to the big goal, and what kinds of people do we need to help us get there?<\/b><\/p>\n<p>Rangan\u00a0Majumder: So the first few things that you\u2019re going to see over the next year or two is multi-modal models. So that\u2019s mixing text, images, and videos together in a single representation. That\u2019s something we\u2019re experimenting with today. We\u2019re seeing some good results.\u00a0So I think you\u2019ll be able to, like, ask questions and get an answer in images\u00a0or in videos, like look inside a video, look inside an image. So I think that\u2019s going to be pretty cool. The other thing is, like, we\u2019re definitely betting big on this deep learning, so you\u2019re going to see us be more and more efficient around, how do we run these models, train them with less compute? How do we get more out of it? Data efficiency is another thing. Given there\u2019s a limited amount of data,\u00a0how do we make sure that we\u2019re maximizing it to build better models? But I\u2019d say, in the long term, the thing that is still missing is, um\u2026 Like, I think there\u2019s two AI camps. There\u2019s this deep learning camp and\u00a0then\u00a0there\u2019s this, what they call, a symbolist camp, which is looking at graphs and structured data, and I think there still needs to be a way to fuse those two so that you can actually take unstructured data and reason over it the way you can with structured data, and you can create new knowledge and things like that, because there\u2019s a lot of questions we\u2019re seeing people ask, and sure, the answer isn\u2019t written there, but if you combine the information in two paragraphs, you can actually get the answer by combining them, so I think that\u2019s something we\u2019re still thinking about. It\u2019s not going to be an easy problem, but I think that\u2019s something the academic field and industry still needs to do.<\/p>\n<p><b>Host:\u00a0<\/b><b>Rangan<\/b><b>\u00a0Majumder, thank you so much for joining us today. It\u2019s been a real pleasure.<\/b><\/p>\n<p>Rangan\u00a0Majumder: Thank you for having me!<\/p>\n<p><b><i>(music plays)<\/i><\/b><\/p>\n<p><b><i>To learn more about\u00a0<\/i><\/b><b><i>Rangan<\/i><\/b><b><i>\u00a0Majumder and the latest advances in Search and AI technology, visit Microsoft.com\/research<\/i><\/b><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Rangan Majumder is the Partner Group Program Manager of Microsoft\u2019s Search and AI, and he has a simple goal: to make the world smarter and more productive. But nobody said simple was easy, so he and his team are working on better \u2013 and faster \u2013 ways to help you find the information you\u2019re looking for, anywhere you\u2019re looking for it. On the podcast, Rangan talks about how three big trends have changed the way Microsoft is building \u2013 and sharing \u2013 AI stacks across product groups. He also tells us about Project Turing, an internal deep learning moonshot that aims to harness the resources of the web and bring the power of deep learning to a search box near you.<\/p>\n","protected":false},"author":37583,"featured_media":644961,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"msr-url-field":"https:\/\/player.blubrry.com\/id\/57607598","msr-podcast-episode":"112","msrModifiedDate":"","msrModifiedDateEnabled":false,"ep_exclude_from_search":false,"_classifai_error":"","msr-author-ordering":[],"msr_hide_image_in_river":0,"footnotes":""},"categories":[240054],"tags":[],"research-area":[13556,13562,13545,13555],"msr-region":[],"msr-event-type":[],"msr-locale":[268875],"msr-post-option":[],"msr-impact-theme":[],"msr-promo-type":[],"msr-podcast-series":[],"class_list":["post-644958","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-msr-podcast","msr-research-area-artificial-intelligence","msr-research-area-computer-vision","msr-research-area-human-language-technologies","msr-research-area-search-information-retrieval","msr-locale-en_us"],"msr_event_details":{"start":"","end":"","location":""},"podcast_url":"https:\/\/player.blubrry.com\/id\/57607598","podcast_episode":"112","msr_research_lab":[],"msr_impact_theme":[],"related-publications":[],"related-downloads":[],"related-videos":[],"related-academic-programs":[],"related-groups":[],"related-projects":[649749],"related-events":[],"related-researchers":[],"msr_type":"Post","featured_image_thumbnail":"<img width=\"960\" height=\"540\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/03\/Research_Podcast_Majumder_Site_1400x788-960x540.png\" class=\"img-object-cover\" alt=\"Headshot of Rangan Majumder for the Microsoft Research Podcast\" decoding=\"async\" loading=\"lazy\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/03\/Research_Podcast_Majumder_Site_1400x788-960x540.png 960w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/03\/Research_Podcast_Majumder_Site_1400x788-300x169.png 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/03\/Research_Podcast_Majumder_Site_1400x788-1024x576.png 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/03\/Research_Podcast_Majumder_Site_1400x788-768x432.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/03\/Research_Podcast_Majumder_Site_1400x788-1066x600.png 1066w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/03\/Research_Podcast_Majumder_Site_1400x788-655x368.png 655w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/03\/Research_Podcast_Majumder_Site_1400x788-343x193.png 343w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/03\/Research_Podcast_Majumder_Site_1400x788-640x360.png 640w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/03\/Research_Podcast_Majumder_Site_1400x788-1280x720.png 1280w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/03\/Research_Podcast_Majumder_Site_1400x788.png 1400w\" sizes=\"auto, (max-width: 960px) 100vw, 960px\" \/>","byline":"","formattedDate":"March 25, 2020","formattedExcerpt":"Rangan Majumder is the Partner Group Program Manager of Microsoft\u2019s Search and AI, and he has a simple goal: to make the world smarter and more productive. But nobody said simple was easy, so he and his team are working on better \u2013 and faster&hellip;","locale":{"slug":"en_us","name":"English","native":"","english":"English"},"_links":{"self":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts\/644958","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/users\/37583"}],"replies":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/comments?post=644958"}],"version-history":[{"count":8,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts\/644958\/revisions"}],"predecessor-version":[{"id":668127,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts\/644958\/revisions\/668127"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media\/644961"}],"wp:attachment":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media?parent=644958"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/categories?post=644958"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/tags?post=644958"},{"taxonomy":"msr-research-area","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/research-area?post=644958"},{"taxonomy":"msr-region","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-region?post=644958"},{"taxonomy":"msr-event-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event-type?post=644958"},{"taxonomy":"msr-locale","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-locale?post=644958"},{"taxonomy":"msr-post-option","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-post-option?post=644958"},{"taxonomy":"msr-impact-theme","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-impact-theme?post=644958"},{"taxonomy":"msr-promo-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-promo-type?post=644958"},{"taxonomy":"msr-podcast-series","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-podcast-series?post=644958"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}