{"id":1098456,"date":"2024-11-13T09:30:00","date_gmt":"2024-11-13T17:30:00","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/?p=1098456"},"modified":"2024-11-14T20:53:17","modified_gmt":"2024-11-15T04:53:17","slug":"toward-modular-models-collaborative-ai-development-enables-model-accountability-and-continuous-learning","status":"publish","type":"post","link":"https:\/\/www.microsoft.com\/en-us\/research\/blog\/toward-modular-models-collaborative-ai-development-enables-model-accountability-and-continuous-learning\/","title":{"rendered":"Toward modular models: Collaborative AI development enables model accountability and continuous learning"},"content":{"rendered":"\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"576\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models-BlogHeroFeature-1400x788-1-1024x576.jpg\" alt=\"Modular Models blog hero\" class=\"wp-image-1098480\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models-BlogHeroFeature-1400x788-1-1024x576.jpg 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models-BlogHeroFeature-1400x788-1-300x169.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models-BlogHeroFeature-1400x788-1-768x432.jpg 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models-BlogHeroFeature-1400x788-1-1066x600.jpg 1066w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models-BlogHeroFeature-1400x788-1-655x368.jpg 655w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models-BlogHeroFeature-1400x788-1-240x135.jpg 240w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models-BlogHeroFeature-1400x788-1-640x360.jpg 640w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models-BlogHeroFeature-1400x788-1-960x540.jpg 960w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models-BlogHeroFeature-1400x788-1-1280x720.jpg 1280w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models-BlogHeroFeature-1400x788-1.jpg 1400w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure>\n\n\n\n<p>Today, development of generalizable AI models requires access to sufficient data and compute resources, which may create challenges for some researchers. Democratizing access to technology across the research community can advance the development of generalizable AI models. By applying the core software development concept of modularity to AI, we can build models that are powerful, efficient, adaptable, and transparent.&nbsp;<\/p>\n\n\n\n<p>Until recently, AI models were primarily built using monolithic architecture. Though powerful, these models can be challenging to customize and edit compared to modular models with easily interpretable functional components. Today, developers employ modularity to make services more reliable, faster to refine, and easier for multiple users to contribute to simultaneously. One promising research direction that supports this involves shifting AI development towards a <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/colinraffel.com\/blog\/a-call-to-build-models-like-we-build-open-source-software.html\" target=\"_blank\" rel=\"noopener noreferrer\">modular approach<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, which could enhance flexibility and improve scalability.&nbsp;<\/p>\n\n\n\n<p>One such approach is to use numerous fine-tuned models designed for specific tasks, known as <em>expert models<\/em>, and coordinate them to solve broader tasks (see&nbsp;<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/towards-modular-llms-by-building-and-reusing-a-library-of-loras\/\">Towards Modular LLMs by Building and Reusing a Library of LoRAs &#8211; Microsoft Research<\/a>,&nbsp;<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/arxiv.org\/abs\/2402.05859\" target=\"_blank\" rel=\"noopener noreferrer\">Learning to Route Among Specialized Experts for Zero-Shot Generalization<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>). These expert models can be developed in a decentralized way. Similar to the benefits of using a microservice architecture, this modular AI approach can be more flexible, cheaper to develop, and more compliant with relevant privacy and legal policies. However, while substantial research has been done on training optimization, coordination methods remain largely unexplored.<\/p>\n\n\n\n<p>Our team is exploring the potential of modular models by focusing on two themes: i) optimizing the training of expert models and ii) refining how expert models coordinate to form a collaborative model. One method for coordinating expert models is to adaptively select the most relevant independently developed expert models for specific tasks or queries. This approach, called <em>MoErging<\/em>, is similar to Mixture-of-Experts (MoE) approaches but differs in that the routing mechanism is learned <em>after<\/em> the individual experts are trained. As an initial step, we contributed to creating a taxonomy for organizing recent MoErging methods with the goal of helping establish a shared language for the research community and facilitating easier and fairer comparisons between different methods.&nbsp;<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"assessing-existing-moerging-methods\">Assessing existing MoErging methods<\/h2>\n\n\n\n<p>Most MoErging methods were developed within the past year, so they don\u2019t reference each and are difficult to compare. To enable comparison of MoErging methods, we recently collaborated on a <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/a-survey-on-model-moerging-recycling-and-routing-among-specialized-experts-for-collaborative-learning\/\" target=\"_blank\" rel=\"noreferrer noopener\">survey<\/a> that establishes a taxonomy for comparing methods and organizes MoErging design choices into three steps:&nbsp;<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Expert design:<\/strong> Identifies and uses expert models trained asynchronously by distributed contributors.&nbsp;<\/li>\n\n\n\n<li><strong>Routing design:<\/strong> Routes tasks to the appropriate expert models.&nbsp;<\/li>\n\n\n\n<li><strong>Application design:<\/strong> Applies the merged models to specific tasks or domains.&nbsp;<\/li>\n<\/ul>\n\n\n\n<p>Each step is broken down into more detailed choices. For example, in <em>expert design<\/em>, expert training can be <em>custom<\/em> or <em>standard<\/em>, and training data can be <em>private<\/em> or <em>shared<\/em>. Custom training requires MoErging to have specific training procedures, while the standard training does not. Similarly, <em>shared data<\/em> means that the training data must be accessible for routing. Otherwise, the training data is considered private.&nbsp;<\/p>\n\n\n\n<p>The benefits of modular models discussed below assume that training data doesn\u2019t need to be shared. However, a review of current MoErging methods finds that some approaches do require sharing training data, making certain benefits no longer applicable.&nbsp;<\/p>\n\n\n\n\t<div class=\"border-bottom border-top border-gray-300 mt-5 mb-5 msr-promo text-center text-md-left alignwide\" data-bi-aN=\"promo\" data-bi-id=\"1002645\">\n\t\t\n\n\t\t<p class=\"msr-promo__label text-gray-800 text-center text-uppercase\">\n\t\t<span class=\"px-4 bg-white display-inline-block font-weight-semibold small\">Spotlight: AI-POWERED EXPERIENCE<\/span>\n\t<\/p>\n\t\n\t<div class=\"row pt-3 pb-4 align-items-center\">\n\t\t\t\t\t\t<div class=\"msr-promo__media col-12 col-md-5\">\n\t\t\t\t<a class=\"bg-gray-300 display-block\" href=\"https:\/\/aka.ms\/research-copilot\/?OCID=msr_researchforum_Copilot_MCR_Blog_Promo\" aria-label=\"Microsoft research copilot experience\" data-bi-cN=\"Microsoft research copilot experience\" target=\"_blank\">\n\t\t\t\t\t<img decoding=\"async\" class=\"w-100 display-block\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/01\/MSR-Chat-Promo.png\" alt=\"\" \/>\n\t\t\t\t<\/a>\n\t\t\t<\/div>\n\t\t\t\n\t\t\t<div class=\"msr-promo__content p-3 px-5 col-12 col-md\">\n\n\t\t\t\t\t\t\t\t\t<h2 class=\"h4\">Microsoft research copilot experience<\/h2>\n\t\t\t\t\n\t\t\t\t\t\t\t\t<p id=\"microsoft-research-copilot-experience\" class=\"large\">Discover more about research at Microsoft through our AI-powered experience<\/p>\n\t\t\t\t\n\t\t\t\t\t\t\t\t<div class=\"wp-block-buttons justify-content-center justify-content-md-start\">\n\t\t\t\t\t<div class=\"wp-block-button\">\n\t\t\t\t\t\t<a href=\"https:\/\/aka.ms\/research-copilot\/?OCID=msr_researchforum_Copilot_MCR_Blog_Promo\" aria-describedby=\"microsoft-research-copilot-experience\" class=\"btn btn-brand glyph-append glyph-append-chevron-right\" data-bi-cN=\"Microsoft research copilot experience\" target=\"_blank\">\n\t\t\t\t\t\t\tStart now\t\t\t\t\t\t<\/a>\n\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t\t\t\t<\/div><!--\/.msr-promo__content-->\n\t<\/div><!--\/.msr-promo__inner-wrap-->\n\t<\/div><!--\/.msr-promo-->\n\t\n\n\n<p>The survey evaluates 29 different MoErging methods using its taxonomy, which categorizes the design choices into two expert design choices, five routing design choices, and two application design options, shown in Figure 1.<\/p>\n\n\n\n<figure class=\"wp-block-image aligncenter size-full\"><img loading=\"lazy\" decoding=\"async\" width=\"1017\" height=\"1056\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models_Figure1.png\" alt=\"Taxonomy of model MoErging design choices. References in the leaf noes link to sections for specific papers that make some particular design choice. We omit references to methods for which a given choice is not applicable.\" class=\"wp-image-1098477\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models_Figure1.png 1017w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models_Figure1-289x300.png 289w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models_Figure1-986x1024.png 986w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models_Figure1-768x797.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models_Figure1-173x180.png 173w\" sizes=\"auto, (max-width: 1017px) 100vw, 1017px\" \/><figcaption class=\"wp-element-caption\">Figure 1: Taxonomy of model MoErging design choices. References in the leaf nodes link to sections of specific papers that implement each choice. We omit references to methods where a particular choice is not applicable.&nbsp;<\/figcaption><\/figure>\n\n\n\n<p>One takeaway from the survey is that most MoErging methods can be grouped into four categories based on their routing design choices:<\/p>\n\n\n\n<ol class=\"wp-block-list\">\n<li><strong>Classifier-based routing:<\/strong> Methods that train the router as a classifier using expert datasets or unseen data.&nbsp;<\/li>\n\n\n\n<li><strong>Embedding-based routing:<\/strong> Methods that compute embeddings of expert training sets and compare them to a query embedding for routing.&nbsp;<\/li>\n\n\n\n<li><strong>Nonrouter methods:<\/strong> Methods that do not explicitly train a router but instead initialize the router in an unsupervised manner.&nbsp;&nbsp;<\/li>\n\n\n\n<li><strong>Task-specific routing:<\/strong> Methods that learn a task-specific routing distribution over the target dataset to improve performance on a specific task.&nbsp;<\/li>\n<\/ol>\n\n\n\n<p>While the differences within each category are minor, the differences across categories are significant because they determine the level of data access required for implementation. As a result, data access is a primary factor in determining which methods are applicable and feasible in various settings.&nbsp;<\/p>\n\n\n\n<p>Our taxonomy also covers recent approaches to building agentic systems, which could be viewed as specific types of MoErging methods where experts are full language models and routing decisions are made on a step-by-step or example-by-example basis. The optimal level for MoErging may vary depending on the task and the computational resources available to each stakeholder.&nbsp;<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"potential-benefits-and-use-cases-of-modular-models\">Potential benefits and use cases of modular models&nbsp;<\/h2>\n\n\n\n<p>Modular models can unlock new benefits and use cases for AI, offering a promising approach to addressing challenges in current AI development. Moving forward, further substantial research is needed to validate this potential and assess feasibility.&nbsp;&nbsp;<\/p>\n\n\n\n<p>Modular AI may:&nbsp;<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Allow privacy-conscious contributions.<\/strong> &nbsp;Teams with sensitive or proprietary data, such as personally identifiable information (PII) and copyrighted content, can contribute expert models and benefit from larger projects without sharing their data. This capacity can make it easier to comply with data privacy and legal standards, which could be valuable for healthcare teams that would benefit from general model capabilities without combining their sensitive data with other training data.&nbsp;<\/li>\n\n\n\n<li><strong>Drive model transparency and accountability.<\/strong> &nbsp;Modular models allow specific expert models to be identified and, if necessary, removed or retrained. For example, if a module trained on PII, copyrighted, or biased data is identified, it can be removed more easily, eliminating the need for retraining and helping ensure compliance with privacy and ethical standards.&nbsp;<\/li>\n\n\n\n<li><strong>Facilitate model extensibility and continual improvement.<\/strong>&nbsp;Modularity supports continual improvements, allowing new capabilities from expert models to be integrated as they are available. This approach is akin to making localized edits, allowing for continuous, cost-effective improvement.&nbsp;<\/li>\n\n\n\n<li><strong>Lower the barrier to AI development for those with limited compute and data resources.<\/strong>&nbsp;Modular AI can reduce the need for extensive data and compute by creating a system where pretrained experts can be reused, benefiting academics, startups, and teams focused on niche use cases. For example, an AI agent tasked with booking flights on a specific website with limited training data could leverage general navigation and booking skills from other trained AI experts, enabling generalizable and broadly applicable skills without requiring domain-specific training data. We explore this process of transferring skills across tasks in our paper \u201c<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/multi-head-adapter-routing-for-cross-task-generalization\/%22%20%EF%B7%9FHYPERLINK%20%22https:\/\/www.microsoft.com\/en-us\/research\/publication\/multi-head-adapter-routing-for-cross-task-generalization\/\">Multi-Head Routing For Cross-Task Generalization<\/a>.\u201d&nbsp;<\/li>\n\n\n\n<li><strong>Support personalization.<\/strong> &nbsp;Modular models make it possible to equip AI agents with experts tailored to individual users or systems. For instance, AI designed to emulate five-time World Chess Champion Magnus Carlsen could enhance a player\u2019s preparation to play a match against him. Experiments suggest that storing knowledge or user profiles in on-demand modules can match or surpass the performance of retrieval-augmented generation (RAG), potentially reducing latency and improving the user\u2019s experience in custom AI applications.&nbsp;<\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"current-limitations-and-looking-forward\">Current limitations and looking forward&nbsp;<\/h2>\n\n\n\n<p>In this blog, we focused on a type of modular approach that involves training foundation models, which requires substantial compute power and large amounts of data. Despite the advantages of modularity, such as increased flexibility, efficiency, and adaptability, the development of foundation models remains resource-intensive, necessitating high-performance computing and robust datasets to support fine-tuning.&nbsp;&nbsp;<\/p>\n\n\n\n<p>Recent work has begun to address these challenges by <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/arxiv.org\/abs\/2311.08105\" target=\"_blank\" rel=\"noopener noreferrer\">distributing the pretraining process of foundation models<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>. Looking ahead, a promising research direction focuses on exploring how to create a minimal dataset for training \u201cempty foundation models\u201d while shifting most of their capabilities to external pluggable modules.&nbsp;<\/p>\n\n\n\n<p>Modular methods are evolving rapidly, and we\u2019re excited by their potential. Modularity has the capacity to democratize AI development, improve model accountability, and support efficient continuous learning. With the MoErging taxonomy, we aim to establish a shared language that fosters engagement within the research community. This research is in the early stages, and we welcome community collaboration. If you&#8217;re interested in working with us, please reach out to <a href=\"mailto:ModularModels@microsoft.com\" target=\"_blank\" rel=\"noreferrer noopener\">ModularModels@microsoft.com<\/a>.&nbsp;<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"acknowledgements\">Acknowledgements<\/h2>\n\n\n\n<p>We would like to thank paper collaborators: Prateek Yadav, Colin Raffel, Mohammed Muqeeth, Haokun Liu, Tianlong Chen, Mohit Bansal, Leshem Choshen, Edoardo Ponti, Zhan Su, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/matpereira\/\">Matheus Pereira<\/a>, Nicolas Le Roux, Nabil Omi, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/sidsen\/\">Siddhartha Sen<\/a>, Anurag Sarkar, Jordan T. Ash, Oleksiy Ostapenko,\u00a0and\u00a0Laurent Charlin.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Modular models can democratize AI development while unlocking new benefits and use cases. Modularized AI can be more flexible, more compliant, and cheaper to develop\u2014requiring less data and fewer compute resources to train expert models.<\/p>\n","protected":false},"author":42735,"featured_media":1098480,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"msr-url-field":"","msr-podcast-episode":"","msrModifiedDate":"","msrModifiedDateEnabled":false,"ep_exclude_from_search":false,"_classifai_error":"","msr-author-ordering":[{"type":"user_nicename","value":"Lucas Caccia","user_id":"43647"},{"type":"user_nicename","value":"Darya Moldavskaya","user_id":"43569"},{"type":"user_nicename","value":"Alessandro Sordoni","user_id":"37230"}],"msr_hide_image_in_river":null,"footnotes":""},"categories":[1],"tags":[],"research-area":[13556],"msr-region":[],"msr-event-type":[],"msr-locale":[268875],"msr-post-option":[269148,243984,269142],"msr-impact-theme":[],"msr-promo-type":[],"msr-podcast-series":[],"class_list":["post-1098456","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-research-blog","msr-research-area-artificial-intelligence","msr-locale-en_us","msr-post-option-approved-for-river","msr-post-option-blog-homepage-featured","msr-post-option-include-in-river"],"msr_event_details":{"start":"","end":"","location":""},"podcast_url":"","podcast_episode":"","msr_research_lab":[199571,437514],"msr_impact_theme":[],"related-publications":[],"related-downloads":[],"related-videos":[],"related-academic-programs":[],"related-groups":[],"related-projects":[],"related-events":[],"related-researchers":[{"type":"user_nicename","value":"Lucas Caccia","user_id":43647,"display_name":"Lucas Caccia","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/lpagecaccia\/\" aria-label=\"Visit the profile page for Lucas Caccia\">Lucas Caccia<\/a>","is_active":false,"last_first":"Caccia, Lucas","people_section":0,"alias":"lpagecaccia"},{"type":"user_nicename","value":"Darya Moldavskaya","user_id":43569,"display_name":"Darya Moldavskaya","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/dmoldavskaya\/\" aria-label=\"Visit the profile page for Darya Moldavskaya\">Darya Moldavskaya<\/a>","is_active":false,"last_first":"Moldavskaya, Darya","people_section":0,"alias":"dmoldavskaya"},{"type":"user_nicename","value":"Alessandro Sordoni","user_id":37230,"display_name":"Alessandro Sordoni","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/alsordon\/\" aria-label=\"Visit the profile page for Alessandro Sordoni\">Alessandro Sordoni<\/a>","is_active":false,"last_first":"Sordoni, Alessandro","people_section":0,"alias":"alsordon"}],"msr_type":"Post","featured_image_thumbnail":"<img width=\"960\" height=\"540\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models-BlogHeroFeature-1400x788-1-960x540.jpg\" class=\"img-object-cover\" alt=\"Modular Models blog hero\" decoding=\"async\" loading=\"lazy\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models-BlogHeroFeature-1400x788-1-960x540.jpg 960w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models-BlogHeroFeature-1400x788-1-300x169.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models-BlogHeroFeature-1400x788-1-1024x576.jpg 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models-BlogHeroFeature-1400x788-1-768x432.jpg 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models-BlogHeroFeature-1400x788-1-1066x600.jpg 1066w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models-BlogHeroFeature-1400x788-1-655x368.jpg 655w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models-BlogHeroFeature-1400x788-1-240x135.jpg 240w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models-BlogHeroFeature-1400x788-1-640x360.jpg 640w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models-BlogHeroFeature-1400x788-1-1280x720.jpg 1280w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/10\/Modular-Models-BlogHeroFeature-1400x788-1.jpg 1400w\" sizes=\"auto, (max-width: 960px) 100vw, 960px\" \/>","byline":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/lpagecaccia\/\" title=\"Go to researcher profile for Lucas Caccia\" aria-label=\"Go to researcher profile for Lucas Caccia\" data-bi-type=\"byline author\" data-bi-cN=\"Lucas Caccia\">Lucas Caccia<\/a>, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/dmoldavskaya\/\" title=\"Go to researcher profile for Darya Moldavskaya\" aria-label=\"Go to researcher profile for Darya Moldavskaya\" data-bi-type=\"byline author\" data-bi-cN=\"Darya Moldavskaya\">Darya Moldavskaya<\/a>, and <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/alsordon\/\" title=\"Go to researcher profile for Alessandro Sordoni\" aria-label=\"Go to researcher profile for Alessandro Sordoni\" data-bi-type=\"byline author\" data-bi-cN=\"Alessandro Sordoni\">Alessandro Sordoni<\/a>","formattedDate":"November 13, 2024","formattedExcerpt":"Modular models can democratize AI development while unlocking new benefits and use cases. Modularized AI can be more flexible, more compliant, and cheaper to develop\u2014requiring less data and fewer compute resources to train expert models.","locale":{"slug":"en_us","name":"English","native":"","english":"English"},"_links":{"self":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts\/1098456","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/users\/42735"}],"replies":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/comments?post=1098456"}],"version-history":[{"count":21,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts\/1098456\/revisions"}],"predecessor-version":[{"id":1103436,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts\/1098456\/revisions\/1103436"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media\/1098480"}],"wp:attachment":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media?parent=1098456"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/categories?post=1098456"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/tags?post=1098456"},{"taxonomy":"msr-research-area","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/research-area?post=1098456"},{"taxonomy":"msr-region","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-region?post=1098456"},{"taxonomy":"msr-event-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event-type?post=1098456"},{"taxonomy":"msr-locale","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-locale?post=1098456"},{"taxonomy":"msr-post-option","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-post-option?post=1098456"},{"taxonomy":"msr-impact-theme","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-impact-theme?post=1098456"},{"taxonomy":"msr-promo-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-promo-type?post=1098456"},{"taxonomy":"msr-podcast-series","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-podcast-series?post=1098456"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}