{"id":1138863,"date":"2025-05-19T09:00:11","date_gmt":"2025-05-19T16:00:11","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/?p=1138863"},"modified":"2025-11-26T14:38:25","modified_gmt":"2025-11-26T22:38:25","slug":"magentic-ui-an-experimental-human-centered-web-agent","status":"publish","type":"post","link":"https:\/\/www.microsoft.com\/en-us\/research\/blog\/magentic-ui-an-experimental-human-centered-web-agent\/","title":{"rendered":"Magentic-UI, an experimental human-centered web agent"},"content":{"rendered":"\n<figure class=\"wp-block-image size-full\"><img loading=\"lazy\" decoding=\"async\" width=\"1400\" height=\"788\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/MagenticUI-BlogHeroFeature-1400x788-1.jpg\" alt=\"This figure denotes a human figure above a small monitor on the left and a gear on the right with arrows pointing to each.\" class=\"wp-image-1139340\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/MagenticUI-BlogHeroFeature-1400x788-1.jpg 1400w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/MagenticUI-BlogHeroFeature-1400x788-1-300x169.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/MagenticUI-BlogHeroFeature-1400x788-1-1024x576.jpg 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/MagenticUI-BlogHeroFeature-1400x788-1-768x432.jpg 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/MagenticUI-BlogHeroFeature-1400x788-1-1066x600.jpg 1066w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/MagenticUI-BlogHeroFeature-1400x788-1-655x368.jpg 655w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/MagenticUI-BlogHeroFeature-1400x788-1-240x135.jpg 240w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/MagenticUI-BlogHeroFeature-1400x788-1-640x360.jpg 640w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/MagenticUI-BlogHeroFeature-1400x788-1-960x540.jpg 960w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/MagenticUI-BlogHeroFeature-1400x788-1-1280x720.jpg 1280w\" sizes=\"auto, (max-width: 1400px) 100vw, 1400px\" \/><\/figure>\n\n\n\n<p>Modern productivity is rooted in the web\u2014from searching for information and filling in forms to navigating dashboards. Yet, many of these tasks remain manual and repetitive. Today, we are introducing Magentic-UI, a new open-source research prototype of a <em>human-centered<\/em> agent that is meant to help researchers study open questions on human-in-the-loop approaches and oversight mechanisms for AI agents. This prototype <strong>collaborates with users on web-based tasks<\/strong> and operates in real time over a web browser. Unlike other computer use agents that aim for full autonomy, Magentic-UI offers a transparent and controllable experience for tasks that are <em>action-oriented <\/em>and<em> <\/em>require activities beyond just performing simple web searches.<\/p>\n\n\n\n<p>Magentic-UI builds on <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/ai.azure.com\/labs\/projects\/magentic-one\" target=\"_blank\" rel=\"noopener noreferrer\">Magentic-One<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, a powerful multi-agent team we released last year, and is powered by <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/github.com\/microsoft\/autogen\" target=\"_blank\" rel=\"noopener noreferrer\">AutoGen<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, our leading agent framework. It is available under MIT license at <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/github.com\/microsoft\/Magentic-UI\" target=\"_blank\" rel=\"noopener noreferrer\">https:\/\/github.com\/microsoft\/Magentic-UI<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> and on <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/ai.azure.com\/labs\/projects\/magentic-ui\" target=\"_blank\" rel=\"noopener noreferrer\">Azure AI Foundry Labs<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, the hub where developers, startups, and enterprises can explore groundbreaking innovations from Microsoft Research. Magentic-UI is integrated with Azure AI Foundry models and agents. Learn more about how to integrate Azure AI agents into the Magentic-UI multi-agent architecture by following this <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/github.com\/microsoft\/magentic-ui\/blob\/main\/samples\/sample_azure_agent.py\" target=\"_blank\" rel=\"noopener noreferrer\">code sample<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>.&nbsp;<\/p>\n\n\n\n\t<div class=\"border-bottom border-top border-gray-300 mt-5 mb-5 msr-promo text-center text-md-left alignwide\" data-bi-aN=\"promo\" data-bi-id=\"1144027\">\n\t\t\n\n\t\t<p class=\"msr-promo__label text-gray-800 text-center text-uppercase\">\n\t\t<span class=\"px-4 bg-white display-inline-block font-weight-semibold small\">PODCAST SERIES<\/span>\n\t<\/p>\n\t\n\t<div class=\"row pt-3 pb-4 align-items-center\">\n\t\t\t\t\t\t<div class=\"msr-promo__media col-12 col-md-5\">\n\t\t\t\t<a class=\"bg-gray-300 display-block\" href=\"https:\/\/www.microsoft.com\/en-us\/research\/story\/ai-testing-and-evaluation-learnings-from-science-and-industry\/\" aria-label=\"AI Testing and Evaluation: Learnings from Science and Industry\" data-bi-cN=\"AI Testing and Evaluation: Learnings from Science and Industry\" target=\"_blank\">\n\t\t\t\t\t<img decoding=\"async\" class=\"w-100 display-block\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/06\/EP2-AI-TE_Hero_Feature_River_No_Text_1400x788.jpg\" alt=\"Illustrated headshots of Daniel Carpenter, Timo Minssen, Chad Atalla, and Kathleen Sullivan for the Microsoft Research Podcast\" \/>\n\t\t\t\t<\/a>\n\t\t\t<\/div>\n\t\t\t\n\t\t\t<div class=\"msr-promo__content p-3 px-5 col-12 col-md\">\n\n\t\t\t\t\t\t\t\t\t<h2 class=\"h4\">AI Testing and Evaluation: Learnings from Science and Industry<\/h2>\n\t\t\t\t\n\t\t\t\t\t\t\t\t<p id=\"ai-testing-and-evaluation-learnings-from-science-and-industry\" class=\"large\">Discover how Microsoft is learning from other domains to advance evaluation and testing as a pillar of AI governance.<\/p>\n\t\t\t\t\n\t\t\t\t\t\t\t\t<div class=\"wp-block-buttons justify-content-center justify-content-md-start\">\n\t\t\t\t\t<div class=\"wp-block-button\">\n\t\t\t\t\t\t<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/story\/ai-testing-and-evaluation-learnings-from-science-and-industry\/\" aria-describedby=\"ai-testing-and-evaluation-learnings-from-science-and-industry\" class=\"btn btn-brand glyph-append glyph-append-chevron-right\" data-bi-cN=\"AI Testing and Evaluation: Learnings from Science and Industry\" target=\"_blank\">\n\t\t\t\t\t\t\tListen now\t\t\t\t\t\t<\/a>\n\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t\t\t\t<\/div><!--\/.msr-promo__content-->\n\t<\/div><!--\/.msr-promo__inner-wrap-->\n\t<\/div><!--\/.msr-promo-->\n\t\n\n\n<p>Magentic-UI can perform tasks that require browsing the web, writing and executing Python and shell code, and understanding files. Its key features include:<\/p>\n\n\n\n<ol start=\"1\" class=\"wp-block-list\">\n<li><strong>Collaborative planning with users (co-planning)<\/strong>. Magentic-UI allows users to directly modify its plan through a plan editor or by providing textual feedback before Magentic-UI executes any actions.&nbsp;<\/li>\n\n\n\n<li><strong>Collaborative execution with users (co-tasking)<\/strong>. Users can pause the system and give feedback in natural language or demonstrate it by directly taking control of the browser.<\/li>\n\n\n\n<li><strong>Safety with human-in-the-loop (action guards)<\/strong>. Magentic-UI seeks user approval before executing potentially irreversible actions, and the user can specify how often Magentic-UI needs approvals. Furthermore, Magentic-UI is sandboxed for the safe operation of tools such as browsers and code executors.<\/li>\n\n\n\n<li><strong>Safety with human-in-the-loop<\/strong>. Magentic-UI seeks user approval before executing potentially irreversible actions, and the user can specify how often Magentic-UI needs approvals. Furthermore, Magentic-UI is sandboxed for the safe operation of tools such as browsers and code executors.&nbsp;<\/li>\n\n\n\n<li><strong>Learning from experience (plan learning)<\/strong>. Magentic-UI can learn and save plans from previous interactions to improve task completion for future tasks.&nbsp;<\/li>\n<\/ol>\n\n\n\n<figure class=\"wp-block-image aligncenter size-full\"><a data-bi-bhvr=\"14\"  data-bi-cn=\"A screenshot of Magentic-UI actively performing a task. The left side of the screen shows Magentic-UI stating its plan based on the following query \"Find a suitable birthday figt for a 5 years old nephew in Seattle who loves bikes\". On the right hand side the figure shows magentic-ui \" href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/landing-screenshot.png\"><img loading=\"lazy\" decoding=\"async\" width=\"2322\" height=\"1700\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/landing-screenshot.png\" alt=\"A screenshot of Magentic-UI actively performing a task. The left side of the screen shows Magentic-UI stating its plan based on the following query \"Find a suitable birthday figt for a 5 years old nephew in Seattle who loves bikes\". On the right hand side the figure shows magentic-ui 's progress to accomplish this user\u2019s complex goal. The right side shows the browser Magentic-UI is controlling.\" class=\"wp-image-1139443\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/landing-screenshot.png 2322w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/landing-screenshot-300x220.png 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/landing-screenshot-1024x750.png 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/landing-screenshot-768x562.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/landing-screenshot-1536x1125.png 1536w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/landing-screenshot-2048x1499.png 2048w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/landing-screenshot-80x60.png 80w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/landing-screenshot-240x176.png 240w\" sizes=\"auto, (max-width: 2322px) 100vw, 2322px\" \/><\/a><figcaption class=\"wp-element-caption\"><em>Figure 1: Screenshot of Magentic-UI actively performing a task. The left side of the screen shows Magentic-UI stating its plan and progress to accomplish a user\u2019s complex goal. The right side shows the browser Magentic-UI is controlling.&nbsp;<\/em><\/figcaption><\/figure>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"how-is-magentic-ui-human-centered\">How is Magentic-UI human-centered?<\/h2>\n\n\n\n<p>While many web agents promise full autonomy, in practice users can be left unsure of what the agent can do, what it is currently doing, and whether they have enough control to intervene when something goes wrong or doesn\u2019t occur as expected. By contrast, Magentic-UI considers user needs at every stage of interaction. We followed a human-centered design methodology in building Magentic-UI by prototyping and obtaining feedback from pilot users during its design.&nbsp;<\/p>\n\n\n\n<figure class=\"wp-block-image aligncenter size-full\"><a data-bi-bhvr=\"14\"  data-bi-cn=\"Co-planning - This figure describes how users can collaboratively plan with Magentic-UI. On the left hand side, users can accept the plan magentic-ui creates or re-create the plan. On the right hand side they can see the actions magnetic-ui is taking on the browser.\" href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/coplanning.gif\"><img loading=\"lazy\" decoding=\"async\" width=\"1440\" height=\"1080\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/coplanning.gif\" alt=\"Co-planning - This figure describes how users can collaboratively plan with Magentic-UI. On the left hand side, users can accept the plan magentic-ui creates or re-create the plan. On the right hand side they can see the actions magnetic-ui is taking on the browser.\" class=\"wp-image-1139454\"\/><\/a><figcaption class=\"wp-element-caption\"><em>Figure 2: Co-planning &#8211; Users can collaboratively plan with Magentic-UI.<\/em><\/figcaption><\/figure>\n\n\n\n<p>For example, after a person specifies and before Magentic-UI even begins to execute, it creates a clear step-by-step plan that outlines what it would do to accomplish the task. People can collaborate with Magentic-UI to modify this plan and then give final approval for Magentic-UI to begin execution. This is crucial as users may have expectations of how the task should be completed; communicating that information could significantly improve agent <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/navigating-rifts-in-human-llm-grounding-study-and-benchmark\/\">performance<\/a>. We call this feature co-planning.<\/p>\n\n\n\n<p>During execution, Magentic-UI shows in real time what specific actions it\u2019s about to take. For example, whether it is about to click on a button or input a search query. It also shows in real time what it observed on the web pages it is visiting. Users can take control of the action at any point in time and give control back to the agent. We call this feature co-tasking.<\/p>\n\n\n\n<figure class=\"wp-block-image aligncenter size-full\"><a data-bi-bhvr=\"14\"  data-bi-cn=\"Co-tasking. This fiture shows how magentic-ui shows the plan it created to the user and allows the user to push a button to accept the plan or modify it.\" href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/cotasking.gif\"><img loading=\"lazy\" decoding=\"async\" width=\"1428\" height=\"1080\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/cotasking.gif\" alt=\"Co-tasking. This fiture shows how magentic-ui shows the plan it created to the user and allows the user to push a button to accept the plan or modify it.\" class=\"wp-image-1139456\"\/><\/a><figcaption class=\"wp-element-caption\"><em>Figure 3: Co-tasking &#8211; Magentic-UI provides real-time updates about what it is about to do and what it already did, allowing users to collaboratively complete tasks with the agent.<\/em><\/figcaption><\/figure>\n\n\n\n<figure class=\"wp-block-image aligncenter size-full\"><a data-bi-bhvr=\"14\"  data-bi-cn=\"Action-guards. TOn the left hand side,this figure shous how Magentic-UI will ask users for permission using Approve or Reject buttons before executing actions that it deems consequential or important. On the right hand side the figure shows magentic-ui picking Crab Wonton from Thai Thani\" href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/magui-actionguard.png\"><img loading=\"lazy\" decoding=\"async\" width=\"1534\" height=\"1148\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/magui-actionguard.png\" alt=\"Action-guards. TOn the left hand side,this figure shous how Magentic-UI will ask users for permission using Approve or Reject buttons before executing actions that it deems consequential or important. On the right hand side the figure shows magentic-ui picking Crab Wonton from Thai Thani's restaurant menu.\" class=\"wp-image-1139447\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/magui-actionguard.png 1534w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/magui-actionguard-300x225.png 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/magui-actionguard-1024x766.png 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/magui-actionguard-768x575.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/magui-actionguard-80x60.png 80w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/magui-actionguard-240x180.png 240w\" sizes=\"auto, (max-width: 1534px) 100vw, 1534px\" \/><\/a><figcaption class=\"wp-element-caption\"><em>Figure 4: Action-guards \u2013 Magentic-UI will ask users for permission before executing actions that it deems consequential or important.&nbsp;<\/em><\/figcaption><\/figure>\n\n\n\n<p>Additionally, Magentic-UI asks for user permission before performing actions that are deemed irreversible, such as closing a tab or clicking a button with side effects. We call these \u201caction guards\u201d. The user can also configure Magentic-UI\u2019s action guards to always ask for permission before performing any action. If the user deems an action risky (e.g., paying for an item), they can reject it.&nbsp;<\/p>\n\n\n\n<div class=\"wp-block-columns is-layout-flex wp-container-core-columns-is-layout-9d6595d7 wp-block-columns-is-layout-flex\">\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<figure class=\"wp-block-image aligncenter size-full\"><a data-bi-bhvr=\"14\"  data-bi-cn=\"This figure shows how magentic-ui learns a plan by allowing the users to click on a button entitled \"Plan Learned\"\" href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/plan-learning-1.png\"><img loading=\"lazy\" decoding=\"async\" width=\"782\" height=\"790\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/plan-learning-1.png\" alt=\"This figure shows how magentic-ui learns a plan by allowing the users to click on a button entitled \"Plan Learned\"\" class=\"wp-image-1139462\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/plan-learning-1.png 782w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/plan-learning-1-297x300.png 297w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/plan-learning-1-768x776.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/plan-learning-1-178x180.png 178w\" sizes=\"auto, (max-width: 782px) 100vw, 782px\" \/><\/a><\/figure>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<figure class=\"wp-block-image aligncenter size-large\"><a data-bi-bhvr=\"14\"  data-bi-cn=\"This figure shows how users can see their saved plans and click on a button to either run that same plan or edit it before running it.\" href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/plan-learning-2.png\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"777\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/plan-learning-2-1024x777.png\" alt=\"This figure shows how users can see their saved plans and click on a button to either run that same plan or edit it before running it.\" class=\"wp-image-1139463\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/plan-learning-2-1024x777.png 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/plan-learning-2-300x228.png 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/plan-learning-2-768x583.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/plan-learning-2-80x60.png 80w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/plan-learning-2-237x180.png 237w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/plan-learning-2.png 1196w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/a><\/figure>\n<\/div>\n<\/div>\n\n\n\n<figure class=\"wp-block-video aligncenter\"><figcaption class=\"wp-element-caption\"><em>Figure 5: Plan learning \u2013 Once a task is successfully completed, users can request Magentic-UI to learn a step-by-step plan from this experience.<\/em><\/figcaption><\/figure>\n\n\n\n<p>After execution, the user can ask Magentic-UI to reflect on the conversation and infer and save a step-by-step plan for future similar tasks. Users can view and modify saved plans for Magentic-UI to reuse in the future in a saved-plans gallery. In a future session, users can launch Magentic-UI with the saved plan to either execute the same task again, like checking the price of a specific flight, or use the plan as a guide to help complete similar tasks, such as checking the price of a different type of flight.&nbsp;<\/p>\n\n\n\n<p>Combined, these four features\u2014co-planning, co-tasking, action guards, and plan learning\u2014enable users to collaborate effectively with Magentic-UI.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"architecture\">Architecture<\/h2>\n\n\n\n<p>Magentic-UI\u2019s underlying system is a team of specialized agents adapted from AutoGen\u2019s <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/articles\/magentic-one-a-generalist-multi-agent-system-for-solving-complex-tasks\/?msockid=0d5953b355086cfe362446d0547f6d22\" target=\"_blank\" rel=\"noreferrer noopener\">Magentic-One<\/a> system. The agents work together to create a modular system:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Orchestrator <\/strong>is the lead agent, powered by a large language model (LLM), that performs co-planning with the user, decides when to ask the user for feedback, and delegates sub-tasks to the remaining agents to complete.<\/li>\n\n\n\n<li><strong>WebSurfer <\/strong>is an LLM agent equipped with a web browser that it can control. Given a request by the Orchestrator, it can click, type, scroll, and visit pages in multiple rounds to complete the request from the Orchestrator.<\/li>\n\n\n\n<li><strong>Coder <\/strong>is an LLM agent equipped with a Docker code-execution container. It can write and execute Python and shell commands and provide a response back to the Orchestrator.<\/li>\n\n\n\n<li><strong>FileSurfer<\/strong> is an LLM agent equipped with a Docker code-execution container and file-conversion tools from the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/github.com\/microsoft\/markitdown\" target=\"_blank\" rel=\"noopener noreferrer\">MarkItDown<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> package. It can locate files in the directory controlled by Magentic-UI, convert files to markdown, and answer questions about them.<\/li>\n<\/ul>\n\n\n\n<figure class=\"wp-block-image aligncenter size-full\"><a data-bi-bhvr=\"14\"  data-bi-cn=\"This figure shows a histogram. It\" href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/magenticui.png\"><img loading=\"lazy\" decoding=\"async\" width=\"1723\" height=\"1307\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/magenticui.png\" alt=\"This figure shows a histogram. It's a comparison on the GAIA validation set of the accuracy of Magentic-One, Magentic-UI in autonomous mode, Magentic-UI with a simulated user powered by a smarter LLM than the MAGUI agents, Magentic-UI with a simulated user that has a\\access to side information about the tasks, and human performance. This shows that human-in-the-loop can improve the accuracy of autonomous agents, bridging the gap to human performance at a fraction of the cost.  \" class=\"wp-image-1139336\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/magenticui.png 1723w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/magenticui-300x228.png 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/magenticui-1024x777.png 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/magenticui-768x583.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/magenticui-1536x1165.png 1536w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/magenticui-80x60.png 80w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/magenticui-237x180.png 237w\" sizes=\"auto, (max-width: 1723px) 100vw, 1723px\" \/><\/a><figcaption class=\"wp-element-caption\"><em>Figure 6: System architecture diagram of Magentic-UI<\/em><\/figcaption><\/figure>\n\n\n\n<p>To interact with Magentic-UI, users can enter a text message and attach images. In response, Magentic-UI creates a natural-language step-by-step plan with which users can interact through a plan-editing interface. Users can add, delete, edit, regenerate steps, and write follow-up messages to iterate on the plan. While the user editing the plan adds an upfront cost to the interaction, it can potentially save a significant amount of time in the agent executing the plan and increase its chance at success.<\/p>\n\n\n\n<p>The plan is stored inside the Orchestrator and is used to execute the task. For each step of the plan, the Orchestrator determines which of the agents (WebSurfer, Coder, FileSurfer) or the user should complete the step. Once that decision is made, the Orchestrator sends a request to one of the agents or the user and waits for a response. After the response is received, the Orchestrator decides whether that step is complete. If it is, the Orchestrator moves on to the following step.<\/p>\n\n\n\n<p>Once all steps are completed, the Orchestrator generates a final answer that is presented to the user. If, while executing any of the steps, the Orchestrator decides that the plan is inadequate (for example, because a certain website is unreachable), the Orchestrator can replan with user permission and start executing a new plan.<\/p>\n\n\n\n<p>All intermediate progress steps are clearly displayed to the user. Furthermore, the user can pause the execution of the plan and send additional requests or feedback. The user can also configure through the interface whether agent actions (e.g., clicking a button) require approval.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"evaluating-magentic-ui\">Evaluating Magentic-UI<\/h2>\n\n\n\n<p>Magentic-UI innovates through its ability to integrate human feedback in its planning and execution of tasks. We performed a preliminary automated evaluation to showcase this ability on the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/arxiv.org\/abs\/2311.12983\" target=\"_blank\" rel=\"noopener noreferrer\">GAIA benchmark<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> for agents with a user-simulation experiment.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"evaluation-with-simulated-users\">Evaluation with simulated users<\/h3>\n\n\n\n<figure class=\"wp-block-image aligncenter size-full\"><a data-bi-bhvr=\"14\"  data-bi-cn=\"This figure shows a box diagram to describe the architecture of Magentic-UI.\" href=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/Magentic_UI_Figure.png\"><img loading=\"lazy\" decoding=\"async\" width=\"943\" height=\"609\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/Magentic_UI_Figure.png\" alt=\"This figure shows a box diagram to describe the architecture of Magentic-UI.\" class=\"wp-image-1139273\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/Magentic_UI_Figure.png 943w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/Magentic_UI_Figure-300x194.png 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/Magentic_UI_Figure-768x496.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/Magentic_UI_Figure-240x155.png 240w\" sizes=\"auto, (max-width: 943px) 100vw, 943px\" \/><\/a><figcaption class=\"wp-element-caption\"><em>Figure 7: Comparison on the GAIA validation set of the accuracy of Magentic-One, Magentic-UI in autonomous mode, Magentic-UI with a simulated user powered by a smarter LLM than the MAGUI agents, Magentic-UI with a simulated user that has a\\access to side information about the tasks, and human performance. This shows that human-in-the-loop can improve the accuracy of autonomous agents, bridging the gap to human performance at a fraction of the cost.<\/em><\/figcaption><\/figure>\n\n\n\n<p>GAIA is a benchmark for general AI assistants, with multimodal question-answer pairs that are challenging, requiring the agents to navigate the web, process files, and execute code. The traditional evaluation setup with GAIA assumes the system will autonomously complete the task and return an answer, which is compared to the ground-truth answer.&nbsp;<\/p>\n\n\n\n<p>To evaluate the human-in-the-loop capabilities of Magentic-UI, we transform GAIA into an <em>interactive benchmark <\/em>by introducing the concept of a simulated user. Simulated users provide value in two ways: by having specific expertise that the agent may not possess, and by providing guidance on how the task should be performed.<\/p>\n\n\n\n<p>We experiment with two types of simulated users to show the value of human-in-the-loop: (1) a simulated user that is more intelligent than the Magentic-UI agents and (2) a simulated user with the same intelligence as Magentic-UI agents but with additional information about the task. During co-planning, Magentic-UI takes feedback from this simulated user to improve its plan. During co-tasking, Magentic-UI can ask the (simulated) user for help when it gets stuck. Finally, if Magentic-UI does not provide a final answer, then the simulated user provides an answer instead. These experiments reflect a lower bound on the value of human feedback, since real users can step in at any time and offer any kind of input\u2014not just when the system explicitly asks for help.<\/p>\n\n\n\n<p>The simulated user is an LLM without any tools, instructed to interact with Magentic-UI the way we expect a human would act. The&nbsp;first type of&nbsp;simulated user relies on OpenAI\u2019s o4-mini, more performant at many tasks than the one powering the Magentic-UI agents (GPT-4o). For the second type of simulated user, we use GPT-4o for both the simulated user and the rest of the agents, but the user has access to <em>side information<\/em> about each task. Each task in GAIA has side information, which includes a human-written <em>plan<\/em> to solve the task. While this <em>plan<\/em> is not used as input in the traditional benchmark, in our <em>interactive<\/em> setting we provide this information to the second type of <em>simulated<\/em> user,which is powered by an LLM so that it can mimic a knowledgeable user. Importantly, we tuned our simulated user so as not to reveal the ground-truth answer directly as the answer is usually found inside the human written plan. Instead, it is prompted to guide Magentic-UI indirectly. We found that this tuning prevented the simulated user from inadvertently revealing the answer in all but 6% of tasks when Magentic-UI provides a final answer.&nbsp;<\/p>\n\n\n\n<p>On the validation subset of GAIA (162 tasks), we show the results of Magentic-One operating in autonomous mode, Magentic-UI operating in autonomous mode (without the simulated user), Magentic-UI with simulated user (1) (smarter model), Magentic-UI with simulated user (2) (side-information), and human performance. We first note that Magentic-UI in autonomous mode is within a margin of error of the performance of Magentic-One. Note that the same LLM (GPT-4o) is used for Magentic-UI and Magentic-One.<\/p>\n\n\n\n<p>Magentic-UI with the simulated user that has access to side information improves the accuracy of autonomous Magentic-UI by 71%, from a 30.3% task-completion rate to a 51.9% task-completion rate. Moreover, Magentic-UI only asks for help from the simulated user in 10% of tasks and relies on the simulated user for the final answer in 18% of tasks. And in those tasks where it does ask for help, it asks for help on average 1.1 times. Magentic-UI with the simulated user powered by a smarter model improves to 42.6% where Magentic-UI asks for help in only 4.3% of tasks, asking for help an average of 1.7 times in those tasks. This demonstrates the potential of even lightweight human feedback for improving performance (e.g., task completion) over autonomous agents working alone, especially at a fraction of the cost compared to people completing tasks entirely manually.&nbsp;<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"learning-and-reusing-plans\">Learning and reusing plans<\/h2>\n\n\n\n<p>As described above, once Magentic-UI completes a task, users have the option for Magentic-UI to learn a plan based on the execution of the task. These plans are saved in a plan gallery, which users and Magentic-UI can access in the future.<\/p>\n\n\n\n<p>The user can select a plan from the plan gallery, which is displayed by clicking on the Saved Plans button. Alternatively, as a user enters a task that closely matches a previous task, the saved plan will be displayed even before the user is done typing. If no identical task is found, Magentic-UI can use <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/microsoft.github.io\/autogen\/stable\/reference\/python\/autogen_ext.experimental.task_centric_memory.html\" target=\"_blank\" rel=\"noopener noreferrer\">AutoGen&#8217;s Task-Centric Memory<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> to retrieve plans for any similar tasks. Our preliminary evaluations show that this retrieval is highly accurate, and when recalling a saved plan can be around 3x faster than generating a new plan. Once a plan is recalled or generated, the user can always accept it, modify it, or ask Magentic-UI to modify it for the specific task at hand.&nbsp;<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"safety-and-control\">Safety and control<\/h2>\n\n\n\n<p>Magentic-UI can surf the live internet and execute code. With such capabilities, we need to ensure that Magentic-UI acts in a safe and secure manner. The following features, design decisions, and evaluations were made to ensure this:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Allow-list<\/strong>: Users can set a list of websites that Magentic-UI is allowed to access. If Magentic-UI needs to access a website outside of the allow-list, users must explicitly approve it through the interface<\/li>\n\n\n\n<li><strong>Anytime interruptions<\/strong>: At any point of Magentic-UI completing the task, the user can interrupt Magentic-UI and stop any pending code execution or web browsing.<\/li>\n\n\n\n<li><strong>Docker sandboxing<\/strong>: Magentic-UI controls a browser that is launched inside a Docker container with no credentials, which avoids risks with logged-in accounts and credentials. Moreover, any code execution is also performed inside a separate Docker container to avoid affecting the host environment in which Magentic-UI is running. This is illustrated in the system architecture of Magentic-UI (Figure 3).<\/li>\n\n\n\n<li><strong>Detection and approval of irreversible agent actions: <\/strong>Users can configure an action-approval policy (action guards) to determine which actions Magentic-UI can perform without user approval. In the extreme, users can specify that any action (e.g., any button click) needs explicit user approval. Users must press an \u201cAccept\u201d or \u201cDeny\u201d button for each action.<\/li>\n<\/ul>\n\n\n\n<p>In addition to the above design decisions, we performed a red-team evaluation of Magentic-UI on a set of internal scenarios, which we developed to challenge the security and safety of Magentic-UI. Such scenarios include cross-site prompt injection attacks, where web pages contain malicious instructions distinct from the user\u2019s original intent (e.g., to execute risky code, access sensitive files, or perform actions on other websites). It also contains scenarios comparable to phishing, which try to trick Magentic-UI into entering sensitive information, or granting permissions on impostor sites (e.g., a synthetic website that asks Magentic-UI to log in and enter Google credentials to read an article). In our preliminary evaluations, we found that Magentic-UI either refuses to complete the requests, stops to ask the user, or, as a final safety measure, is eventually unable to complete the request due to Docker sandboxing. We have found that this layered approach is effective for thwarting these attacks.<\/p>\n\n\n\n<p>We have also released transparency notes, which can be found at: <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/github.com\/microsoft\/magentic-ui\/blob\/main\/TRANSPARENCY_NOTE.md\" target=\"_blank\" rel=\"noopener noreferrer\">https:\/\/github.com\/microsoft\/magentic-ui\/blob\/main\/TRANSPARENCY_NOTE.md<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"open-research-questions\">Open research questions&nbsp;<\/h2>\n\n\n\n<p>Magentic-UI provides a tool for researchers to study critical questions in agentic systems and particularly on human-agent interaction. In a previous <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/arxiv.org\/abs\/2412.10380\" target=\"_blank\" rel=\"noopener noreferrer\">report<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, we outlined 12 questions for human-agent communication, and Magentic-UI provides a vehicle to study these questions in a realistic setting. A key question among these is how we enable humans to efficiently intervene and provide feedback to the agent while executing a task. Humans should not have to constantly watch the agent. Ideally, the agent should know when to reach out for help and provide the necessary context for the human to assist it. A second question is about safety. As agents interact with the live web, they may become prone to attacks from malicious actors. We need to study what necessary safeguards are needed to protect the human from side effects without adding a heavy burden on the human to verify every agent action. There are also many other questions surrounding security, personalization, and learning that Magentic-UI can help with studying.&nbsp;<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"conclusion\">Conclusion<\/h2>\n\n\n\n<p>Magentic-UI is an open-source agent prototype that works with people to complete complex tasks that require multi-step planning and browser use. As agentic systems expand in the scope of tasks they can complete, Magentic-UI\u2019s design enables better transparency into agent actions and enables human control to ensure safety and reliability. Moreover, by facilitating human intervention, we can improve performance while still reducing human cost in completing tasks on aggregate. Today we have released the first version of Magentic-UI. Looking ahead, we plan to continue developing it in the open with the goal of improving its capabilities and answering research questions on human-agent collaboration. We invite the research community to extend and reuse Magentic-UI for their scientific explorations and domains.&nbsp;<\/p>\n\n\n\n<div class=\"wp-block-buttons is-layout-flex wp-block-buttons-is-layout-flex\">\n<div class=\"wp-block-button is-style-pill\"><a data-bi-type=\"button\" class=\"wp-block-button__link has-white-color has-blue-background-color has-text-color has-background has-link-color wp-element-button\" href=\"https:\/\/ai.azure.com\/labs\/projects\/magentic-ui?\" target=\"_blank\" rel=\"noreferrer noopener\">Explore Magentic-UI on Azure AI Foundry<\/a><\/div>\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>Magentic-UI, new from Microsoft Research, is an open-source research prototype of a human-centered AI agent, designed to work with people to complete complex, web-based tasks in real time over a web browser.<\/p>\n","protected":false},"author":43868,"featured_media":1139340,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"msr-url-field":"","msr-podcast-episode":"","msrModifiedDate":"","msrModifiedDateEnabled":false,"ep_exclude_from_search":false,"_classifai_error":"","msr-author-ordering":null,"msr_hide_image_in_river":null,"footnotes":""},"categories":[1],"tags":[],"research-area":[13556],"msr-region":[],"msr-event-type":[],"msr-locale":[268875],"msr-post-option":[269148,243984,269142],"msr-impact-theme":[],"msr-promo-type":[],"msr-podcast-series":[],"class_list":["post-1138863","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-research-blog","msr-research-area-artificial-intelligence","msr-locale-en_us","msr-post-option-approved-for-river","msr-post-option-blog-homepage-featured","msr-post-option-include-in-river"],"msr_event_details":{"start":"","end":"","location":""},"podcast_url":"","podcast_episode":"","msr_research_lab":[992148],"msr_impact_theme":[],"related-publications":[],"related-downloads":[],"related-videos":[],"related-academic-programs":[],"related-groups":[],"related-projects":[973047],"related-events":[],"related-researchers":[{"type":"user_nicename","value":"Hussein Mozannar","user_id":43671,"display_name":"Hussein Mozannar","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/hmozannar\/\" aria-label=\"Visit the profile page for Hussein Mozannar\">Hussein Mozannar<\/a>","is_active":false,"last_first":"Mozannar, Hussein","people_section":0,"alias":"hmozannar"},{"type":"user_nicename","value":"Gagan Bansal","user_id":41707,"display_name":"Gagan Bansal","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/gaganbansal\/\" aria-label=\"Visit the profile page for Gagan Bansal\">Gagan Bansal<\/a>","is_active":false,"last_first":"Bansal, Gagan","people_section":0,"alias":"gaganbansal"},{"type":"user_nicename","value":"Cheng Tan","user_id":37953,"display_name":"Cheng Tan","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/chetan\/\" aria-label=\"Visit the profile page for Cheng Tan\">Cheng Tan<\/a>","is_active":false,"last_first":"Tan, Cheng","people_section":0,"alias":"chetan"},{"type":"user_nicename","value":"Adam Fourney","user_id":30820,"display_name":"Adam Fourney","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/adamfo\/\" aria-label=\"Visit the profile page for Adam Fourney\">Adam Fourney<\/a>","is_active":false,"last_first":"Fourney, Adam","people_section":0,"alias":"adamfo"},{"type":"user_nicename","value":"Victor Dibia","user_id":41311,"display_name":"Victor Dibia","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/victordibia\/\" aria-label=\"Visit the profile page for Victor Dibia\">Victor Dibia<\/a>","is_active":false,"last_first":"Dibia, Victor","people_section":0,"alias":"victordibia"},{"type":"user_nicename","value":"Friederike Niedtner","user_id":39919,"display_name":"Friederike Niedtner","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/fniedtner\/\" aria-label=\"Visit the profile page for Friederike Niedtner\">Friederike Niedtner<\/a>","is_active":false,"last_first":"Niedtner, Friederike","people_section":0,"alias":"fniedtner"},{"type":"user_nicename","value":"Jacob Alber","user_id":36747,"display_name":"Jacob Alber","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/jaalber\/\" aria-label=\"Visit the profile page for Jacob Alber\">Jacob Alber<\/a>","is_active":false,"last_first":"Alber, Jacob","people_section":0,"alias":"jaalber"},{"type":"guest","value":"jingya-chen","user_id":"767776","display_name":"Jingya Chen","author_link":"<a href=\"https:\/\/www.linkedin.com\/in\/jingya-chen-cc\/\" aria-label=\"Visit the profile page for Jingya Chen\">Jingya Chen<\/a>","is_active":true,"last_first":"Chen, Jingya","people_section":0,"alias":"jingya-chen"},{"type":"user_nicename","value":"Maya Murad","user_id":43879,"display_name":"Maya Murad","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/mayamurad\/\" aria-label=\"Visit the profile page for Maya Murad\">Maya Murad<\/a>","is_active":false,"last_first":"Murad, Maya","people_section":0,"alias":"mayamurad"},{"type":"user_nicename","value":"Rafah Hosn","user_id":36783,"display_name":"Rafah Hosn","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/raaboulh\/\" aria-label=\"Visit the profile page for Rafah Hosn\">Rafah Hosn<\/a>","is_active":false,"last_first":"Hosn, Rafah","people_section":0,"alias":"raaboulh"},{"type":"user_nicename","value":"Ece Kamar","user_id":31710,"display_name":"Ece Kamar","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/eckamar\/\" aria-label=\"Visit the profile page for Ece Kamar\">Ece Kamar<\/a>","is_active":false,"last_first":"Kamar, Ece","people_section":0,"alias":"eckamar"},{"type":"user_nicename","value":"Saleema Amershi","user_id":33505,"display_name":"Saleema Amershi","author_link":"<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/samershi\/\" aria-label=\"Visit the profile page for Saleema Amershi\">Saleema Amershi<\/a>","is_active":false,"last_first":"Amershi, Saleema","people_section":0,"alias":"samershi"}],"msr_type":"Post","featured_image_thumbnail":"<img width=\"960\" height=\"540\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/MagenticUI-BlogHeroFeature-1400x788-1-960x540.jpg\" class=\"img-object-cover\" alt=\"This figure denotes a human figure above a small monitor on the left and a gear on the right with arrows pointing to each.\" decoding=\"async\" loading=\"lazy\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/MagenticUI-BlogHeroFeature-1400x788-1-960x540.jpg 960w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/MagenticUI-BlogHeroFeature-1400x788-1-300x169.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/MagenticUI-BlogHeroFeature-1400x788-1-1024x576.jpg 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/MagenticUI-BlogHeroFeature-1400x788-1-768x432.jpg 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/MagenticUI-BlogHeroFeature-1400x788-1-1066x600.jpg 1066w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/MagenticUI-BlogHeroFeature-1400x788-1-655x368.jpg 655w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/MagenticUI-BlogHeroFeature-1400x788-1-240x135.jpg 240w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/MagenticUI-BlogHeroFeature-1400x788-1-640x360.jpg 640w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/MagenticUI-BlogHeroFeature-1400x788-1-1280x720.jpg 1280w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/05\/MagenticUI-BlogHeroFeature-1400x788-1.jpg 1400w\" sizes=\"auto, (max-width: 960px) 100vw, 960px\" \/>","byline":"","formattedDate":"May 19, 2025","formattedExcerpt":"Magentic-UI, new from Microsoft Research, is an open-source research prototype of a human-centered AI agent, designed to work with people to complete complex, web-based tasks in real time over a web browser.","locale":{"slug":"en_us","name":"English","native":"","english":"English"},"_links":{"self":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts\/1138863","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/users\/43868"}],"replies":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/comments?post=1138863"}],"version-history":[{"count":44,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts\/1138863\/revisions"}],"predecessor-version":[{"id":1145104,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts\/1138863\/revisions\/1145104"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media\/1139340"}],"wp:attachment":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media?parent=1138863"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/categories?post=1138863"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/tags?post=1138863"},{"taxonomy":"msr-research-area","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/research-area?post=1138863"},{"taxonomy":"msr-region","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-region?post=1138863"},{"taxonomy":"msr-event-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event-type?post=1138863"},{"taxonomy":"msr-locale","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-locale?post=1138863"},{"taxonomy":"msr-post-option","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-post-option?post=1138863"},{"taxonomy":"msr-impact-theme","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-impact-theme?post=1138863"},{"taxonomy":"msr-promo-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-promo-type?post=1138863"},{"taxonomy":"msr-podcast-series","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-podcast-series?post=1138863"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}