{"id":1139371,"date":"2025-06-04T10:56:30","date_gmt":"2025-06-04T17:56:30","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/?post_type=msr-event&#038;p=1139371"},"modified":"2025-06-05T08:11:11","modified_gmt":"2025-06-05T15:11:11","slug":"embodied-ai-workshop-at-cvpr-2025","status":"publish","type":"msr-event","link":"https:\/\/www.microsoft.com\/en-us\/research\/event\/embodied-ai-workshop-at-cvpr-2025\/","title":{"rendered":"Embodied AI Workshop at CVPR 2025"},"content":{"rendered":"\n\n\n\n\n<h2 class=\"wp-block-heading\" id=\"about-the-workshop\">About the workshop<\/h2>\n\n\n\n<p><strong>Host conference:<\/strong> <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/cvpr.thecvf.com\/Conferences\/2025\" target=\"_blank\" rel=\"noopener noreferrer\">The Conference on Computer Vision and Pattern Recognition (CVPR)<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> | June 11-15, 2025<\/p>\n\n\n\n<p><strong>Panel speakers:<\/strong> <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/jianwyan\/\">Jianwei Yang<\/a><\/p>\n\n\n\n<p><strong>Workshop scientific advisor<\/strong>:&nbsp;<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/adfamoti\/\">Ade Famoti<\/a>, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/akolobov\/\">Andrey Kolobov<\/a><\/p>\n\n\n\n<p><strong>Workshop organizer<\/strong>:&nbsp;<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/aminvivan\/\">Vivan Amin<\/a>, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/jiaoyan\/\">Jiaolong Yang<\/a> (<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/embodied-ai.org\/#organizers\" target=\"_blank\" rel=\"noopener noreferrer\">see all workshop organizers<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>)<\/p>\n\n\n\n<p>The <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/embodied-ai.org\/cvpr2025\/\">Embodied AI Workshop at CVPR 2025<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, will be held in conjunction with The Conference on Computer Vision and Pattern Recognition (CVPR) in Nashville, Tennessee. This year&#8217;s workshop focuses on the overarching theme of <strong>Real-World Applications<\/strong>: creating embodied AI solutions that are deployed in real-world environments, ideally in the service of real-world tasks. As embodied AI agents mature, the community is encouraged to promote work that transitions research from simulation and laboratory settings into practical, real-world applications.<\/p>\n\n\n\n<p>This umbrella theme is divided into four topics:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Embodied AI Solutions:<\/strong> As embodied AI solutions become more powerful, they should address more complex problems, particularly real-world challenges outside of simulation and the laboratory. While scientific advances are of interest, we actively seek work that applies embodied AI to real-world industry applications.<\/li>\n\n\n\n<li><strong>Advances in Simulation:<\/strong> Advances in simulation have enabled many embodied AI algorithms. Procedural simulation, parameterized simulation, differentiable simulation, and world models are of interest, as are simulations based on the increasing numbers of large embodied datasets.<\/li>\n\n\n\n<li><strong>Generative Methods for Embodied AI:<\/strong> Generative AI is becoming increasingly important for embodied artificial intelligence research. Topics such as generative AI for simulation, data generation, and policies (e.g., diffusion policies and world models) are of great interest.<\/li>\n\n\n\n<li><strong>Foundation Models:<\/strong> Large-scale pretrained models adaptable to new tasks first emerged in language, speech, and vision domains. Increasingly, foundation models are being developed in robotics domains, including action, perception, problem-solving, and simulation. We invite research on adapting existing models to embodied problems and training embodied foundation models directly on such tasks.<\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"agenda\">Agenda<\/h2>\n\n\n\n<p>A detailed agenda will be posted on this page as soon as it&#8217;s available.<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Workshop talks: 8:50AM-5:30PM PT<\/li>\n\n\n\n<li>Poster session: 1:00PM-2:00PM PT<\/li>\n<\/ul>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"challenges\">Challenges<\/h2>\n\n\n\n<p>The Embodied AI Workshop at CVPR 2025 will host several challenges:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Social Mobile Manipulation Challenge:<\/strong> Developing embodied AI agents capable of performing long sequences of complex tasks through social interactions in dynamic, multi-agent environments.<\/li>\n\n\n\n<li><strong>Multi-Object Rescue Challenge:<\/strong> Tasks involving reasoning about human intentions and planning within dynamic environments.<\/li>\n\n\n\n<li><strong>Vision-Tactile Fusion Manipulation Challenge:<\/strong> Focusing on integrating vision and tactile signals for manipulation tasks.<\/li>\n\n\n\n<li><strong>Open Vocabulary Mobile Manipulation Challenge:<\/strong> Encouraging agents to perform tasks using open vocabulary instructions.<\/li>\n<\/ul>\n\n\n\n<p>Each challenge will have its own dataset, evaluation criteria, and submission guidelines. Winners will be announced during the workshop and may be invited to present their work.<\/p>\n\n\n\n<div class=\"wp-block-buttons is-layout-flex wp-block-buttons-is-layout-flex\">\n<div class=\"wp-block-button\"><a data-bi-type=\"button\" class=\"wp-block-button__link wp-element-button\" href=\"https:\/\/embodied-ai.org\/cvpr2025\/#workshop-schedule\" target=\"_blank\" rel=\"noreferrer noopener\">View the schedule<\/a><\/div>\n<\/div>\n\n\n\n\n\n<h2 class=\"wp-block-heading\" id=\"agenda-june-12-2025\">Agenda: June 12, 2025<\/h2>\n\n\n\n<figure class=\"wp-block-table\"><table><thead><tr><th>Time<\/th><th>Description<\/th><\/tr><\/thead><tbody><tr><td>08:50-09:00<\/td><td><strong>Welcome<\/strong><br>Moderator &#8211; <\/td><\/tr><tr><td>09:00-09:30<\/td><td><strong>Title<\/strong><br>Name, Org<\/td><\/tr><tr><td>09:30-10:00<\/td><td><strong>Title<\/strong><br>Name, Org<\/td><\/tr><tr><td>10:00-10:30<\/td><td><strong>Title<\/strong><br>Name, Org<\/td><\/tr><tr><td>10:30-11:00<\/td><td><strong>Invited Panel &#8211; <\/strong><br>Jianwei Yang, Microsoft<\/td><\/tr><tr><td>11:00-11:30<\/td><td><strong>Title<\/strong><br>Name, Org<\/td><\/tr><tr><td>11:30-12:00<\/td><td><strong>Title<\/strong><br>Name, Org<\/td><\/tr><tr><td>12:00-13:00<\/td><td>Lunch<\/td><\/tr><tr><td>13:00-14:00<\/td><td><strong>Accepted Papers Poster Session<\/strong><\/td><\/tr><tr><td>14:00-14:30<\/td><td><strong>Title<\/strong><br>Name, Org<\/td><\/tr><tr><td>14:30-15:30<\/td><td><strong><strong>Title<\/strong><br>Name, Org<\/strong><\/td><\/tr><tr><td>15:30-16:00<\/td><td><strong>Title<\/strong><br>Name, Org<\/td><\/tr><tr><td>16:00-16:30<\/td><td><strong>Title<\/strong><br>Name, Org<\/td><\/tr><tr><td>16:30-17:00<\/td><td><strong>Title<\/strong><br>Name, Org<\/td><\/tr><tr><td>17:00-17:30<\/td><td><strong>Title<\/strong><br>Name, Org<\/td><\/tr><\/tbody><\/table><\/figure>\n\n\n\n<div style=\"height:30px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n","protected":false},"excerpt":{"rendered":"<p>Host conference: The Conference on Computer Vision and Pattern Recognition (CVPR) (opens in new tab) | June 11-15, 2025 Panel speakers: Jianwei Yang Workshop scientific advisor:&nbsp;Ade Famoti, Andrey Kolobov Workshop organizer:&nbsp;Vivan Amin, Jiaolong Yang (see all workshop organizers (opens in new tab)) The Embodied AI Workshop at CVPR 2025 (opens in new tab), will be [&hellip;]<\/p>\n","protected":false},"featured_media":1034910,"template":"","meta":{"msr-url-field":"","msr-podcast-episode":"","msrModifiedDate":"","msrModifiedDateEnabled":false,"ep_exclude_from_search":false,"_classifai_error":"","msr_startdate":"2025-06-12","msr_enddate":"2025-06-12","msr_location":"Nashville, Tennessee","msr_expirationdate":"","msr_event_recording_link":"","msr_event_link":"","msr_event_link_redirect":false,"msr_event_time":"Central Daylight Time (UTC - 5)","msr_hide_region":false,"msr_private_event":false,"msr_hide_image_in_river":null,"footnotes":""},"research-area":[13556],"msr-region":[],"msr-event-type":[210063],"msr-video-type":[],"msr-locale":[268875],"msr-program-audience":[],"msr-post-option":[269148,269142],"msr-impact-theme":[],"class_list":["post-1139371","msr-event","type-msr-event","status-publish","has-post-thumbnail","hentry","msr-research-area-artificial-intelligence","msr-event-type-workshop","msr-locale-en_us","msr-post-option-approved-for-river","msr-post-option-include-in-river"],"msr_about":"<!-- wp:msr\/event-details {\"title\":\"Embodied AI Workshop at CVPR 2025\",\"subTitle\":\"An ACM Interaction Design and Children (IDC) 2024 Workshop\",\"image\":{\"id\":1034910,\"url\":\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/05\/CVPR_WebBanner_1920x720.png\",\"alt\":\"teal background triangular pattern\"}} \/-->\n\n<!-- wp:msr\/content-tabs -->\n<!-- wp:msr\/content-tab -->\n<!-- wp:heading -->\n<h2 class=\"wp-block-heading\" id=\"about-the-workshop\">About the workshop<\/h2>\n<!-- \/wp:heading -->\n\n<!-- wp:paragraph -->\n<p><strong>Host conference:<\/strong> <a href=\"https:\/\/cvpr.thecvf.com\/Conferences\/2025\" target=\"_blank\" rel=\"noreferrer noopener\">The Conference on Computer Vision and Pattern Recognition (CVPR)<\/a> | June 11-15, 2025<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p><strong>Panel speakers:<\/strong> <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/jianwyan\/\">Jianwei Yang<\/a><\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p><strong>Workshop scientific advisor<\/strong>:&nbsp;<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/adfamoti\/\">Ade Famoti<\/a>, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/akolobov\/\">Andrey Kolobov<\/a><\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p><strong>Workshop organizer<\/strong>:&nbsp;<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/aminvivan\/\">Vivan Amin<\/a>, <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/jiaoyan\/\">Jiaolong Yang<\/a> (<a href=\"https:\/\/embodied-ai.org\/#organizers\" target=\"_blank\" rel=\"noreferrer noopener\">see all workshop organizers<\/a>)<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p>The <a href=\"https:\/\/embodied-ai.org\/cvpr2025\/\">Embodied AI Workshop at CVPR 2025<\/a>, will be held in conjunction with The Conference on Computer Vision and Pattern Recognition (CVPR) in Nashville, Tennessee. This year's workshop focuses on the overarching theme of <strong>Real-World Applications<\/strong>: creating embodied AI solutions that are deployed in real-world environments, ideally in the service of real-world tasks. As embodied AI agents mature, the community is encouraged to promote work that transitions research from simulation and laboratory settings into practical, real-world applications.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p>This umbrella theme is divided into four topics:<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:list -->\n<ul class=\"wp-block-list\"><!-- wp:list-item -->\n<li><strong>Embodied AI Solutions:<\/strong> As embodied AI solutions become more powerful, they should address more complex problems, particularly real-world challenges outside of simulation and the laboratory. While scientific advances are of interest, we actively seek work that applies embodied AI to real-world industry applications.<\/li>\n<!-- \/wp:list-item -->\n\n<!-- wp:list-item -->\n<li><strong>Advances in Simulation:<\/strong> Advances in simulation have enabled many embodied AI algorithms. Procedural simulation, parameterized simulation, differentiable simulation, and world models are of interest, as are simulations based on the increasing numbers of large embodied datasets.<\/li>\n<!-- \/wp:list-item -->\n\n<!-- wp:list-item -->\n<li><strong>Generative Methods for Embodied AI:<\/strong> Generative AI is becoming increasingly important for embodied artificial intelligence research. Topics such as generative AI for simulation, data generation, and policies (e.g., diffusion policies and world models) are of great interest.<\/li>\n<!-- \/wp:list-item -->\n\n<!-- wp:list-item -->\n<li><strong>Foundation Models:<\/strong> Large-scale pretrained models adaptable to new tasks first emerged in language, speech, and vision domains. Increasingly, foundation models are being developed in robotics domains, including action, perception, problem-solving, and simulation. We invite research on adapting existing models to embodied problems and training embodied foundation models directly on such tasks.<\/li>\n<!-- \/wp:list-item --><\/ul>\n<!-- \/wp:list -->\n\n<!-- wp:heading -->\n<h2 class=\"wp-block-heading\" id=\"agenda\">Agenda<\/h2>\n<!-- \/wp:heading -->\n\n<!-- wp:paragraph -->\n<p>A detailed agenda will be posted on this page as soon as it's available.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:list -->\n<ul class=\"wp-block-list\"><!-- wp:list-item -->\n<li>Workshop talks: 8:50AM-5:30PM PT<\/li>\n<!-- \/wp:list-item -->\n\n<!-- wp:list-item -->\n<li>Poster session: 1:00PM-2:00PM PT<\/li>\n<!-- \/wp:list-item --><\/ul>\n<!-- \/wp:list -->\n\n<!-- wp:heading -->\n<h2 class=\"wp-block-heading\" id=\"challenges\">Challenges<\/h2>\n<!-- \/wp:heading -->\n\n<!-- wp:paragraph -->\n<p>The Embodied AI Workshop at CVPR 2025 will host several challenges:<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:list -->\n<ul class=\"wp-block-list\"><!-- wp:list-item -->\n<li><strong>Social Mobile Manipulation Challenge:<\/strong> Developing embodied AI agents capable of performing long sequences of complex tasks through social interactions in dynamic, multi-agent environments.<\/li>\n<!-- \/wp:list-item -->\n\n<!-- wp:list-item -->\n<li><strong>Multi-Object Rescue Challenge:<\/strong> Tasks involving reasoning about human intentions and planning within dynamic environments.<\/li>\n<!-- \/wp:list-item -->\n\n<!-- wp:list-item -->\n<li><strong>Vision-Tactile Fusion Manipulation Challenge:<\/strong> Focusing on integrating vision and tactile signals for manipulation tasks.<\/li>\n<!-- \/wp:list-item -->\n\n<!-- wp:list-item -->\n<li><strong>Open Vocabulary Mobile Manipulation Challenge:<\/strong> Encouraging agents to perform tasks using open vocabulary instructions.<\/li>\n<!-- \/wp:list-item --><\/ul>\n<!-- \/wp:list -->\n\n<!-- wp:paragraph -->\n<p>Each challenge will have its own dataset, evaluation criteria, and submission guidelines. Winners will be announced during the workshop and may be invited to present their work.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:buttons -->\n<div class=\"wp-block-buttons\"><!-- wp:button -->\n<div class=\"wp-block-button\"><a class=\"wp-block-button__link wp-element-button\" href=\"https:\/\/embodied-ai.org\/cvpr2025\/#workshop-schedule\" target=\"_blank\" rel=\"noreferrer noopener\">View the schedule<\/a><\/div>\n<!-- \/wp:button --><\/div>\n<!-- \/wp:buttons -->\n<!-- \/wp:msr\/content-tab -->\n\n<!-- wp:msr\/content-tab {\"title\":\"Agenda\",\"visible\":false} -->\n<!-- wp:heading -->\n<h2 class=\"wp-block-heading\" id=\"agenda-june-12-2025\">Agenda: June 12, 2025<\/h2>\n<!-- \/wp:heading -->\n\n<!-- wp:table {\"hasFixedLayout\":false} -->\n<figure class=\"wp-block-table\"><table><thead><tr><th>Time<\/th><th>Description<\/th><\/tr><\/thead><tbody><tr><td>08:50-09:00<\/td><td><strong>Welcome<\/strong><br>Moderator - <\/td><\/tr><tr><td>09:00-09:30<\/td><td><strong>Title<\/strong><br>Name, Org<\/td><\/tr><tr><td>09:30-10:00<\/td><td><strong>Title<\/strong><br>Name, Org<\/td><\/tr><tr><td>10:00-10:30<\/td><td><strong>Title<\/strong><br>Name, Org<\/td><\/tr><tr><td>10:30-11:00<\/td><td><strong>Invited Panel - <\/strong><br>Jianwei Yang, Microsoft<\/td><\/tr><tr><td>11:00-11:30<\/td><td><strong>Title<\/strong><br>Name, Org<\/td><\/tr><tr><td>11:30-12:00<\/td><td><strong>Title<\/strong><br>Name, Org<\/td><\/tr><tr><td>12:00-13:00<\/td><td>Lunch<\/td><\/tr><tr><td>13:00-14:00<\/td><td><strong>Accepted Papers Poster Session<\/strong><\/td><\/tr><tr><td>14:00-14:30<\/td><td><strong>Title<\/strong><br>Name, Org<\/td><\/tr><tr><td>14:30-15:30<\/td><td><strong><strong>Title<\/strong><br>Name, Org<\/strong><\/td><\/tr><tr><td>15:30-16:00<\/td><td><strong>Title<\/strong><br>Name, Org<\/td><\/tr><tr><td>16:00-16:30<\/td><td><strong>Title<\/strong><br>Name, Org<\/td><\/tr><tr><td>16:30-17:00<\/td><td><strong>Title<\/strong><br>Name, Org<\/td><\/tr><tr><td>17:00-17:30<\/td><td><strong>Title<\/strong><br>Name, Org<\/td><\/tr><\/tbody><\/table><\/figure>\n<!-- \/wp:table -->\n\n<!-- wp:spacer {\"height\":\"30px\"} -->\n<div style=\"height:30px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n<!-- \/wp:spacer -->\n<!-- \/wp:msr\/content-tab -->\n<!-- \/wp:msr\/content-tabs -->","tab-content":[],"msr_startdate":"2025-06-12","msr_enddate":"2025-06-12","msr_event_time":"Central Daylight Time (UTC - 5)","msr_location":"Nashville, Tennessee","msr_event_link":"","msr_event_recording_link":"","msr_startdate_formatted":"June 12, 2025","msr_register_text":"Watch now","msr_cta_link":"","msr_cta_text":"","msr_cta_bi_name":"","featured_image_thumbnail":"<img width=\"960\" height=\"540\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/05\/CVPR_WebBanner_1920x720-960x540.png\" class=\"img-object-cover\" alt=\"teal background triangular pattern\" decoding=\"async\" loading=\"lazy\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/05\/CVPR_WebBanner_1920x720-960x540.png 960w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/05\/CVPR_WebBanner_1920x720-1066x600.png 1066w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/05\/CVPR_WebBanner_1920x720-655x368.png 655w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/05\/CVPR_WebBanner_1920x720-640x360.png 640w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/05\/CVPR_WebBanner_1920x720-1280x720.png 1280w\" sizes=\"auto, (max-width: 960px) 100vw, 960px\" \/>","event_excerpt":"Host conference: The Conference on Computer Vision and Pattern Recognition (CVPR) (opens in new tab) | June 11-15, 2025 Panel speakers: Jianwei Yang Workshop scientific advisor:&nbsp;Ade Famoti, Andrey Kolobov Workshop organizer:&nbsp;Vivan Amin, Jiaolong Yang (see all workshop organizers (opens in new tab)) The Embodied AI Workshop at CVPR 2025 (opens in new tab), will be held in conjunction with The Conference on Computer Vision and Pattern Recognition (CVPR) in Nashville, Tennessee. This year's workshop focuses&hellip;","msr_research_lab":[],"related-researchers":[],"msr_impact_theme":[],"related-academic-programs":[],"related-groups":[1057371],"related-projects":[],"related-opportunities":[],"related-publications":[],"related-videos":[],"related-posts":[],"_links":{"self":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event\/1139371","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event"}],"about":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/types\/msr-event"}],"version-history":[{"count":9,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event\/1139371\/revisions"}],"predecessor-version":[{"id":1141266,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event\/1139371\/revisions\/1141266"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media\/1034910"}],"wp:attachment":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media?parent=1139371"}],"wp:term":[{"taxonomy":"msr-research-area","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/research-area?post=1139371"},{"taxonomy":"msr-region","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-region?post=1139371"},{"taxonomy":"msr-event-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event-type?post=1139371"},{"taxonomy":"msr-video-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-video-type?post=1139371"},{"taxonomy":"msr-locale","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-locale?post=1139371"},{"taxonomy":"msr-program-audience","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-program-audience?post=1139371"},{"taxonomy":"msr-post-option","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-post-option?post=1139371"},{"taxonomy":"msr-impact-theme","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-impact-theme?post=1139371"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}