{"id":876336,"date":"2022-10-04T08:58:53","date_gmt":"2022-10-04T15:58:53","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/?post_type=msr-event&#038;p=876336"},"modified":"2022-10-25T08:48:10","modified_gmt":"2022-10-25T15:48:10","slug":"creation-for-rich-world-workshop","status":"publish","type":"msr-event","link":"https:\/\/www.microsoft.com\/en-us\/research\/event\/creation-for-rich-world-workshop\/","title":{"rendered":"Creation for Rich World Workshop"},"content":{"rendered":"\n\n\n\n\n<p>As our world goes more and more digitalized (especially in the era of metaverse), there is an increasing interest to automatically create content (visual, audio, text, etc) to provide media-rich and immersive experiences.<\/p>\n\n\n\n<p>This workshop invites top researchers and academics in this area to discuss the recent progress, share the thoughts, and envision the future of content creation for rich world. Discussion topics include:&nbsp;&nbsp;<\/p>\n\n\n\n<ul class=\"wp-block-list\"><li>Speech, music, sound, and spatial audio creation<\/li><li>Image, video, scene, and environment creation<\/li><li>Digital human creation<\/li><\/ul>\n\n\n\n<p>The goal of this workshop is to discuss, communicate, share, and learn with each other on the topics of content creation, which can help better understand this area and come up with future research opportunities.<\/p>\n\n\n\n<h2 id=\"speakers\">Speakers<\/h2>\n\n\n\n<div class=\"wp-block-columns is-layout-flex wp-container-core-columns-is-layout-9d6595d7 wp-block-columns-is-layout-flex\">\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<figure class=\"wp-block-image aligncenter size-thumbnail is-style-rounded\"><img loading=\"lazy\" decoding=\"async\" width=\"150\" height=\"150\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/08\/tadas-baltrusaitis-150x150.jpg\" alt=\"Tadas Baltrusaitis\" class=\"wp-image-868479\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/08\/tadas-baltrusaitis-150x150.jpg 150w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/08\/tadas-baltrusaitis-180x180.jpg 180w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/08\/tadas-baltrusaitis.jpg 256w\" sizes=\"auto, (max-width: 150px) 100vw, 150px\" \/><\/figure>\n\n\n\n<div style=\"height:7px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<h5 id=\"tadas-baltrusaitis\" class=\"has-text-align-center\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/tabaltru\/\" target=\"_blank\" rel=\"noreferrer noopener\">Tadas Baltrusaitis<\/a><\/h5>\n\n\n\n<p class=\"has-text-align-center\">Principal Scientist<br><em>Microsoft Mixed Reality and AI Lab<\/em><\/p>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<figure class=\"wp-block-image aligncenter size-thumbnail is-style-rounded\"><img loading=\"lazy\" decoding=\"async\" width=\"150\" height=\"150\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Jesse-Engel_360x360-150x150.jpg\" alt=\"Jesse Engel - Research Summit speaker\" class=\"wp-image-881673\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Jesse-Engel_360x360-150x150.jpg 150w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Jesse-Engel_360x360-300x300.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Jesse-Engel_360x360-180x180.jpg 180w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Jesse-Engel_360x360.jpg 360w\" sizes=\"auto, (max-width: 150px) 100vw, 150px\" \/><\/figure>\n\n\n\n<div style=\"height:7px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<h5 id=\"jesse-engel-opens-in-new-tab\" class=\"has-text-align-center\"><a href=\"https:\/\/research.google\/people\/JesseEngel\/\" target=\"_blank\" rel=\"noreferrer noopener\">Jesse Engel<\/a><\/h5>\n\n\n\n<p class=\"has-text-align-center\">Staff Research Scientist<br><em>Google Research<\/em><\/p>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<figure class=\"wp-block-image aligncenter size-thumbnail is-style-rounded\"><img loading=\"lazy\" decoding=\"async\" width=\"150\" height=\"150\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Jia-Tao_360x360-150x150.jpg\" alt=\"Jiatao Gu - Research Summit speaker\" class=\"wp-image-881676\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Jia-Tao_360x360-150x150.jpg 150w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Jia-Tao_360x360-300x300.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Jia-Tao_360x360-180x180.jpg 180w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Jia-Tao_360x360.jpg 360w\" sizes=\"auto, (max-width: 150px) 100vw, 150px\" \/><\/figure>\n\n\n\n<div style=\"height:7px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<h5 id=\"jiatao-gu-opens-in-new-tab\" class=\"has-text-align-center\"><a href=\"https:\/\/jiataogu.me\/\" target=\"_blank\" rel=\"noreferrer noopener\">Jiatao Gu<\/a><\/h5>\n\n\n\n<p class=\"has-text-align-center\">Researcher<br><em>Apple<\/em><\/p>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<figure class=\"wp-block-image aligncenter size-thumbnail is-style-rounded\"><img loading=\"lazy\" decoding=\"async\" width=\"150\" height=\"150\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Xiaoguang-Han_360x360-150x150.jpg\" alt=\"Xiaoguang Han - Research Summit speaker\" class=\"wp-image-881682\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Xiaoguang-Han_360x360-150x150.jpg 150w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Xiaoguang-Han_360x360-300x300.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Xiaoguang-Han_360x360-180x180.jpg 180w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Xiaoguang-Han_360x360.jpg 360w\" sizes=\"auto, (max-width: 150px) 100vw, 150px\" \/><\/figure>\n\n\n\n<div style=\"height:7px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<h5 id=\"xiaoguang-han-opens-in-new-tab\" class=\"has-text-align-center\"><a href=\"https:\/\/gaplab.cuhk.edu.cn\/\" target=\"_blank\" rel=\"noreferrer noopener\">Xiaoguang Han<\/a><\/h5>\n\n\n\n<p class=\"has-text-align-center\">Assistant Professor<br><em>The Chinese University of Hong Kong, Shenzhen<\/em><\/p>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<figure class=\"wp-block-image aligncenter size-thumbnail is-style-rounded\"><img loading=\"lazy\" decoding=\"async\" width=\"150\" height=\"150\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Seungyong-Lee_360x360-150x150.jpg\" alt=\"Seungyong Lee - Research Summit speaker\" class=\"wp-image-881679\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Seungyong-Lee_360x360-150x150.jpg 150w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Seungyong-Lee_360x360-300x300.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Seungyong-Lee_360x360-180x180.jpg 180w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Seungyong-Lee_360x360.jpg 360w\" sizes=\"auto, (max-width: 150px) 100vw, 150px\" \/><\/figure>\n\n\n\n<div style=\"height:7px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<h5 id=\"seungyong-lee-opens-in-new-tab\" class=\"has-text-align-center\"><a href=\"http:\/\/cg.postech.ac.kr\/leesy\/\" target=\"_blank\" rel=\"noreferrer noopener\">Seungyong Lee<\/a><\/h5>\n\n\n\n<p class=\"has-text-align-center\">Professor<br><em>POSTECH<\/em><\/p>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<figure class=\"wp-block-image aligncenter size-thumbnail is-style-rounded\"><img loading=\"lazy\" decoding=\"async\" width=\"150\" height=\"150\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Yang-Liu_360x360-150x150.jpg\" alt=\"Yang Liu - Research Summit speaker\" class=\"wp-image-881649\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Yang-Liu_360x360-150x150.jpg 150w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Yang-Liu_360x360-300x300.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Yang-Liu_360x360-180x180.jpg 180w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Yang-Liu_360x360.jpg 360w\" sizes=\"auto, (max-width: 150px) 100vw, 150px\" \/><\/figure>\n\n\n\n<div style=\"height:7px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<h5 id=\"yang-liu\" class=\"has-text-align-center\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/yangliu\/\" target=\"_blank\" rel=\"noreferrer noopener\">Yang Liu<\/a><\/h5>\n\n\n\n<p class=\"has-text-align-center\">Principal Researcher<br><em>Microsoft Research Asia<\/em><\/p>\n<\/div>\n<\/div>\n\n\n\n<div class=\"wp-block-columns is-layout-flex wp-container-core-columns-is-layout-9d6595d7 wp-block-columns-is-layout-flex\">\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<figure class=\"wp-block-image aligncenter size-thumbnail is-style-rounded\"><img loading=\"lazy\" decoding=\"async\" width=\"150\" height=\"150\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Wei-Ping_square-150x150.jpg\" alt=\"Wei Ping - Research Summit speaker\" class=\"wp-image-881700\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Wei-Ping_square-150x150.jpg 150w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Wei-Ping_square-180x180.jpg 180w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Wei-Ping_square.jpg 192w\" sizes=\"auto, (max-width: 150px) 100vw, 150px\" \/><\/figure>\n\n\n\n<div style=\"height:7px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<h5 id=\"wei-ping-opens-in-new-tab\" class=\"has-text-align-center\"><a href=\"https:\/\/wpingnet.github.io\/\" target=\"_blank\" rel=\"noreferrer noopener\">Wei Ping<\/a><\/h5>\n\n\n\n<p class=\"has-text-align-center\">Principal Research Scientist<br><em>NVIDIA<\/em><\/p>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<figure class=\"wp-block-image aligncenter size-thumbnail is-style-rounded\"><img loading=\"lazy\" decoding=\"async\" width=\"150\" height=\"150\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Alexander-Richard_360x360-150x150.jpg\" alt=\"Alexander Richard - Research Summit speaker\" class=\"wp-image-881667\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Alexander-Richard_360x360-150x150.jpg 150w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Alexander-Richard_360x360-300x300.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Alexander-Richard_360x360-180x180.jpg 180w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Alexander-Richard_360x360.jpg 360w\" sizes=\"auto, (max-width: 150px) 100vw, 150px\" \/><\/figure>\n\n\n\n<div style=\"height:7px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<h5 id=\"alexander-richard-opens-in-new-tab\" class=\"has-text-align-center\"><a href=\"https:\/\/alexanderrichard.github.io\/\" target=\"_blank\" rel=\"noreferrer noopener\">Alexander Richard<\/a><\/h5>\n\n\n\n<p class=\"has-text-align-center\">Research Scientist<br><em>Meta Reality Labs Research<\/em><\/p>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<figure class=\"wp-block-image aligncenter size-thumbnail is-style-rounded\"><img loading=\"lazy\" decoding=\"async\" width=\"150\" height=\"150\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Xu-Tan_360x360-150x150.jpg\" alt=\"Xu Tan - Research Summit speaker\" class=\"wp-image-881688\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Xu-Tan_360x360-150x150.jpg 150w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Xu-Tan_360x360-300x300.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Xu-Tan_360x360-180x180.jpg 180w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Xu-Tan_360x360.jpg 360w\" sizes=\"auto, (max-width: 150px) 100vw, 150px\" \/><\/figure>\n\n\n\n<div style=\"height:7px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<h5 id=\"xu-tan\" class=\"has-text-align-center\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/xuta\/\" target=\"_blank\" rel=\"noreferrer noopener\">Xu Tan<\/a><\/h5>\n\n\n\n<p class=\"has-text-align-center\">Principal Research Manager<br><em>Microsoft Research Asia<\/em><\/p>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<figure class=\"wp-block-image aligncenter size-thumbnail is-style-rounded\"><img loading=\"lazy\" decoding=\"async\" width=\"150\" height=\"150\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Xin-Tong_square-150x150.jpg\" alt=\"Xin Tong - Research Summit speaker\" class=\"wp-image-881685\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Xin-Tong_square-150x150.jpg 150w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Xin-Tong_square-300x300.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Xin-Tong_square-180x180.jpg 180w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Xin-Tong_square.jpg 310w\" sizes=\"auto, (max-width: 150px) 100vw, 150px\" \/><\/figure>\n\n\n\n<div style=\"height:7px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<h5 id=\"xin-tong\" class=\"has-text-align-center\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/xtong\/\" target=\"_blank\" rel=\"noreferrer noopener\">Xin Tong<\/a><\/h5>\n\n\n\n<p class=\"has-text-align-center\">Partner Research Manager<br><em>Microsoft Research Asia<\/em><\/p>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<figure class=\"wp-block-image aligncenter size-thumbnail is-style-rounded\"><img loading=\"lazy\" decoding=\"async\" width=\"150\" height=\"150\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Baoyuan-Wang_360x360-150x150.jpg\" alt=\"Baoyuan Wang - Research Summit speaker\" class=\"wp-image-881670\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Baoyuan-Wang_360x360-150x150.jpg 150w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Baoyuan-Wang_360x360-300x300.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Baoyuan-Wang_360x360-180x180.jpg 180w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Baoyuan-Wang_360x360.jpg 360w\" sizes=\"auto, (max-width: 150px) 100vw, 150px\" \/><\/figure>\n\n\n\n<div style=\"height:7px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<h5 id=\"baoyuan-wang-opens-in-new-tab\" class=\"has-text-align-center\"><a href=\"https:\/\/www.linkedin.com\/in\/baoyuan-wang-59153424\/\" target=\"_blank\" rel=\"noreferrer noopener\">Baoyuan Wang<\/a><\/h5>\n\n\n\n<p class=\"has-text-align-center\">Cofounder & VP<br><em>Xiaobing.ai<\/em><\/p>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\"><\/div>\n<\/div>\n\n\n\n<div style=\"height:15px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<h2 id=\"agenda\">Agenda<\/h2>\n\n\n\n<figure class=\"wp-block-table is-style-stripes\"><table><thead><tr><th>Time (CST)<\/th><th>Session<\/th><th>Speaker<\/th><\/tr><\/thead><tbody><tr><td>09:30 AM<\/td><td>Welcome & Overview of Workshop<\/td><td><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/xuta\/\">Xu Tan<\/a><br>Principal Research Manager<br><em>Microsoft Research Asia<\/em><\/td><\/tr><tr><td><\/td><td><strong>Session 1: Speech\/Music\/Sound and Spatial Audio Creation<\/strong><\/td><td><\/td><\/tr><tr><td>09:45 AM<\/td><td>Explore the Limit of Zero-shot Audio Synthesis with Large-scale GAN Training<br><br>Abstract: In this talk, I will present some state-of-the-art results for raw audio synthesis. We will compare different family of methods for the universal vocoding task and introduce BigVGAN that can generalize well under various unseen conditions in zero-shot setting.<\/td><td><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/wpingnet.github.io\/\" target=\"_blank\" rel=\"noopener noreferrer\">Wei Ping<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><br>Principal Research Scientist<br><em>NVIDIA<\/em><\/td><\/tr><tr><td>10:05 AM<\/td><td>Magenta: Empowering Creativity in the Age of Machine Learning<br><br>Abstract: Since 2016, the Magenta research group has investigated the role of machine learning in empowering the creativity of artists, musicians, and novices alike. In this talk, we&#8217;ll examine recent advances by the group the role of interpretable decompositions for music understanding and generation, including state-of-the-art models in music transcription (MT3), Composition (Perceiver AR), synthesis (Spectrogram Diffusion), and user interaction (DDSP-VST, MIDI-DDSP). Finally, we&#8217;ll explore how the combination of expressive generative models and intuitive controls can power a new generation of creative tools.<\/td><td><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/research.google\/people\/JesseEngel\/\" target=\"_blank\" rel=\"noopener noreferrer\">Jesse Engel<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><br>Staff Research Scientist<br><em>Google Research<\/em><\/td><\/tr><tr><td>10:25 AM<\/td><td>Neural Audio Rendering for Social Telepresence<br><br>Abstract: These days, physical distance between people is one of the biggest obstacles to maintaining meaningful social relationships with family, friends, and co-workers. Even with today\u2019s technology, remote communication is limited to a two-dimensional audio-visual experience and lacks the availability of a shared, three-dimensional space in which people can interact with each other over the distance. Our mission at Reality Labs Research (RLR) in Pittsburgh is to develop a telepresence system that is indistinguishable from reality, i.e., a system that provides photo- and phono-realistic social interactions in VR. Highly realistic spatial audio rendering is a key ingredient to achieve the desired level of realism. While computer graphics has long moved from traditional rendering to neural rendering, the audio community is just in the early stages of this process. I will discuss the advantages of neural sound rendering and outline the challenges in data collection for those typically data-hungry machine learning approaches. I will further demonstrate that realism and accuracy of neural spatial audio methods exceed traditional signal processing. In the future, these technologies will help build a realistic virtual environment with lifelike avatars that allow for authentic social interactions, connecting people all over the world, anywhere and at any time.<\/td><td><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/alexanderrichard.github.io\/\" target=\"_blank\" rel=\"noopener noreferrer\">Alexander Richard<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><br>Research Scientist<br><em>Meta Reality Labs Research<\/em><\/td><\/tr><tr><td>10:45 AM<\/td><td>Panel Discussions<\/td><td><strong>Host<\/strong>: <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/research.google\/people\/JesseEngel\/\" target=\"_blank\" rel=\"noopener noreferrer\">Jesse Engel<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><br><strong>Panelists<\/strong>:<br><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/wpingnet.github.io\/\" target=\"_blank\" rel=\"noopener noreferrer\">Wei Ping<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/alexanderrichard.github.io\/\" target=\"_blank\" rel=\"noopener noreferrer\">Alexander Richard<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, <br><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/xuta\/\">Xu Tan<\/a><\/td><\/tr><tr><td>11:30 AM<\/td><td><em>Lunch break<\/em><\/td><td><\/td><\/tr><tr><td><\/td><td><strong>Session 2: Image\/Video\/Scene Creation<\/strong><\/td><td><\/td><\/tr><tr><td>12:30 PM<\/td><td>Semantic Instance Reconstruction for 3D Scene Understanding<br><br>Abstract: 3D scene understanding and reconstruction plays very important roles in many application scenarios, like robot perception and also AR\/VR etc. Currently, most of existing works treated the 3D scene as a whole stuff for reconstruction. In this talk, I will introduce our recent techniques that conducted reconstruction together with instance understanding, which is termed as \u201cSemantic Instance Reconstruction\u201d. The main content includes three published works: Total3DUnderstanding (CVPR 2020), RfD-Net (CVPR 2021) and InstPIFu (ECCV 2022). They are all aiming to semantic instance reconstruction, but Total3D focuses on reconstruction from single images while RfD-Net takes point cloud as input, InstPIFu targets high-fidelity single-view Reconstruction. Furthermore, I will also introduce our recent work on 3D table-scene dataset.<\/td><td><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/gaplab.cuhk.edu.cn\/\" target=\"_blank\" rel=\"noopener noreferrer\">Xiaoguang Han<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><br>Assistant Professor<br><em>The Chinese University of Hong Kong, Shenzhen<\/em><\/td><\/tr><tr><td>12:50 PM<\/td><td>Towards High-fidelity 3D Shape and Scene Generation<br><br>Abstract: Digital 3D contents equipped with high-fidelity shape geometry, scene layout, and visual appearance are the basis for building the digital world and facilitating 3D understanding, interaction, and exploration. The main barriers to generating vibrant 3D content include insufficient number of labeled 3D data, diversity of 3D representations, and lack of editability. In this talk, we will present a set of our 3D generation work that incorporates neural-based 3D representations and generative models to overcome these barriers and create high-fidelity 3D shapes and scenes.<\/td><td><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/xueyuhanlang.github.io\/\" target=\"_blank\" rel=\"noopener noreferrer\">Yang Liu<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><br>Principal Researcher<br><em>Microsoft Research Asia<\/em><\/td><\/tr><tr><td>1:10 PM<\/td><td>Diffusion Models for Image and Neural Field Generation<br><br>Abstract: Diffusion probabilistic models have quickly become the de-facto choice for generative modeling of images, text or 3D geometry. In this talk, we will introduce some of our recent works on how to generalize diffusion models for more efficient and controllable image generation, as well as general field data.<\/td><td><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/jiataogu.me\/\" target=\"_blank\" rel=\"noopener noreferrer\">Jiatao Gu<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><br>Researcher<br><em>Apple<\/em><\/td><\/tr><tr><td>1:30 PM<\/td><td>Panel Discussions<\/td><td><strong>Host<\/strong>: <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/xtong\/\">Xin Tong<\/a><br><strong>Panelists<\/strong>: <br><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/gaplab.cuhk.edu.cn\/\" target=\"_blank\" rel=\"noopener noreferrer\">Xiaoguang Han<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, <br><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/xueyuhanlang.github.io\/\" target=\"_blank\" rel=\"noopener noreferrer\">Yang Liu<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, <br><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/jiataogu.me\/\" target=\"_blank\" rel=\"noopener noreferrer\">Jiatao Gu<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/td><\/tr><tr><td>2:15 PM<\/td><td><em>Short break<\/em><\/td><td><br><\/td><\/tr><tr><td><\/td><td><strong>Session 3: Digital Human Creation<\/strong><\/td><td><\/td><\/tr><tr><td>2:25 PM<\/td><td>From Human to AI Being Intelligence: Challenges and Opportunities<br><br>Abstract: Digital human or AI being (named by Xiaobing.ai) has already shown tremendous potential value for helping digital transformations in various industry domains. It is one of the interesting testbeds for building next generations of artificial intelligence. In this talk, from an industry perspective, we will discuss the fundamental technical challenges and share some learning experiences of developing different AI beings under Xiaoice avatar framework.<\/td><td><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/www.linkedin.com\/in\/baoyuan-wang-59153424\/\" target=\"_blank\" rel=\"noopener noreferrer\">Baoyuan Wang<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><br>Cofounder & VP<br><em>Xiaobing.ai<\/em><\/td><\/tr><tr><td>2:45 PM<\/td><td>MLPs for Reconstruction and Control of Explicit and Detailed 3D Human Models<br><br>Abstract: MLPs (Multi-Layer Perceptrons) have been widely used for reconstructing implicit 3D representations of human models. In this talk, in contrast, I will focus on utilizing MLPs for reconstructing and controlling explicit polygonal representations of 3D human models. I will first present an MLP-based framework for building a deformable surface model, which takes a latent code and produces a 3D caricature model. The framework captures the variations of 3D caricatures in a compact parameter space and provides a useful data-driven toolkit for handling 3D caricature deformations. I will then present LaplacianFusion, a novel framework that reconstructs a detailed and controllable 3D clothed human model from a point cloud sequence. In the framework, an MLP is used to learn and predict Laplacian coordinates representing the details on the body surface. The talk will be concluded with discussion on the pros and cons of implicit and explicit representations when they are combined with MLPs for reconstruction and control of 3D human models.<\/td><td><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"http:\/\/cg.postech.ac.kr\/leesy\/\" target=\"_blank\" rel=\"noopener noreferrer\">Seungyong Lee<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><br>Professor<br><em>POSTECH<\/em><\/td><\/tr><tr><td>3:05 PM<\/td><td>Synthetic Data is All You Need: Face Analysis Using Synthetic Data Alone<br><br>Abstract: In this talk I will demonstrate how synthetic data alone can be used to perform face-related computer vision. The community has long enjoyed the benefits of synthesizing training data with graphics, but the domain gap between real and synthetic data has remained a problem, especially for human faces. I will show that it is possible to synthesize data with minimal domain gap, so that models trained on synthetic data generalize to real in-the-wild datasets. I will describe how to combine a procedurally-generated parametric 3D face model with a comprehensive library of hand-crafted assets to render training images with unprecedented realism and diversity. I will show that models trained using synthetic data alone can both match and exceed real data in accuracy as well as open up new approaches where manual labelling would be impossible.<\/td><td><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/tabaltru\/\">Tadas Baltrusaitis<\/a><br>Principal Scientist<br><em>Microsoft Mixed Reality and AI <\/em>l<em>ab<\/em><\/td><\/tr><tr><td>3:25 PM<\/td><td>Panel Discussions<\/td><td><strong>Host<\/strong>: <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/xuta\/\">Xu Tan<\/a><br><strong>Panelists<\/strong>:<br><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/www.linkedin.com\/in\/baoyuan-wang-59153424\/\" target=\"_blank\" rel=\"noopener noreferrer\">Baoyuan Wang<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, <br><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"http:\/\/cg.postech.ac.kr\/leesy\/\" target=\"_blank\" rel=\"noopener noreferrer\">Seungyong Lee<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, <br><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/tabaltru\/\">Tadas Baltrusaitis<\/a>, <br><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/xtong\/\">Xin Tong<\/a><\/td><\/tr><tr><td>4:10 PM<\/td><td>Next Steps\/Closing Remarks<\/td><td><\/td><\/tr><\/tbody><\/table><\/figure>\n\n\n\n<div style=\"height:40px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<h2 id=\"workshop-organizers\">Workshop organizers<\/h2>\n\n\n\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/xuta\/\" target=\"_blank\" rel=\"noreferrer noopener\">Xu Tan<\/a>, Microsoft Research Asia<br><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/xtong\/\" target=\"_blank\" rel=\"noreferrer noopener\">Xin Tong<\/a>, Microsoft Research Asia<br><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/www.linkedin.com\/in\/miran-lee-37996b36\/\" target=\"_blank\" rel=\"noopener noreferrer\">Miran Lee<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, Microsoft Research Asia<\/p>\n\n\n\n<div style=\"height:30px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<h4 id=\"microsofts-event-code-of-conduct\">Microsoft\u2019s Event Code of Conduct<\/h4>\n\n\n\n<p>Microsoft\u2019s mission is to empower every person and every organization on the planet to achieve more. This includes events Microsoft hosts and participates in, where we seek to create a respectful, friendly, and inclusive experience for all participants. As such, we do not tolerate harassing or disrespectful behavior, messages, images, or interactions by any event participant, in any form, at any aspect of the program including business and social activities, regardless of location. <\/p>\n\n\n\n<p>We do not tolerate any behavior that is degrading to any gender, race, sexual orientation or disability, or any behavior that would violate <a href=\"https:\/\/www.microsoft.com\/en-us\/legal\/compliance\/default.aspx\" target=\"_blank\" rel=\"noreferrer noopener\">Microsoft\u2019s Anti-Harassment and Anti-Discrimination Policy, Equal Employment Opportunity Policy, or&nbsp;Standards of Business Conduct<\/a>. In short, the entire experience at the venue must meet our culture standards. We encourage everyone to assist in creating a welcoming and safe environment. Please <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/aka.ms\/reportconcern\" target=\"_blank\" rel=\"noopener noreferrer\">report<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> any concerns, harassing behavior, or suspicious or disruptive activity to venue staff, the event host or owner, or event staff. Microsoft reserves the right to refuse admittance to or remove any person from company-sponsored events at any time in its sole discretion.<\/p>\n\n\n\n<div class=\"wp-block-buttons is-layout-flex wp-block-buttons-is-layout-flex\">\n<div class=\"wp-block-button is-style-outline is-style-outline--1\"><a data-bi-type=\"button\" class=\"wp-block-button__link\" href=\"https:\/\/aka.ms\/reportconcern\" target=\"_blank\" rel=\"noreferrer noopener\">Report a concern<\/a><\/div>\n<\/div>\n\n\n\n\n\n<p><\/p>\n\n\n\n<iframe loading=\"lazy\" width=\"1000px\" height=\"4600px\" src=\"https:\/\/forms.office.com\/r\/LBinzjZBg3;embed=true\" frameborder=\"0\" marginwidth=\"0\" marginheight=\"0\" style=\"border: none; max-width:100%; max-height:100vh\" allowfullscreen=\"\" webkitallowfullscreen=\"\" mozallowfullscreen=\"\" msallowfullscreen=\"\"> <\/iframe>\n\n\n","protected":false},"excerpt":{"rendered":"<p>As our world goes more and more digitalized (especially in the era of metaverse), there is an increasing interest to automatically create content (visual, audio, text, etc) to provide media-rich and immersive experiences. This workshop invites top researchers and academics in this area to discuss the recent progress, share the thoughts, and envision the future [&hellip;]<\/p>\n","protected":false},"featured_media":874611,"template":"","meta":{"msr-url-field":"","msr-podcast-episode":"","msrModifiedDate":"","msrModifiedDateEnabled":false,"ep_exclude_from_search":false,"_classifai_error":"","msr_startdate":"2022-10-27","msr_enddate":"","msr_location":"Virtual","msr_expirationdate":"","msr_event_recording_link":"","msr_event_link":"https:\/\/www.microsoft.com\/en-us\/research\/?post_type=msr-event&p=876336&msr-tab=register&secret=X4hE9g","msr_event_link_redirect":false,"msr_event_time":"China Standard Time (GMT +8)","msr_hide_region":false,"msr_private_event":true,"msr_hide_image_in_river":0,"footnotes":""},"research-area":[13556],"msr-region":[],"msr-event-type":[197944],"msr-video-type":[],"msr-locale":[268875],"msr-program-audience":[],"msr-post-option":[],"msr-impact-theme":[],"class_list":["post-876336","msr-event","type-msr-event","status-publish","has-post-thumbnail","hentry","msr-research-area-artificial-intelligence","msr-event-type-hosted-by-microsoft","msr-locale-en_us"],"msr_about":"<!-- wp:msr\/event-details {\"title\":\"Creation for Rich World Workshop\",\"image\":{\"id\":874611,\"url\":\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/WebsiteHero_1400x788_B.jpg\",\"alt\":\"Abstract image with blue, purple, and orange tiles moving upward\"}} \/-->\n\n<!-- wp:msr\/content-tabs -->\n<!-- wp:msr\/content-tab -->\n<!-- wp:paragraph -->\n<p>As our world goes more and more digitalized (especially in the era of metaverse), there is an increasing interest to automatically create content (visual, audio, text, etc) to provide media-rich and immersive experiences.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p>This workshop invites top researchers and academics in this area to discuss the recent progress, share the thoughts, and envision the future of content creation for rich world. Discussion topics include:&nbsp;&nbsp;<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:list -->\n<ul><li>Speech, music, sound, and spatial audio creation<\/li><li>Image, video, scene, and environment creation<\/li><li>Digital human creation<\/li><\/ul>\n<!-- \/wp:list -->\n\n<!-- wp:paragraph -->\n<p>The goal of this workshop is to discuss, communicate, share, and learn with each other on the topics of content creation, which can help better understand this area and come up with future research opportunities.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:heading -->\n<h2>Speakers<\/h2>\n<!-- \/wp:heading -->\n\n<!-- wp:columns -->\n<div class=\"wp-block-columns\"><!-- wp:column -->\n<div class=\"wp-block-column\"><!-- wp:image {\"align\":\"center\",\"id\":868479,\"sizeSlug\":\"thumbnail\",\"linkDestination\":\"none\",\"className\":\"is-style-rounded\"} -->\n<figure class=\"wp-block-image aligncenter size-thumbnail is-style-rounded\"><img src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/08\/tadas-baltrusaitis-150x150.jpg\" alt=\"Tadas Baltrusaitis\" class=\"wp-image-868479\"\/><\/figure>\n<!-- \/wp:image -->\n\n<!-- wp:spacer {\"height\":\"7px\"} -->\n<div style=\"height:7px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n<!-- \/wp:spacer -->\n\n<!-- wp:heading {\"textAlign\":\"center\",\"level\":5} -->\n<h5 class=\"has-text-align-center\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/tabaltru\/\" target=\"_blank\" rel=\"noreferrer noopener\">Tadas Baltrusaitis<\/a><\/h5>\n<!-- \/wp:heading -->\n\n<!-- wp:paragraph {\"align\":\"center\"} -->\n<p class=\"has-text-align-center\">Principal Scientist<br><em>Microsoft Mixed Reality and AI Lab<\/em><\/p>\n<!-- \/wp:paragraph --><\/div>\n<!-- \/wp:column -->\n\n<!-- wp:column -->\n<div class=\"wp-block-column\"><!-- wp:image {\"align\":\"center\",\"id\":881673,\"sizeSlug\":\"thumbnail\",\"linkDestination\":\"none\",\"className\":\"is-style-rounded\"} -->\n<figure class=\"wp-block-image aligncenter size-thumbnail is-style-rounded\"><img src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Jesse-Engel_360x360-150x150.jpg\" alt=\"Jesse Engel - Research Summit speaker\" class=\"wp-image-881673\"\/><\/figure>\n<!-- \/wp:image -->\n\n<!-- wp:spacer {\"height\":\"7px\"} -->\n<div style=\"height:7px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n<!-- \/wp:spacer -->\n\n<!-- wp:heading {\"textAlign\":\"center\",\"level\":5} -->\n<h5 class=\"has-text-align-center\"><a href=\"https:\/\/research.google\/people\/JesseEngel\/\" target=\"_blank\" rel=\"noreferrer noopener\">Jesse Engel<\/a><\/h5>\n<!-- \/wp:heading -->\n\n<!-- wp:paragraph {\"align\":\"center\"} -->\n<p class=\"has-text-align-center\">Staff Research Scientist<br><em>Google Research<\/em><\/p>\n<!-- \/wp:paragraph --><\/div>\n<!-- \/wp:column -->\n\n<!-- wp:column -->\n<div class=\"wp-block-column\"><!-- wp:image {\"align\":\"center\",\"id\":881676,\"sizeSlug\":\"thumbnail\",\"linkDestination\":\"none\",\"className\":\"is-style-rounded\"} -->\n<figure class=\"wp-block-image aligncenter size-thumbnail is-style-rounded\"><img src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Jia-Tao_360x360-150x150.jpg\" alt=\"Jiatao Gu - Research Summit speaker\" class=\"wp-image-881676\"\/><\/figure>\n<!-- \/wp:image -->\n\n<!-- wp:spacer {\"height\":\"7px\"} -->\n<div style=\"height:7px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n<!-- \/wp:spacer -->\n\n<!-- wp:heading {\"textAlign\":\"center\",\"level\":5} -->\n<h5 class=\"has-text-align-center\"><a href=\"https:\/\/jiataogu.me\/\" target=\"_blank\" rel=\"noreferrer noopener\">Jiatao Gu<\/a><\/h5>\n<!-- \/wp:heading -->\n\n<!-- wp:paragraph {\"align\":\"center\"} -->\n<p class=\"has-text-align-center\">Researcher<br><em>Apple<\/em><\/p>\n<!-- \/wp:paragraph --><\/div>\n<!-- \/wp:column -->\n\n<!-- wp:column -->\n<div class=\"wp-block-column\"><!-- wp:image {\"align\":\"center\",\"id\":881682,\"sizeSlug\":\"thumbnail\",\"linkDestination\":\"none\",\"className\":\"is-style-rounded\"} -->\n<figure class=\"wp-block-image aligncenter size-thumbnail is-style-rounded\"><img src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Xiaoguang-Han_360x360-150x150.jpg\" alt=\"Xiaoguang Han - Research Summit speaker\" class=\"wp-image-881682\"\/><\/figure>\n<!-- \/wp:image -->\n\n<!-- wp:spacer {\"height\":\"7px\"} -->\n<div style=\"height:7px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n<!-- \/wp:spacer -->\n\n<!-- wp:heading {\"textAlign\":\"center\",\"level\":5} -->\n<h5 class=\"has-text-align-center\"><a href=\"https:\/\/gaplab.cuhk.edu.cn\/\" target=\"_blank\" rel=\"noreferrer noopener\">Xiaoguang Han<\/a><\/h5>\n<!-- \/wp:heading -->\n\n<!-- wp:paragraph {\"align\":\"center\"} -->\n<p class=\"has-text-align-center\">Assistant Professor<br><em>The Chinese University of Hong Kong, Shenzhen<\/em><\/p>\n<!-- \/wp:paragraph --><\/div>\n<!-- \/wp:column -->\n\n<!-- wp:column -->\n<div class=\"wp-block-column\"><!-- wp:image {\"align\":\"center\",\"id\":881679,\"sizeSlug\":\"thumbnail\",\"linkDestination\":\"none\",\"className\":\"is-style-rounded\"} -->\n<figure class=\"wp-block-image aligncenter size-thumbnail is-style-rounded\"><img src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Seungyong-Lee_360x360-150x150.jpg\" alt=\"Seungyong Lee - Research Summit speaker\" class=\"wp-image-881679\"\/><\/figure>\n<!-- \/wp:image -->\n\n<!-- wp:spacer {\"height\":\"7px\"} -->\n<div style=\"height:7px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n<!-- \/wp:spacer -->\n\n<!-- wp:heading {\"textAlign\":\"center\",\"level\":5} -->\n<h5 class=\"has-text-align-center\"><a href=\"http:\/\/cg.postech.ac.kr\/leesy\/\" target=\"_blank\" rel=\"noreferrer noopener\">Seungyong Lee<\/a><\/h5>\n<!-- \/wp:heading -->\n\n<!-- wp:paragraph {\"align\":\"center\"} -->\n<p class=\"has-text-align-center\">Professor<br><em>POSTECH<\/em><\/p>\n<!-- \/wp:paragraph --><\/div>\n<!-- \/wp:column -->\n\n<!-- wp:column -->\n<div class=\"wp-block-column\"><!-- wp:image {\"align\":\"center\",\"id\":881649,\"sizeSlug\":\"thumbnail\",\"linkDestination\":\"none\",\"className\":\"is-style-rounded\"} -->\n<figure class=\"wp-block-image aligncenter size-thumbnail is-style-rounded\"><img src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Yang-Liu_360x360-150x150.jpg\" alt=\"Yang Liu - Research Summit speaker\" class=\"wp-image-881649\"\/><\/figure>\n<!-- \/wp:image -->\n\n<!-- wp:spacer {\"height\":\"7px\"} -->\n<div style=\"height:7px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n<!-- \/wp:spacer -->\n\n<!-- wp:heading {\"textAlign\":\"center\",\"level\":5} -->\n<h5 class=\"has-text-align-center\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/yangliu\/\" target=\"_blank\" rel=\"noreferrer noopener\">Yang Liu<\/a><\/h5>\n<!-- \/wp:heading -->\n\n<!-- wp:paragraph {\"align\":\"center\"} -->\n<p class=\"has-text-align-center\">Principal Researcher<br><em>Microsoft Research Asia<\/em><\/p>\n<!-- \/wp:paragraph --><\/div>\n<!-- \/wp:column --><\/div>\n<!-- \/wp:columns -->\n\n<!-- wp:columns -->\n<div class=\"wp-block-columns\"><!-- wp:column -->\n<div class=\"wp-block-column\"><!-- wp:image {\"align\":\"center\",\"id\":881700,\"sizeSlug\":\"thumbnail\",\"linkDestination\":\"none\",\"className\":\"is-style-rounded\"} -->\n<figure class=\"wp-block-image aligncenter size-thumbnail is-style-rounded\"><img src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Wei-Ping_square-150x150.jpg\" alt=\"Wei Ping - Research Summit speaker\" class=\"wp-image-881700\"\/><\/figure>\n<!-- \/wp:image -->\n\n<!-- wp:spacer {\"height\":\"7px\"} -->\n<div style=\"height:7px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n<!-- \/wp:spacer -->\n\n<!-- wp:heading {\"textAlign\":\"center\",\"level\":5} -->\n<h5 class=\"has-text-align-center\"><a href=\"https:\/\/wpingnet.github.io\/\" target=\"_blank\" rel=\"noreferrer noopener\">Wei Ping<\/a><\/h5>\n<!-- \/wp:heading -->\n\n<!-- wp:paragraph {\"align\":\"center\"} -->\n<p class=\"has-text-align-center\">Principal Research Scientist<br><em>NVIDIA<\/em><\/p>\n<!-- \/wp:paragraph --><\/div>\n<!-- \/wp:column -->\n\n<!-- wp:column -->\n<div class=\"wp-block-column\"><!-- wp:image {\"align\":\"center\",\"id\":881667,\"sizeSlug\":\"thumbnail\",\"linkDestination\":\"none\",\"className\":\"is-style-rounded\"} -->\n<figure class=\"wp-block-image aligncenter size-thumbnail is-style-rounded\"><img src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Alexander-Richard_360x360-150x150.jpg\" alt=\"Alexander Richard - Research Summit speaker\" class=\"wp-image-881667\"\/><\/figure>\n<!-- \/wp:image -->\n\n<!-- wp:spacer {\"height\":\"7px\"} -->\n<div style=\"height:7px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n<!-- \/wp:spacer -->\n\n<!-- wp:heading {\"textAlign\":\"center\",\"level\":5} -->\n<h5 class=\"has-text-align-center\"><a href=\"https:\/\/alexanderrichard.github.io\/\" target=\"_blank\" rel=\"noreferrer noopener\">Alexander Richard<\/a><\/h5>\n<!-- \/wp:heading -->\n\n<!-- wp:paragraph {\"align\":\"center\"} -->\n<p class=\"has-text-align-center\">Research Scientist<br><em>Meta Reality Labs Research<\/em><\/p>\n<!-- \/wp:paragraph --><\/div>\n<!-- \/wp:column -->\n\n<!-- wp:column -->\n<div class=\"wp-block-column\"><!-- wp:image {\"align\":\"center\",\"id\":881688,\"sizeSlug\":\"thumbnail\",\"linkDestination\":\"none\",\"className\":\"is-style-rounded\"} -->\n<figure class=\"wp-block-image aligncenter size-thumbnail is-style-rounded\"><img src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Xu-Tan_360x360-150x150.jpg\" alt=\"Xu Tan - Research Summit speaker\" class=\"wp-image-881688\"\/><\/figure>\n<!-- \/wp:image -->\n\n<!-- wp:spacer {\"height\":\"7px\"} -->\n<div style=\"height:7px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n<!-- \/wp:spacer -->\n\n<!-- wp:heading {\"textAlign\":\"center\",\"level\":5} -->\n<h5 class=\"has-text-align-center\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/xuta\/\" target=\"_blank\" rel=\"noreferrer noopener\">Xu Tan<\/a><\/h5>\n<!-- \/wp:heading -->\n\n<!-- wp:paragraph {\"align\":\"center\"} -->\n<p class=\"has-text-align-center\">Principal Research Manager<br><em>Microsoft Research Asia<\/em><\/p>\n<!-- \/wp:paragraph --><\/div>\n<!-- \/wp:column -->\n\n<!-- wp:column -->\n<div class=\"wp-block-column\"><!-- wp:image {\"align\":\"center\",\"id\":881685,\"sizeSlug\":\"thumbnail\",\"linkDestination\":\"none\",\"className\":\"is-style-rounded\"} -->\n<figure class=\"wp-block-image aligncenter size-thumbnail is-style-rounded\"><img src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Xin-Tong_square-150x150.jpg\" alt=\"Xin Tong - Research Summit speaker\" class=\"wp-image-881685\"\/><\/figure>\n<!-- \/wp:image -->\n\n<!-- wp:spacer {\"height\":\"7px\"} -->\n<div style=\"height:7px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n<!-- \/wp:spacer -->\n\n<!-- wp:heading {\"textAlign\":\"center\",\"level\":5} -->\n<h5 class=\"has-text-align-center\"><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/xtong\/\" target=\"_blank\" rel=\"noreferrer noopener\">Xin Tong<\/a><\/h5>\n<!-- \/wp:heading -->\n\n<!-- wp:paragraph {\"align\":\"center\"} -->\n<p class=\"has-text-align-center\">Partner Research Manager<br><em>Microsoft Research Asia<\/em><\/p>\n<!-- \/wp:paragraph --><\/div>\n<!-- \/wp:column -->\n\n<!-- wp:column -->\n<div class=\"wp-block-column\"><!-- wp:image {\"align\":\"center\",\"id\":881670,\"sizeSlug\":\"thumbnail\",\"linkDestination\":\"none\",\"className\":\"is-style-rounded\"} -->\n<figure class=\"wp-block-image aligncenter size-thumbnail is-style-rounded\"><img src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/Baoyuan-Wang_360x360-150x150.jpg\" alt=\"Baoyuan Wang - Research Summit speaker\" class=\"wp-image-881670\"\/><\/figure>\n<!-- \/wp:image -->\n\n<!-- wp:spacer {\"height\":\"7px\"} -->\n<div style=\"height:7px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n<!-- \/wp:spacer -->\n\n<!-- wp:heading {\"textAlign\":\"center\",\"level\":5} -->\n<h5 class=\"has-text-align-center\"><a href=\"https:\/\/www.linkedin.com\/in\/baoyuan-wang-59153424\/\" target=\"_blank\" rel=\"noreferrer noopener\">Baoyuan Wang<\/a><\/h5>\n<!-- \/wp:heading -->\n\n<!-- wp:paragraph {\"align\":\"center\"} -->\n<p class=\"has-text-align-center\">Cofounder &amp; VP<br><em>Xiaobing.ai<\/em><\/p>\n<!-- \/wp:paragraph --><\/div>\n<!-- \/wp:column -->\n\n<!-- wp:column -->\n<div class=\"wp-block-column\"><\/div>\n<!-- \/wp:column --><\/div>\n<!-- \/wp:columns -->\n\n<!-- wp:spacer {\"height\":\"15px\"} -->\n<div style=\"height:15px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n<!-- \/wp:spacer -->\n\n<!-- wp:heading -->\n<h2>Agenda<\/h2>\n<!-- \/wp:heading -->\n\n<!-- wp:table {\"className\":\"is-style-stripes\"} -->\n<figure class=\"wp-block-table is-style-stripes\"><table><thead><tr><th>Time (CST)<\/th><th>Session<\/th><th>Speaker<\/th><\/tr><\/thead><tbody><tr><td>09:30 AM<\/td><td>Welcome &amp; Overview of Workshop<\/td><td><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/xuta\/\">Xu Tan<\/a><br>Principal Research Manager<br><em>Microsoft Research Asia<\/em><\/td><\/tr><tr><td><\/td><td><strong>Session 1: Speech\/Music\/Sound and Spatial Audio Creation<\/strong><\/td><td><\/td><\/tr><tr><td>09:45 AM<\/td><td>Explore the Limit of Zero-shot Audio Synthesis with Large-scale GAN Training<br><br>Abstract: In this talk, I will present some state-of-the-art results for raw audio synthesis. We will compare different family of methods for the universal vocoding task and introduce BigVGAN that can generalize well under various unseen conditions in zero-shot setting.<\/td><td><a href=\"https:\/\/wpingnet.github.io\/\" target=\"_blank\" rel=\"noreferrer noopener\">Wei Ping<\/a><br>Principal Research Scientist<br><em>NVIDIA<\/em><\/td><\/tr><tr><td>10:05 AM<\/td><td>Magenta: Empowering Creativity in the Age of Machine Learning<br><br>Abstract: Since 2016, the Magenta research group has investigated the role of machine learning in empowering the creativity of artists, musicians, and novices alike. In this talk, we'll examine recent advances by the group the role of interpretable decompositions for music understanding and generation, including state-of-the-art models in music transcription (MT3), Composition (Perceiver AR), synthesis (Spectrogram Diffusion), and user interaction (DDSP-VST, MIDI-DDSP). Finally, we'll explore how the combination of expressive generative models and intuitive controls can power a new generation of creative tools.<\/td><td><a href=\"https:\/\/research.google\/people\/JesseEngel\/\" target=\"_blank\" rel=\"noreferrer noopener\">Jesse Engel<\/a><br>Staff Research Scientist<br><em>Google Research<\/em><\/td><\/tr><tr><td>10:25 AM<\/td><td>Neural Audio Rendering for Social Telepresence<br><br>Abstract: These days, physical distance between people is one of the biggest obstacles to maintaining meaningful social relationships with family, friends, and co-workers. Even with today\u2019s technology, remote communication is limited to a two-dimensional audio-visual experience and lacks the availability of a shared, three-dimensional space in which people can interact with each other over the distance. Our mission at Reality Labs Research (RLR) in Pittsburgh is to develop a telepresence system that is indistinguishable from reality, i.e., a system that provides photo- and phono-realistic social interactions in VR. Highly realistic spatial audio rendering is a key ingredient to achieve the desired level of realism. While computer graphics has long moved from traditional rendering to neural rendering, the audio community is just in the early stages of this process. I will discuss the advantages of neural sound rendering and outline the challenges in data collection for those typically data-hungry machine learning approaches. I will further demonstrate that realism and accuracy of neural spatial audio methods exceed traditional signal processing. In the future, these technologies will help build a realistic virtual environment with lifelike avatars that allow for authentic social interactions, connecting people all over the world, anywhere and at any time.<\/td><td><a href=\"https:\/\/alexanderrichard.github.io\/\" target=\"_blank\" rel=\"noreferrer noopener\">Alexander Richard<\/a><br>Research Scientist<br><em>Meta Reality Labs Research<\/em><\/td><\/tr><tr><td>10:45 AM<\/td><td>Panel Discussions<\/td><td><strong>Host<\/strong>: <a href=\"https:\/\/research.google\/people\/JesseEngel\/\" target=\"_blank\" rel=\"noreferrer noopener\">Jesse Engel<\/a><br><strong>Panelists<\/strong>:<br><a href=\"https:\/\/wpingnet.github.io\/\" target=\"_blank\" rel=\"noreferrer noopener\">Wei Ping<\/a>, <a href=\"https:\/\/alexanderrichard.github.io\/\" target=\"_blank\" rel=\"noreferrer noopener\">Alexander Richard<\/a>, <br><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/xuta\/\">Xu Tan<\/a><\/td><\/tr><tr><td>11:30 AM<\/td><td><em>Lunch break<\/em><\/td><td><\/td><\/tr><tr><td><\/td><td><strong>Session 2: Image\/Video\/Scene Creation<\/strong><\/td><td><\/td><\/tr><tr><td>12:30 PM<\/td><td>Semantic Instance Reconstruction for 3D Scene Understanding<br><br>Abstract: 3D scene understanding and reconstruction plays very important roles in many application scenarios, like robot perception and also AR\/VR etc. Currently, most of existing works treated the 3D scene as a whole stuff for reconstruction. In this talk, I will introduce our recent techniques that conducted reconstruction together with instance understanding, which is termed as \u201cSemantic Instance Reconstruction\u201d. The main content includes three published works: Total3DUnderstanding (CVPR 2020), RfD-Net (CVPR 2021) and InstPIFu (ECCV 2022). They are all aiming to semantic instance reconstruction, but Total3D focuses on reconstruction from single images while RfD-Net takes point cloud as input, InstPIFu targets high-fidelity single-view Reconstruction. Furthermore, I will also introduce our recent work on 3D table-scene dataset.<\/td><td><a href=\"https:\/\/gaplab.cuhk.edu.cn\/\" target=\"_blank\" rel=\"noreferrer noopener\">Xiaoguang Han<\/a><br>Assistant Professor<br><em>The Chinese University of Hong Kong, Shenzhen<\/em><\/td><\/tr><tr><td>12:50 PM<\/td><td>Towards High-fidelity 3D Shape and Scene Generation<br><br>Abstract: Digital 3D contents equipped with high-fidelity shape geometry, scene layout, and visual appearance are the basis for building the digital world and facilitating 3D understanding, interaction, and exploration. The main barriers to generating vibrant 3D content include insufficient number of labeled 3D data, diversity of 3D representations, and lack of editability. In this talk, we will present a set of our 3D generation work that incorporates neural-based 3D representations and generative models to overcome these barriers and create high-fidelity 3D shapes and scenes.<\/td><td><a href=\"https:\/\/xueyuhanlang.github.io\/\" target=\"_blank\" rel=\"noreferrer noopener\">Yang Liu<\/a><br>Principal Researcher<br><em>Microsoft Research Asia<\/em><\/td><\/tr><tr><td>1:10 PM<\/td><td>Diffusion Models for Image and Neural Field Generation<br><br>Abstract: Diffusion probabilistic models have quickly become the de-facto choice for generative modeling of images, text or 3D geometry. In this talk, we will introduce some of our recent works on how to generalize diffusion models for more efficient and controllable image generation, as well as general field data.<\/td><td><a href=\"https:\/\/jiataogu.me\/\" target=\"_blank\" rel=\"noreferrer noopener\">Jiatao Gu<\/a><br>Researcher<br><em>Apple<\/em><\/td><\/tr><tr><td>1:30 PM<\/td><td>Panel Discussions<\/td><td><strong>Host<\/strong>: <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/xtong\/\">Xin Tong<\/a><br><strong>Panelists<\/strong>: <br><a href=\"https:\/\/gaplab.cuhk.edu.cn\/\" target=\"_blank\" rel=\"noreferrer noopener\">Xiaoguang Han<\/a>, <br><a href=\"https:\/\/xueyuhanlang.github.io\/\" target=\"_blank\" rel=\"noreferrer noopener\">Yang Liu<\/a>, <br><a href=\"https:\/\/jiataogu.me\/\" target=\"_blank\" rel=\"noreferrer noopener\">Jiatao Gu<\/a><\/td><\/tr><tr><td>2:15 PM<\/td><td><em>Short break<\/em><\/td><td><br><\/td><\/tr><tr><td><\/td><td><strong>Session 3: Digital Human Creation<\/strong><\/td><td><\/td><\/tr><tr><td>2:25 PM<\/td><td>From Human to AI Being Intelligence: Challenges and Opportunities<br><br>Abstract: Digital human or AI being (named by Xiaobing.ai) has already shown tremendous potential value for helping digital transformations in various industry domains. It is one of the interesting testbeds for building next generations of artificial intelligence. In this talk, from an industry perspective, we will discuss the fundamental technical challenges and share some learning experiences of developing different AI beings under Xiaoice avatar framework.<\/td><td><a href=\"https:\/\/www.linkedin.com\/in\/baoyuan-wang-59153424\/\" target=\"_blank\" rel=\"noreferrer noopener\">Baoyuan Wang<\/a><br>Cofounder &amp; VP<br><em>Xiaobing.ai<\/em><\/td><\/tr><tr><td>2:45 PM<\/td><td>MLPs for Reconstruction and Control of Explicit and Detailed 3D Human Models<br><br>Abstract: MLPs (Multi-Layer Perceptrons) have been widely used for reconstructing implicit 3D representations of human models. In this talk, in contrast, I will focus on utilizing MLPs for reconstructing and controlling explicit polygonal representations of 3D human models. I will first present an MLP-based framework for building a deformable surface model, which takes a latent code and produces a 3D caricature model. The framework captures the variations of 3D caricatures in a compact parameter space and provides a useful data-driven toolkit for handling 3D caricature deformations. I will then present LaplacianFusion, a novel framework that reconstructs a detailed and controllable 3D clothed human model from a point cloud sequence. In the framework, an MLP is used to learn and predict Laplacian coordinates representing the details on the body surface. The talk will be concluded with discussion on the pros and cons of implicit and explicit representations when they are combined with MLPs for reconstruction and control of 3D human models.<\/td><td><a href=\"http:\/\/cg.postech.ac.kr\/leesy\/\" target=\"_blank\" rel=\"noreferrer noopener\">Seungyong Lee<\/a><br>Professor<br><em>POSTECH<\/em><\/td><\/tr><tr><td>3:05 PM<\/td><td>Synthetic Data is All You Need: Face Analysis Using Synthetic Data Alone<br><br>Abstract: In this talk I will demonstrate how synthetic data alone can be used to perform face-related computer vision. The community has long enjoyed the benefits of synthesizing training data with graphics, but the domain gap between real and synthetic data has remained a problem, especially for human faces. I will show that it is possible to synthesize data with minimal domain gap, so that models trained on synthetic data generalize to real in-the-wild datasets. I will describe how to combine a procedurally-generated parametric 3D face model with a comprehensive library of hand-crafted assets to render training images with unprecedented realism and diversity. I will show that models trained using synthetic data alone can both match and exceed real data in accuracy as well as open up new approaches where manual labelling would be impossible.<\/td><td><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/tabaltru\/\">Tadas Baltrusaitis<\/a><br>Principal Scientist<br><em>Microsoft Mixed Reality and AI <\/em>l<em>ab<\/em><\/td><\/tr><tr><td>3:25 PM<\/td><td>Panel Discussions<\/td><td><strong>Host<\/strong>: <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/xuta\/\">Xu Tan<\/a><br><strong>Panelists<\/strong>:<br><a href=\"https:\/\/www.linkedin.com\/in\/baoyuan-wang-59153424\/\" target=\"_blank\" rel=\"noreferrer noopener\">Baoyuan Wang<\/a>, <br><a href=\"http:\/\/cg.postech.ac.kr\/leesy\/\" target=\"_blank\" rel=\"noreferrer noopener\">Seungyong Lee<\/a>, <br><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/tabaltru\/\">Tadas Baltrusaitis<\/a>, <br><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/xtong\/\">Xin Tong<\/a><\/td><\/tr><tr><td>4:10 PM<\/td><td>Next Steps\/Closing Remarks<\/td><td><\/td><\/tr><\/tbody><\/table><\/figure>\n<!-- \/wp:table -->\n\n<!-- wp:spacer {\"height\":\"40px\"} -->\n<div style=\"height:40px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n<!-- \/wp:spacer -->\n\n<!-- wp:heading -->\n<h2>Workshop organizers<\/h2>\n<!-- \/wp:heading -->\n\n<!-- wp:paragraph -->\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/xuta\/\" target=\"_blank\" rel=\"noreferrer noopener\">Xu Tan<\/a>, Microsoft Research Asia<br><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/xtong\/\" target=\"_blank\" rel=\"noreferrer noopener\">Xin Tong<\/a>, Microsoft Research Asia<br><a href=\"https:\/\/www.linkedin.com\/in\/miran-lee-37996b36\/\" target=\"_blank\" rel=\"noreferrer noopener\">Miran Lee<\/a>, Microsoft Research Asia<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:spacer {\"height\":\"30px\"} -->\n<div style=\"height:30px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n<!-- \/wp:spacer -->\n\n<!-- wp:heading {\"level\":4} -->\n<h4>Microsoft\u2019s Event Code of Conduct<\/h4>\n<!-- \/wp:heading -->\n\n<!-- wp:paragraph -->\n<p>Microsoft\u2019s mission is to empower every person and every organization on the planet to achieve more. This includes events Microsoft hosts and participates in, where we seek to create a respectful, friendly, and inclusive experience for all participants. As such, we do not tolerate harassing or disrespectful behavior, messages, images, or interactions by any event participant, in any form, at any aspect of the program including business and social activities, regardless of location. <\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:paragraph -->\n<p>We do not tolerate any behavior that is degrading to any gender, race, sexual orientation or disability, or any behavior that would violate <a href=\"https:\/\/www.microsoft.com\/en-us\/legal\/compliance\/default.aspx\" target=\"_blank\" rel=\"noreferrer noopener\">Microsoft\u2019s Anti-Harassment and Anti-Discrimination Policy, Equal Employment Opportunity Policy, or&nbsp;Standards of Business Conduct<\/a>. In short, the entire experience at the venue must meet our culture standards. We encourage everyone to assist in creating a welcoming and safe environment. Please <a href=\"https:\/\/aka.ms\/reportconcern\" target=\"_blank\" rel=\"noreferrer noopener\">report<\/a> any concerns, harassing behavior, or suspicious or disruptive activity to venue staff, the event host or owner, or event staff. Microsoft reserves the right to refuse admittance to or remove any person from company-sponsored events at any time in its sole discretion.<\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:buttons -->\n<div class=\"wp-block-buttons\"><!-- wp:button {\"className\":\"is-style-outline\"} -->\n<div class=\"wp-block-button is-style-outline\"><a class=\"wp-block-button__link\" href=\"https:\/\/aka.ms\/reportconcern\" target=\"_blank\" rel=\"noreferrer noopener\">Report a concern<\/a><\/div>\n<!-- \/wp:button --><\/div>\n<!-- \/wp:buttons -->\n<!-- \/wp:msr\/content-tab -->\n\n<!-- wp:msr\/content-tab {\"title\":\"Register\"} -->\n<!-- wp:paragraph {\"placeholder\":\"Write content\u2026\"} -->\n<p><\/p>\n<!-- \/wp:paragraph -->\n\n<!-- wp:html -->\n<iframe width=\"1000px\" height=\"4600px\" src=\"https:\/\/forms.office.com\/r\/LBinzjZBg3;embed=true\" frameborder=\"0\" marginwidth=\"0\" marginheight=\"0\" style=\"border: none; max-width:100%; max-height:100vh\" allowfullscreen=\"\" webkitallowfullscreen=\"\" mozallowfullscreen=\"\" msallowfullscreen=\"\"> <\/iframe>\n<!-- \/wp:html -->\n<!-- \/wp:msr\/content-tab -->\n<!-- \/wp:msr\/content-tabs -->","tab-content":[],"msr_startdate":"2022-10-27","msr_enddate":"","msr_event_time":"China Standard Time (GMT +8)","msr_location":"Virtual","msr_event_link":"https:\/\/www.microsoft.com\/en-us\/research\/?post_type=msr-event&p=876336&msr-tab=register&secret=X4hE9g","msr_event_recording_link":"","msr_startdate_formatted":"October 27, 2022","msr_register_text":"Watch now","msr_cta_link":"https:\/\/www.microsoft.com\/en-us\/research\/?post_type=msr-event&p=876336&msr-tab=register&secret=X4hE9g","msr_cta_text":"Watch now","msr_cta_bi_name":"Event Register","featured_image_thumbnail":"<img width=\"960\" height=\"540\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/WebsiteHero_1400x788_B-960x540.jpg\" class=\"img-object-cover\" alt=\"Abstract image with blue, purple, and orange tiles moving upward\" decoding=\"async\" loading=\"lazy\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/WebsiteHero_1400x788_B-960x540.jpg 960w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/WebsiteHero_1400x788_B-300x169.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/WebsiteHero_1400x788_B-1024x576.jpg 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/WebsiteHero_1400x788_B-768x432.jpg 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/WebsiteHero_1400x788_B-1066x600.jpg 1066w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/WebsiteHero_1400x788_B-655x368.jpg 655w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/WebsiteHero_1400x788_B-343x193.jpg 343w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/WebsiteHero_1400x788_B-240x135.jpg 240w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/WebsiteHero_1400x788_B-640x360.jpg 640w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/WebsiteHero_1400x788_B-1280x720.jpg 1280w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2022\/09\/WebsiteHero_1400x788_B.jpg 1400w\" sizes=\"auto, (max-width: 960px) 100vw, 960px\" \/>","event_excerpt":"As our world goes more and more digitalized (especially in the era of metaverse), there is an increasing interest to automatically create content (visual, audio, text, etc) to provide media-rich and immersive experiences. This workshop invites top researchers and academics in this area to discuss the recent progress, share the thoughts, and envision the future of content creation for rich world. Discussion topics include:&nbsp;&nbsp; Speech, music, sound, and spatial audio creationImage, video, scene, and environment&hellip;","msr_research_lab":[],"related-researchers":[],"msr_impact_theme":[],"related-academic-programs":[],"related-groups":[],"related-projects":[],"related-opportunities":[],"related-publications":[],"related-videos":[],"related-posts":[],"_links":{"self":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event\/876336","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event"}],"about":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/types\/msr-event"}],"version-history":[{"count":37,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event\/876336\/revisions"}],"predecessor-version":[{"id":922170,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event\/876336\/revisions\/922170"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media\/874611"}],"wp:attachment":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media?parent=876336"}],"wp:term":[{"taxonomy":"msr-research-area","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/research-area?post=876336"},{"taxonomy":"msr-region","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-region?post=876336"},{"taxonomy":"msr-event-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event-type?post=876336"},{"taxonomy":"msr-video-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-video-type?post=876336"},{"taxonomy":"msr-locale","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-locale?post=876336"},{"taxonomy":"msr-program-audience","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-program-audience?post=876336"},{"taxonomy":"msr-post-option","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-post-option?post=876336"},{"taxonomy":"msr-impact-theme","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-impact-theme?post=876336"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}