{"id":144711,"date":"2020-02-25T11:36:42","date_gmt":"2000-03-27T00:12:25","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/group\/internet-media\/"},"modified":"2025-07-01T02:32:19","modified_gmt":"2025-07-01T09:32:19","slug":"internet-media","status":"publish","type":"msr-group","link":"https:\/\/www.microsoft.com\/en-us\/research\/group\/internet-media\/","title":{"rendered":"Intelligent Multimedia Group"},"content":{"rendered":"<div>\n<div class=\"asset-content\" style=\"text-align: left;\">\n<p>The Intelligent Multimedia (IM) group aims to build seamless yet efficient multimedia applications and services through breakthroughs in fundamental theory and innovations in algorithm and system technology. We address the problems of intelligent multimedia content sensing, processing, analysis, services, and the generic scalability issues of multimedia computing systems. Current research focus is on video analytics to support intelligent cloud and intelligent edge media services.\u00a0Current research interests include, but are not limited to, object detection, tracking, semantic segmentation, human pose estimation, people re-ID, action recognition, depth estimation, SLAM, scene understanding, multimodality analysis, etc.<\/p>\n<\/div>\n<\/div>\n<div class=\"asset-content\" style=\"text-align: left;\"><\/div>\n<h2 class=\"asset-content\" style=\"text-align: left;\"><strong style=\"font-size: 1rem;\">Areas of Focus:<\/strong><\/h2>\n<div class=\"asset-content\" style=\"text-align: left;\">\n<div><strong>Deep Video Analytics <\/strong><\/div>\n<div>\n<p>Video is the biggest big data that contains an enormous amount of information. We are leveraging computer vision and deep learning to develop both cloud-based and edge-based intelligence engines that can turn raw video data into insights to facilitate various applications and services. Target application scenarios include video augmented reality, smart home surveillance,\u00a0business (retail store, office) intelligence, public security, video storytelling and sharing, etc. We have taken a human centric approach where\u00a0a significant effort has been focused on understanding human, human attributes and human behaviors. Our research has\u00a0contributed to\u00a0a number of video APIs offered in Microsoft Cognitive Services (<a href=\"https:\/\/www.microsoft.com\/cognitive-services\">https:\/\/www.microsoft.com\/cognitive-services<\/a>), Azure Media Analytics Services, Windows Machine Learning, Office Media (Stream\/Teams), and Dynamics\/Connected Store.<\/p>\n<p>&#8211; Video API R&D, 3 technologies (intelligent motion detection, face detection\/tracking, face redaction), deployed in Microsoft Cognitive Services and Azure Media Services (2016)<br \/>\n\uf0a7 <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/azure.microsoft.com\/en-us\/blog\/motion-detection\/\"> Announcing: Motion detection for Azure Media Analytics <span class=\"sr-only\"> (opens in new tab)<\/span><\/a> (2016)<br \/>\n\uf0a7 <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/azure.microsoft.com\/en-us\/blog\/face-and-emotion-detection\/\"> Announcing face and emotion detection for Azure Media Analytics | Azure Blog and Updates | Microsoft Azure <span class=\"sr-only\"> (opens in new tab)<\/span><\/a> (2016)<br \/>\n\uf0a7 <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/azure.microsoft.com\/en-us\/blog\/azure-media-redactor\/?cdn=disable\"> Announcing Face Redaction for Azure Media Analytics | Azure Blog and Updates | Microsoft Azure <span class=\"sr-only\"> (opens in new tab)<\/span><\/a> (2016)<br \/>\n\uf0a7 <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/docs.microsoft.com\/en-us\/azure\/media-services\/previous\/media-services-face-redaction\"> Redact faces with Azure Media Analytics | Microsoft Docs <span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/p>\n<p>&#8211; Developed, released\/deployed human pose estimation (2019.5) and object tracking (2019.10) technologies as vision skills on the Windows Machine Learning platform.<br \/>\n\uf0a7 <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/mp.weixin.qq.com\/s\/u0K5jqEaPfuiWCGWihWKzQ\"> \u5fae\u8f6f\u53d1\u5e03Windows Vision Skills\u9884\u89c8\u7248\uff0c\u8f7b\u677e\u8c03\u7528\u8ba1\u7b97\u673a\u89c6\u89c9 <span class=\"sr-only\"> (opens in new tab)<\/span><\/a><br \/>\n\uf0a7 <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/www.nuget.org\/packages\/Microsoft.AI.Skills.Vision.ObjectTrackerPreview\/\"> NuGet Gallery | Microsoft.AI.Skills.Vision.ObjectTrackerPreview 0.0.0.3 <span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/p>\n<p>&#8211; Speech denoising technologies deployed in Microsoft Stream 1.0 (GA, 2020.6) and 2.0 (Internal Preview 2020.12)<br \/>\n\uf0a7 <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/mp.weixin.qq.com\/s\/Xc-OmRpDPgw2qIgswsCeSA\"> \u4ece\u5608\u6742\u89c6\u9891\u4e2d\u63d0\u53d6\u8d85\u6e05\u4eba\u58f0\uff0c\u8bed\u97f3\u589e\u5f3a\u6a21\u578bPHASEN\u5df2\u52a0\u5165\u5fae\u8f6f\u89c6\u9891\u670d\u52a1 <span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/p>\n<p>&#8211; Multi object tracking (FairMOT), Multiview 3D pose estimation (VoxelPose), person re-ID technologies shipped to the Microsoft Dynamics\/Connected Store Product. (2020, and ongoing)<br \/>\n\uf0a7 <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/mp.weixin.qq.com\/s\/bMBBYWk5-o5XEOSzI7DyDg\"> \u4eceFairMOT\u5230VoxelPose\uff0c\u63ed\u79d8\u5fae\u8f6f\u4ee5\u201c\u4eba\u201d\u4e3a\u4e2d\u5fc3\u7684\u6700\u65b0\u89c6\u89c9\u7406\u89e3\u6210\u679c <span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/p>\n<p>&#8211; Screen content understanding (element detection\/screen tree) technologies shipped to Microsoft\u2019s mobile robotic process automation (RPA) product (2020, and ongoing)<\/p>\n<\/div>\n<div><strong> Open Source Projects: <\/strong><\/div>\n<div>.\u00a0Human Pose Estimation:\u00a0<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/github.com\/microsoft\/voxelpose-pytorch\">VoxelPose <span class=\"sr-only\"> (opens in new tab)<\/span><\/a><br \/>\n<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/github.com\/microsoft\/multiview-human-pose-estimation-pytorch\"> Cross View Fusion for 3D Human Pose Estimation <span class=\"sr-only\"> (opens in new tab)<\/span><\/a><br \/>\n<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/github.com\/CHUNYUWANG\/imu-human-pose-pytorch\"> Fusing Wearable IMUs with Multi-View Images for Human Pose Estimation: A Geometric Approach<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/div>\n<div>.\u00a0Object Tracking:\u00a0<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/github.com\/microsoft\/SA-Siam\">SA-Siam<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><br \/>\n<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/github.com\/microsoft\/SPM-Tracker\"> SPM-Tracker<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/github.com\/microsoft\/PySiamTracking\"> Siamese network based tracker <span class=\"sr-only\"> (opens in new tab)<\/span><\/a> (a comprehensive PyTorch based toolbox that supports a series of Siamese-network-based tracking methods like SiamFC \/ SiamRPN \/ SPM)<br \/>\n<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/github.com\/ifzhang\/FairMOT\"> A Simple Baseline for One-Shot Multi-Object Tracking <span class=\"sr-only\"> (opens in new tab)<\/span><\/a> (2.2K stars)<\/div>\n<div>. Re-Identification: <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" style=\"font-size: 1rem;\" href=\"https:\/\/github.com\/microsoft\/Semantics-Aligned-Representation-Learning-for-Person-Re-identification\">Semantics-aligned representation learning for person re-identification (SAN)<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/div>\n<div>. Action Recognition: <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" style=\"font-size: 1rem;\" href=\"https:\/\/github.com\/microsoft\/View-Adaptive-Neural-Networks-for-Skeleton-based-Human-Action-Recognition\">View adaptive neural networks <span class=\"sr-only\"> (opens in new tab)<\/span><\/a><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" style=\"font-size: 1rem;\" href=\"https:\/\/github.com\/microsoft\/SGN\"> Semantics-Guided Neural Networks for Efficient Skeleton-Based Human Action Recognition<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/div>\n<div>. Domain Generalization\/Adaptation:\u00a0<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/github.com\/microsoft\/SNR\"> Style Normalization and Restitution for Domain Generalization and Adaptation<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/div>\n<div>.<\/div>\n<\/div>\n<div><\/div>\n<div class=\"asset-content\" style=\"text-align: left;\">\n<div><strong>Titanium (past project) <\/strong><\/div>\n<div><\/div>\n<div>\n<p>Project Titanium aims at bringing new computing experiences through enriched cloud-client computing. While data and programs can be provided as services from the cloud, the screen, referring to the entire collection of data involved in user interface, constitutes the missing third dimension. Titanium will address the problems of adaptive screen composition, representation, and processing, following the roadmap of Titanium Screen, Titanium Remote, Titanium Live, and Titanium Cloud. As \u201cTitanium\u201d suggests, it will provide a light-weight yet efficient solution towards ultimate computing experiences in the cloud plus service era.<\/p>\n<\/div>\n<div><\/div>\n<div><strong>Mira (past project) <\/strong><\/div>\n<div><\/div>\n<div>\n<p>Project Mira aims at enabling multimedia representation and processing towards perceptual quality rather than pixel-wise fidelity through a joint effort of signal processing, computer vision, and machine learning. In particular, it seeks to build systems not only incorporating this newly developed vision and learning technologies into compression but also inspiring new vision technologies by looking at the problem from the view of signal processing. By bridging vision and signal processing, this project is expected to offer a fresh frame of mind to multimedia representation and processing.<\/p>\n<\/div>\n<div><\/div>\n\t<div data-wp-context='{\"items\":[]}' data-wp-interactive=\"msr\/accordion\">\n\t\t\t\t\t<div class=\"clearfix\">\n\t\t\t\t<div\n\t\t\t\t\tclass=\"btn-group align-items-center mb-g float-sm-right\"\n\t\t\t\t\tdata-bi-aN=\"accordion-collapse-controls\"\n\t\t\t\t>\n\t\t\t\t\t<button\n\t\t\t\t\t\tclass=\"btn btn-link m-0\"\n\t\t\t\t\t\tdata-bi-cN=\"Expand all\"\n\t\t\t\t\t\tdata-wp-bind--aria-controls=\"state.ariaControls\"\n\t\t\t\t\t\tdata-wp-bind--aria-expanded=\"state.ariaExpanded\"\n\t\t\t\t\t\tdata-wp-bind--disabled=\"state.isAllExpanded\"\n\t\t\t\t\t\tdata-wp-class--inactive=\"state.isAllExpanded\"\n\t\t\t\t\t\tdata-wp-on--click=\"actions.onExpandAll\"\n\t\t\t\t\t\ttype=\"button\"\n\t\t\t\t\t>\n\t\t\t\t\t\tExpand all\t\t\t\t\t<\/button>\n\t\t\t\t\t<span aria-hidden=\"true\"> | <\/span>\n\t\t\t\t\t<button\n\t\t\t\t\t\tclass=\"btn btn-link m-0\"\n\t\t\t\t\t\tdata-bi-cN=\"Collapse all\"\n\t\t\t\t\t\tdata-wp-bind--aria-controls=\"state.ariaControls\"\n\t\t\t\t\t\tdata-wp-bind--aria-expanded=\"state.ariaExpanded\"\n\t\t\t\t\t\tdata-wp-bind--disabled=\"state.isAllCollapsed\"\n\t\t\t\t\t\tdata-wp-class--inactive=\"state.isAllCollapsed\"\n\t\t\t\t\t\tdata-wp-on--click=\"actions.onCollapseAll\"\n\t\t\t\t\t\ttype=\"button\"\n\t\t\t\t\t>\n\t\t\t\t\t\tCollapse all\t\t\t\t\t<\/button>\n\t\t\t\t<\/div>\n\t\t\t<\/div>\n\t\t\t\t<ul class=\"msr-accordion\">\n\t\t\t\t\t\t\t\t<li class=\"m-0\" data-wp-context='{\"id\":\"accordion-content-2\"}' data-wp-init=\"callbacks.init\">\n\t\t<div class=\"accordion-header\">\n\t\t\t<button\n\t\t\t\taria-controls=\"accordion-content-2\"\n\t\t\t\tclass=\"btn btn-collapse\"\n\t\t\t\tdata-wp-bind--aria-expanded=\"state.isExpanded\"\n\t\t\t\tdata-wp-on--click=\"actions.onClick\"\n\t\t\t\tid=\"accordion-button-1\"\n\t\t\t\ttype=\"button\"\n\t\t\t>\n\t\t\t\tLatest News\t\t\t<\/button>\n\t\t<\/div>\n\t\t<div\n\t\t\taria-labelledby=\"accordion-button-1\"\n\t\t\tclass=\"msr-accordion__content\"\n\t\t\tdata-wp-bind--inert=\"!state.isExpanded\"\n\t\t\tdata-wp-run=\"callbacks.run\"\n\t\t\tid=\"accordion-content-2\"\n\t\t>\n\t\t\t<div class=\"msr-accordion__body\">\n\t\t\t\t<div>(Aug. 2021) <strong>Congratulations<\/strong> to Yifu, Chunyu, Xin, Cuiling for the following accepted papers!:1. Y. Zhang, C. Wang, X. Wang, W. Zeng, and W. Liu, \u201cFairMOT: On the Fairness of Detection and Re-Identification in Multiple Object Tracking,\u201d to appear in International Journal of Computer Vision;2. Xin Jin, Cuiling Lan, Wenjun Zeng, and Zhibo Chen, &#8220;Style Normalization and Restitution for Domain Generalization and Adaptation,&#8221; to appear in IEEE Trans. on Multimedia.<\/div>\n<div>*<\/div>\n<div>(July 2021) <strong>Congratulations<\/strong> to Xin Jin, Cuiling, Rongchang, Chunyu, Yucheng, Guangting, Chong for the following papers accepted by ICCV 2021!:1. Re-energizing Domain Discriminator with Sample Relabeling for Adversarial Domain Adaptation (Xin Jin, Cuiling Lan, Wenjun Zeng, Zhibo Chen);2. An Empirical Study of the Collapsing Problem in Semi-Supervised 2D Human Pose Estimation (Rongchang Xie, Chunyu Wang, Wenjun Zeng, Yizhou Wang);3. Self-Supervised Visual Representations Learning by Contrastive Mask Prediction (Yucheng Zhao, Guangting Wang, Chong Luo, Wenjun Zeng, Zheng-Jun Zha).<\/div>\n<div>*<\/div>\n<div>(July 2021) <strong>Congratulations<\/strong> to Kecheng, Cuiling and Zhizheng for the following paper accepted by ACM Multimedia 2021 !: Kecheng Zheng, Cuiling Lan, Wenjun Zeng, Jiawei Liu, Zhizheng Zhang, and Zheng-Jun Zha, \u201cPose-Guided Feature Learning with Knowledge Distillation for Occluded Person Re-Identification,\u201d to appear in ACM Multimedia, 2021.<\/div>\n<div>*<\/div>\n<div>(June 2021) <strong>Congratulations<\/strong> to Zhizheng and Cuiling for the following papers accepted by IJCAI 2021 !: 1. Zhizheng Zhang, Cuiling Lan, Wenjun Zeng, Zhibo Chen, and Shih-Fu Chang, &#8220;Uncertainty-Aware Few-Shot Image Classification&#8221;; 2: Jindong Wang, Cuiling Lan, Chang Liu, Yidong Ouyang, and Tao Qin, &#8220;Generalizing to Unseen Domains: A Survey on Domain Generalization&#8221; (survey track)<\/div>\n<div>*<\/div>\n<div>(March 2021) <strong>Congratulations<\/strong> to authors of the following papers accepted by CVPR2021: 1. Guoqiang Wei, Cuiling Lan, Wenjun Zeng, Zhibo Chen, &#8220;MetaAlign: Coordinating Domain Alignment and Classification for Unsupervised Domain Adaptation&#8221;;2. Guangting Wang, Yizhou Zhou, Chong Luo, Wenxuan Xie, Wenjun Zeng, Zhiwei Xiong, \u201cUnsupervised Visual Representation Learning by Tracking Patches in Video\u201d; 3. Xiaotian Chen, Yuwang Wang, Xuejin Chen, Wenjun Zeng, &#8220;S2R-DepthNet: Learning a Generalizable Depth-specific Structural Representation.&#8221; (Oral Paper)<\/div>\n<div>*<\/div>\n<div>(Jan. 2021) <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/github.com\/microsoft\/SNR\"> Style Normalization and Restitution for Domain Generalization and Adaptation <span class=\"sr-only\"> (opens in new tab)<\/span><\/a> is open sourced, and is also on <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/arxiv.org\/pdf\/2101.00588.pdf\"> arxiv <span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/div>\n<div>*<\/div>\n<div>(Dec. 2020) <strong>Congratulations<\/strong> to Kecheng, Cuiling, Zhizheng, and Zheng for their following papers accepted by AAAI2021: 1. Kecheng Zheng, Cuiling Lan, Wenjun Zeng, Zhizheng Zhang, Zheng-Jun Zha, \u201cExploiting Sample Uncertainty for Domain Adaptive Person Re-Identification\u201d;2. Xiao Wang, Zheng Wang, Toshihiko Yamasaki, Wenjun Zeng, \u201cVery Important Person Localization in Unconstrained Conditions: A New Benchmark\u201d<\/div>\n<div>*<\/div>\n<div>(Dec. 2020) <strong>Congratulations<\/strong> to Zhe and Chunyu for their following paper accepted by Inter. Journal of Computer Vision: Zhe Zhang, Chunyu Wang, Weichao Qiu, Wenhu Qin, and Wenjun Zeng, \u201cAdaFuse: Adaptive Multiview Fusion for Accurate Human Pose Estimation in the Wild.\u201d<\/div>\n<div>*<\/div>\n<div>(July 2020) <strong>Congratulations<\/strong> to Hanyue, Chunyu, Xin, Cuiling for their following papers accepted by ECCV2020: 1. Hanyue Tu, Chunyu Wang, and Wenjun Zeng, \u201cEnd-to-End Estimation of Multi-Person 3D Poses from Multiple Cameras\u201d (Oral Paper) ;2. Xin Jin, Cuiling Lan, Wenjun Zeng, Zhibo Chen, \u201cGlobal Distance-distributions Separation for Unsupervised Person Re-identification\u201d<\/div>\n<div>*<\/div>\n<div>(April 2020) <strong>Congratulations<\/strong> to Chuanxin, Chong, Zhiyuan, Wenxuan, Yucheng for their following papers accepted by IJCAI: 1. Joint Time-Frequency and Time Domain Learning for Speech Enhancement (Chuanxin Tang, Chong Luo, Zhiyuan Zhao, Wenxuan Xie, Wenjun Zeng);2. Multi-Scale Group Transformer for Long Sequence Modeling in Speech Separation (Yucheng Zhao, Chong Luo, Zheng-Jun Zha, Wenjun Zeng)<\/div>\n<div>*<\/div>\n<div>(March 2020) <strong>Congratulations<\/strong> to Guangting, Chong, and Yizhou for their following CVPR papers accepted as Oral Papers: 1. Tracking by Instance Detection: A Meta-Learning Approach (Guangting Wang, Chong Luo, Xiaoyan Sun, Zhiwei Xiong, Wenjun Zeng);2. Spatiotemporal Fusion in 3D CNNs: A Probabilistic View (Yizhou Zhou, Xiaoyan Sun, Chong Luo, Zheng-Jun Zha, Wenjun Zeng)<\/div>\n<div>*<\/div>\n<div>(Feb. 2020) <strong>Congratulations<\/strong> to Yizhou Zhou, Xiaoyan Sun, Chong Luo, Xin Jin, Cuiling Lan, Zhizheng Zhang, Guangting Wang, Zhe Zhang, Chunyu Wang, Pengfei Zhang, for the acceptance of their papers by CVPR 2020!! The Intelligent Multimedia Group has a total of 8 papers accepted.<\/div>\n<div>*<\/div>\n<div>(July 2019) <strong>Congratulations<\/strong> to Junsheng and Yuwang for the acceptance of their papers entitled \u201cUnsupervised High-Resolution Depth Learning from Videos with Dual Networks\u201d and &#8220;Moving Indoor: Unsupervised Video Depth Learning in Challenging Environments&#8221; by ICCV 2019!!<\/div>\n<div>*<\/div>\n<div>(July 2019) <strong>Congratulations<\/strong> to Haibo, Chunyu, and Jingdong for the acceptance of their paper entitled \u201cCross View Fusion for 3D Human Pose Estimation\u201d by ICCV 2019!!<\/div>\n<div>*<\/div>\n<div>(July 10, 2019) Dr. Wenjun Zeng presented as a panelist at the industry panel on &#8220;From Papers to Products: Bridging the Gap between Multimedia Research and Practical Applications&#8221; at the 2019 IEEE Inter. Conf. Multimedia & Expo held in Shanghai, July 8-12!!<\/div>\n<div>*<\/div>\n<div>(July 2019) <strong>Congratulations<\/strong> to our collaborator (as part of the MSRA Collaborative Research Program) Prof. Wei-shi Zheng and his team at Sun Yat-Sen University for the acceptance of their paper entitled \u201cPredicting Future Instance Segmentation with Contextual Pyramid ConvLSTMs&#8221; by ACM Multimedia 2019!!<\/div>\n<div>*<\/div>\n<div>(July 2019) <strong>Congratulations<\/strong> to Guoqiang and Cuiling for the acceptance of their paper entitled \u201cView Invariant 3D Human Pose Estimation\u201d by IEEE Trans. on Cir. and Sys. for Video Technology!<\/div>\n<div>*<\/div>\n<div>(April 2019) <strong>Congratulations<\/strong> to Peng, Chunyu, and Jingdong for the acceptance of their paper entitled \u201cObject Detection in Videos by High Quality Object Linking\u201d by IEEE Transactions on Pattern Analysis and Machine Intelligence!!<\/div>\n<div>*<\/div>\n<div>(Feb. 2019) <strong>Congratulations<\/strong> to Zhizheng and Cuiling for the acceptance of their paper entitled \u201cDensely Semantically Aligned Person Re-Identification\u201d by CVPR 2019!!<\/div>\n<div>*<\/div>\n<div>(Feb. 2019) <strong>Congratulations<\/strong> to Guangting and Chong for the acceptance of their paper entitled \u201cSPM-Tracker: Series-Parallel Matching for Real-Time Visual Object Tracking\u201d by CVPR 2019!!<\/div>\n<div>*<\/div>\n<div>(Feb. 2019) <strong>Congratulations<\/strong> to Yizhou and Xiaoyan for the acceptance of their paper entitled \u201cContext-Reinforced Semantic Segmentation\u201d by CVPR 2019!!<\/div>\n<div>*<\/div>\n<div>(Feb. 2019) <strong>Congratulations<\/strong> to Bi and Wenxuan for the acceptance of the paper entitled \u201cLearning to Update for Object Tracking with Recurrent Meta-learner\u201d by the IEEE Transactions on Image Processing!!<\/div>\n<div>*<\/div>\n<div>(Jan. 2019) <strong>Congratulations<\/strong> to Pengfei and Cuiling for the acceptance of the paper entitled \u201cView Adaptive Neural Networks for High Performance Skeleton-based Human Action Recognition\u201d by the IEEE Transactions on Pattern Analysis and Machine Intelligence!!<\/div>\n<div>*<\/div>\n<div>(Jan. 2019) <strong> Congratulations <\/strong> to Xiaolin, Cuiling, and Xiaoyan for the acceptance of the paper entitled \u201cTemporal-Spatial Mapping for Action Recognition\u201d by the IEEE Transactions on Circuits and Systems for Video Technology!!<\/div>\n<div>*<\/div>\n<div>(Dec. 2018) Dr. Wenjun Zeng served on the judging committee of the AI Challenger Global AI Contest (https:\/\/challenger.ai\/?lan=en ).<\/div>\n<div>*<\/div>\n<div>(Oct. 2018) <strong>Congratulations<\/strong> to Dr. Wenjun Zeng for receiving the <strong>2018 Industrial Distinguished Leader Award<\/strong> from APSIPA (Asia Pacific Signal and Information Processing Association, www.apsipa.org)!!<\/div>\n<div>*<\/div>\n<div>(Sept. 2018) <strong> Congratulations <\/strong> to Anfeng and Chong for winning the second place, among 72 submissions\/entries, in the 6th Visual Object Tracking Challenge VOT2018 (http:\/\/www.votchallenge.net\/vot2018\/ ) real-time tracker sub-challenge, held in conjunction with ECCV2018!!<\/div>\n\t\t\t<\/div>\n\t\t<\/div>\n\t<\/li>\n\t\t\t\t\t\t<\/ul>\n\t<\/div>\n\t\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>The Intelligent Multimedia (IM) group aims to build seamless yet efficient multimedia applications and services through breakthroughs in fundamental theory and innovations in algorithm and system technology. We address the problems of intelligent multimedia content sensing, processing, analysis, services, and the generic scalability issues of multimedia computing systems. Current research focus is on video analytics [&hellip;]<\/p>\n","protected":false},"featured_media":0,"template":"","meta":{"msr-url-field":"","msr-podcast-episode":"","msrModifiedDate":"","msrModifiedDateEnabled":false,"ep_exclude_from_search":false,"_classifai_error":"","msr_group_start":"","footnotes":""},"research-area":[13556,13562,13551],"msr-group-type":[243694],"msr-locale":[268875],"msr-impact-theme":[],"class_list":["post-144711","msr-group","type-msr-group","status-publish","hentry","msr-research-area-artificial-intelligence","msr-research-area-computer-vision","msr-research-area-graphics-and-multimedia","msr-group-type-group","msr-locale-en_us"],"msr_group_start":"","msr_detailed_description":"","msr_further_details":"","msr_hero_images":[],"msr_research_lab":[199560,1012650],"related-researchers":[{"type":"user_nicename","display_name":"Yan Lu","user_id":34969,"people_section":"Group 1","alias":"yanlu"},{"type":"user_nicename","display_name":"Chong Luo","user_id":31450,"people_section":"Group 1","alias":"cluo"},{"type":"user_nicename","display_name":"Cuiling Lan","user_id":31487,"people_section":"Group 1","alias":"culan"},{"type":"user_nicename","display_name":"Wenxuan Xie","user_id":34826,"people_section":"Group 1","alias":"wenxie"},{"type":"user_nicename","display_name":"Xiaoyi Zhang","user_id":40573,"people_section":"Group 1","alias":"xiaoyizhang"}],"related-publications":[503525,487076,503543,503555,503207,381029,487070,446622,487031,487058,487064,487043,487052,487025,487019,446637,446607,446616,446643,446631,380945,381044,381089,381074,373211,327203,373220,327221,327209,373232,327230,374480,327227,268446,268464,268482,374510,265095,238322,327218,374519,265092,238252,238251,238318,238320,238321,238314,265098,268419,504548,504566,506147,567030,567045,567057,567072,570015,570027,570039,571770,579259,599202,599208,599214,599220,599232,600555,600600,600606,600624,600729,608238,628890,628896,628902,628908,633105,639132,639144,639156,639168,639180,639192,639204,644952,651510,651531,652149,671655,693123,693135,714028,714043,717547,723916,730822,730828,730834,740146,756991,756997,759631,759652,759658,759664,759670,759679,759721,759727,759766,759775,759949,763102,763108,763114,764494,764500],"related-downloads":[],"related-videos":[793352,793517],"related-projects":[],"related-events":[],"related-opportunities":[1059639],"related-posts":[299177,307334],"tab-content":[],"msr_impact_theme":[],"_links":{"self":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-group\/144711","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-group"}],"about":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/types\/msr-group"}],"version-history":[{"count":99,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-group\/144711\/revisions"}],"predecessor-version":[{"id":1143392,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-group\/144711\/revisions\/1143392"}],"wp:attachment":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media?parent=144711"}],"wp:term":[{"taxonomy":"msr-research-area","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/research-area?post=144711"},{"taxonomy":"msr-group-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-group-type?post=144711"},{"taxonomy":"msr-locale","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-locale?post=144711"},{"taxonomy":"msr-impact-theme","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-impact-theme?post=144711"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}