{"id":1114599,"date":"2025-01-03T23:11:11","date_gmt":"2025-01-04T07:11:11","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/?post_type=msr-academic-program&#038;p=1114599"},"modified":"2025-06-11T06:42:22","modified_gmt":"2025-06-11T13:42:22","slug":"photorealistic-avatar-challenge-cvpr-2025","status":"publish","type":"msr-academic-program","link":"https:\/\/www.microsoft.com\/en-us\/research\/academic-program\/photorealistic-avatar-challenge-cvpr-2025\/","title":{"rendered":"Photorealistic Avatar Challenge CVPR 2025"},"content":{"rendered":"\n\n<p><\/p>\n\n\n\n\n\n\n<p><strong>Program dates<\/strong>: January 15, 2025, to June 10, 2025<\/p>\n\n\n\n<p>The CVPR 2025 Photorealistic Avatar Challenge is intended to stimulate research in the field of photorealistic avatars. The challenge provides a test set and methodology to subjectively evaluate photorealistic avatars for news anchor and telecommunication scenarios. The evaluation task is audio\/video driven self-reenactment.&nbsp;Test subjects will be sitting or standing but only the upper half of the body is rendered for tracks 1 and 2; track 3 will have head-only video. Speech, facial emotions, head turning, and hand gestures sequences are included. The inputs to the challenge are enrollment and test video clips. The output are self-reenactment video clips which we will evaluate with our subjective test framework and are used as the challenge metric.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"challenge-tracks\">Challenge tracks<\/h3>\n\n\n\n<p>There are three challenge tracks:<\/p>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Half-body real-time<\/li>\n\n\n\n<li>Half-body non-real-time<\/li>\n\n\n\n<li>Head-only non-real-time &#8211; <strong>NEW!<\/strong><\/li>\n<\/ol>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"challenge-tracks\">Registration procedure<\/h3>\n\n\n\n<p><strong>Registration is open!<\/strong> To register for the challenge,\u202fparticipants are required to email the CVPR Photorealistic Avatar Challenge organizers&nbsp;<a href=\"mailto:avatar_challenge@microsoft.com\">avatar_challenge@microsoft.com<\/a>&nbsp;with the name of their team members, emails, affiliations, team name, track(s) participating in, team captain, and tentative paper title. Participants also need to register on the&nbsp;<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/cmt3.research.microsoft.com\/AvatarChallenge2025\/\" target=\"_blank\" rel=\"noopener noreferrer\">Challenge CMT<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>&nbsp;site where they can submit the enhanced clips. Registration data is captured and stored in the US.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"challenge-tracks\">Submission instructions<\/h3>\n\n\n\n<p>Please use&nbsp;<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/cmt3.research.microsoft.com\/AvatarChallenge2025\/\" target=\"_blank\" rel=\"noopener noreferrer\">Microsoft Conference Management Toolkit<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>&nbsp;for submitting the results. After logging in, complete the following steps to submit the results<\/p>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Choose \u201cCreate new submission\u201d in the Author Console.<\/li>\n\n\n\n<li>Enter the title, abstract, and co-authors, and upload a&nbsp;<em>lastname<\/em>.txt file (can be empty or contain additional information regarding the submission).<\/li>\n\n\n\n<li>Compress the enhanced results files to a single&nbsp;<em>lastname<\/em>.zip file, retaining the same folder and file names as the blind test set (max file size: 1.8 GB).<\/li>\n\n\n\n<li>After creating the submission, return to the \u201cAuthor Console\u201d (by clicking on \u201cSubmissions\u201d at the top of the page) and upload the&nbsp;<em>lastname<\/em>.zip file via \u201cUpload Supplementary Material\u201d.<\/li>\n<\/ol>\n\n\n\n<p><strong>Contact us:<\/strong>&nbsp;For questions, please contact&nbsp;<a href=\"mailto:avatar_challenge@microsoft.com\"><\/a><a href=\"mailto:avatar_challenge@microsoft.com\">avatar_challenge@microsoft.com<\/a>&nbsp;<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"sample-images-from-some-of-the-clips-in-the-test-set-tracks-1-and-2\">Sample images from some of the clips in the test set (tracks 1 and 2)<\/h3>\n\n\n\n<figure class=\"wp-block-table\"><table class=\"has-fixed-layout\"><tbody><tr><td><img decoding=\"async\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/ilya_emotion.gif\" alt=\"Video clip with gestures and emotions\"><\/td><td><img loading=\"lazy\" decoding=\"async\" width=\"720\" height=\"1280\" class=\"wp-image-1128789\" style=\"width: 400px\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/id3.png\" alt=\"id3\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/id3.png 720w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/id3-169x300.png 169w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/id3-576x1024.png 576w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/id3-101x180.png 101w\" sizes=\"auto, (max-width: 720px) 100vw, 720px\" \/><\/td><td><img loading=\"lazy\" decoding=\"async\" width=\"720\" height=\"1280\" class=\"wp-image-1128783\" style=\"width: 400px\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/id1-2.png\" alt=\"id1\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/id1-2.png 720w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/id1-2-169x300.png 169w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/id1-2-576x1024.png 576w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/id1-2-101x180.png 101w\" sizes=\"auto, (max-width: 720px) 100vw, 720px\" \/><\/td><\/tr><\/tbody><\/table><\/figure>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"sample-image-from-some-of-the-clips-in-the-test-set-track-3\">Sample image from some of the clips in the test set (track 3)<\/h2>\n\n\n\n<figure class=\"wp-block-image size-large is-resized\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"794\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/Sample-1024x794.png\" alt=\"Track 3\" class=\"wp-image-1128384\" style=\"width:476px;height:auto\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/Sample-1024x794.png 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/Sample-300x233.png 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/Sample-768x596.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/Sample-232x180.png 232w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/Sample.png 1497w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure>\n\n\n\n\n\n<p>Photorealistic avatars are human avatars that look, move, and talk like real people. Photorealistic avatars can be used for various applications, such as telecommunication, health care, education, retail and e-commerce, and entertainment. There has been a significant increase in research and publications about photorealistic avatars. CVPR 2024 included 36 papers on photorealistic avatars using a variety of methods. While there are a few common test sets that are used, there is no one test set that all use. Also, only quantitative metrics are used (PSNR, SSIM, and LPIPS) which have well known limitations like weak correlation to subjective realism. None of the CVPR 2024 photorealistic avatar papers included subjective tests for avatar performance, in part because this is a challenging task but also because there is no standardized or readily available method to do so.<\/p>\n\n\n\n<p>In [1] we define and implement the first multidimensional measurement of photorealistic avatar quality of experience. We provide an open source implementation of the subjective test framework based on our extension to ITU-T P.910. We include subjective measurements of avatar realism, affinity, trust, comfortableness using, comfortableness interacting, appropriateness for work, creepiness, formality, resemblance to the person, emotion accuracy, and gesture accuracy. We show that except for resemblance the correlation of these subjective metrics to PSNR, SSIM, and LPIPS is weak; the correlation for emotion accuracy is moderate. For example, the avatar with the best PSNR, SSIM, and LPIPS in [1] (MS3_0 in Table 4) is average in terms of the subjective metrics. In other words, the objective metrics PSNR, SSIM, and LPIPS cannot be used to accurately stack rank the subjective performance of photorealistic avatars. The crowdsourced subjective test framework we have developed has been shown to be highly reproducible and accurate compared to a panel of experts. We also found that for avatars above a certain level of realism (mean opinion score > 2.5 in a 1-5 scale) these measured dimensions are highly correlated. In particular, for photorealistic avatars there is a linear relationship between avatar affinity and realism.<\/p>\n\n\n\n<p>This challenge will provide a test set and methodology to subjectively evaluate photorealistic avatars for news anchor and telecommunication scenarios. Test subjects will be sitting or standing but only the upper half of the body is rendered. Speech, facial emotions, head turning, and hand gestures sequences are included.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"tracks-1-and-2\">Tracks 1 and 2<\/h3>\n\n\n\n<p>The challenge includes two tracks for half-body, (1) real-time and (2) non-real-time. Tracks 1 and 2 are evaluated on the same test set and target a face and upper torso avatar including hands. The real-time track must be evaluated on a NVIDIA RTX 4090 or equivalent GPU and must render the avatar at 1080p 30 FPS (each frame must be captured, processed, and rendered in less than 33 ms total). Teams can submit entries to tracks 1 and 2, but for the non-real-time track the total processing time per frame must be greater than 33 ms. The avatars must be causal and not use any future frames in the rendering of the current frame. <\/p>\n\n\n\n<p>The <strong>challenge metric<\/strong> is the mean of the subjective dimensions defined in [1] with the addition of gesture accuracy. Specifically:<\/p>\n\n\n\n<p><strong>Challenge metric<\/strong> = mean(realism, resemblance to the person, emotion accuracy, gesture accuracy)<\/p>\n\n\n\n<p>At least N=30 ratings per clip will be used for ranking the entries, and statistical tests will be done to determine ties. Additional metrics defined in [1] will be measured but will not be used as the challenge metric.<\/p>\n\n\n\n<p>The test set will consist of a data from 10 people. It will consist of 5 males, 5 females, with a mix of Caucasian, Asian, and Black races for diversity. Each person has a 60 second enrollment clip with head motions, expressions, and speaking. The enrollment clip may be captured on different days as the test set. The test set consists of a 10 second speaking clip, a 15 second non-speaking clip that includes 6 emotions (happy, sad, anger, fear, surprised, and disgust), and a 20 second clip that includes hand gestures and head turns up to 90 degrees. An initial test set of 5 subjects will be provided at the beginning of the challenge, and the final test set will be provided 1 week before challenge submissions. The enrollment and test clips will be captured with a white background. Challenge submissions must be provided in 1080p mp4 format with CRF=17. The clips will be captured on recent iPhone cameras and will be provided in 4K 30 FPS. The baseline avatars will be [2] (which does well with gestures but is not as realistic as live video) and [3] (which is very realistic but does not handle gestures).<\/p>\n\n\n\n<p>The test set is available at <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/github.com\/microsoft\/PhotorealisticAvatarChallenge\">CVPR 2025 Photorealistic Avatar Challenge test set<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"challenge-input-output\">Challenge input\/output<\/h3>\n\n\n\n<p>The general process for the challenge is as follows:<\/p>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Create an avatar model<\/li>\n\n\n\n<li>For each enrollment clip\n<ul class=\"wp-block-list\">\n<li>Enroll the avatar with the enrollment clip<\/li>\n\n\n\n<li>For each test clip for that enrollment clip\n<ul class=\"wp-block-list\">\n<li>Use the enrolled avatar to drive the avatar with the test clip<\/li>\n\n\n\n<li>The rendered avatar should be shown from two viewpoints: 0 degrees (frontal) and 45 degrees. See the below figure. <\/li>\n\n\n\n<li>The rendered avatar should look like the enrollment clip. The enrollment clip is the same person as the test clip but will likely be captured on different days and will be dressed differently.<\/li>\n<\/ul>\n<\/li>\n<\/ul>\n<\/li>\n<\/ol>\n\n\n\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"571\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/Viewpoints-1024x571.png\" alt=\"Render viewpoints\" class=\"wp-image-1124661\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/Viewpoints-1024x571.png 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/Viewpoints-300x167.png 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/Viewpoints-768x428.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/Viewpoints-1536x856.png 1536w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/Viewpoints-240x134.png 240w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/Viewpoints.png 1802w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"illustration-of-the-output-from-the-evaluations-1\">Illustration of the output from the evaluations [1]:<\/h3>\n\n\n\n<figure class=\"wp-block-image size-large is-resized\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"710\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/MOS_Scores_Model_AllAngles_Font20-1024x710.png\" alt=\"Illustration of the output from the evals\" class=\"wp-image-1120233\" style=\"width:596px;height:auto\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/MOS_Scores_Model_AllAngles_Font20-1024x710.png 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/MOS_Scores_Model_AllAngles_Font20-300x208.png 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/MOS_Scores_Model_AllAngles_Font20-768x533.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/MOS_Scores_Model_AllAngles_Font20-240x167.png 240w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2025\/01\/MOS_Scores_Model_AllAngles_Font20.png 1401w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"track-3\">Track 3<\/h3>\n\n\n\n<p>Track 3 is identical to Track 2 except the video input includes only the head and upper torso and does not include hand gestures. The challenge metric is:<\/p>\n\n\n\n<p><strong>Challenge metric<\/strong> = mean(realism, resemblance to the person, emotion accuracy)<\/p>\n\n\n\n<p>The test set for track 3 will be very similar to track 1 and 2 and will include similar enrollment data. Track 3 must render the avatar at 1080p 30 FPS and at two viewpoints as described in tracks 1 and 2.<\/p>\n\n\n\n<p>Teams can enter all tracks. <\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"references\">References<\/h3>\n\n\n\n<ol class=\"wp-block-list\">\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/scholar.google.com\/scholar?oi=bibs&cluster=2998110979420883852&btnI=1&hl=en\">A multidimensional measurement of photorealistic avatar quality of experience<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>. R Cutler, B Naderi, V Gopal, D Palle&nbsp;&#8211; arXiv preprint arXiv:2411.09066, 2024<\/li>\n\n\n\n<li>Z. Huang, F. Tang, Y. Zhang, X. Cun, J. Cao, J. Li, and T.-Y. Lee, \u201cMake-Your-Anchor: A Diffusion-based 2D Avatar Generation Framework,\u201d in CVPR, 2024<\/li>\n\n\n\n<li>J. Guo, D. Zhang, X. Liu, Z. Zhong, Y. Zhang, P. Wan, and D. Zhang, \u201cLivePortrait: Efficient Portrait Animation with Stitching and Retargeting Control,\u201d July 2024, arXiv:2407.03168.<\/li>\n<\/ol>\n\n\n\n\n\n<h3 class=\"wp-block-heading\" id=\"official-rules\">Official rules<\/h3>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"heading-sponsor\">Sponsor<\/h3>\n\n\n\n<p>These Official Rules (\u201cRules\u201d) govern the operation of the CVPR 2025 Photorealistic Avatar Challenge (see overview) Event Contest (\u201cContest\u201d). Microsoft Corporation, One Microsoft Way, Redmond, WA, 98052, USA, is the Contest sponsor (\u201cSponsor\u201d).<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"heading-definitions\">Definitions<\/h3>\n\n\n\n<p>In these Rules, \u201cMicrosoft\u201d, \u201cwe\u201d, \u201cour\u201d, and \u201cus\u201d, refer to Sponsor and \u201cyou\u201d and \u201cyourself\u201d refers to a Contest participant, or the parent\/legal guardian of any Contest entrant who has not reached the age of majority to contractually obligate themselves in their legal place of residence. \u201cEvent\u201d refers to the CVPR 2025 Photorealistic Avatar Challenge (the \u201cEvent\u201d). By entering you (your parent\/legal guardian if you are not the age of majority in your legal place of residence) agree to be bound by these Rules.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"heading-entry-period\">Entry period<\/h3>\n\n\n\n<p>The Contest will operate from Jan 15, 2025 to April 7, 2025 (\u201cEntry Period\u201d). The Entry Period is divided into several periods as described in section How to Enter.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"heading-eligibility\">Eligibility<\/h3>\n\n\n\n<p>Open to any registered Event attendee 18 years of age or older. If you are 18 years of age or older but have not reached the age of majority in your legal place of residence, then you must have consent of a parent\/legal guardian. Employees and directors of Microsoft Corporation and its subsidiaries, affiliates, advertising agencies, and Contest Parties are not eligible, nor are persons involved in the execution or administration of this promotion, or the family members of each above (parents, children, siblings, spouse\/domestic partners, or individuals residing in the same household). Void in Cuba, Iran, North Korea, Sudan, Syria, Region of Crimea, and where prohibited. For business\/tradeshow events: If you are attending the Event in your capacity as an employee, it is your sole responsibility to comply with your employer\u2019s gift policies. Microsoft will not be party to any disputes or actions related to this matter.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"heading-how-to-enter\">How to enter<\/h3>\n\n\n\n<p>The Contest Objective is to promote collaborative research in photorealistic avatars aimed to maximize the subjective (perceptual) quality of avatars. Winners will be determined based on the subjective quality of the avatars using an online test framework described <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/arxiv.org\/pdf\/2411.09066\">here<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>. Only models described in accepted CVPR Photorealistic Avatar Challenge papers will be eligible for winning the Contest. You may participate as an individual or a team. If forming a team, you must designate a \u201cTeam Captain\u201d who will submit all entry materials on behalf of the team. Once you register as part of a Team, you cannot change Teams or alter your current team (either by adding or removing members) after the submission of your Entry. Limit one Entry per person and per team. You may not compete on multiple teams and you may not enter individually and on a team. We are not responsible for Entries that we do not receive for any reason, or for Entries that we receive but are not decipherable or not functional for any reason. Each Team is solely responsible for its own cooperation and teamwork. In no event will Microsoft officiate in any dispute regarding the conduct or cooperation of any Team or its members. The Contest will operate as follows: Registration \/ Development Period: January 15, 2025 to April 17, 2025. To register, please send an email to&nbsp;<a href=\"mailto:aec_challenge@microsoft.com\">avatar_challenge@microsoft.com<\/a>&nbsp;stating that you are interested to participate in the challenge. Please include the following details in your email:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Names of the participants and name of the team captain<\/li>\n\n\n\n<li>Institution\/Company<\/li>\n\n\n\n<li>Email<\/li>\n\n\n\n<li>Tentative title of the paper<\/li>\n<\/ul>\n\n\n\n<p>Then, i. develop an avatar model that best meets the Contest Objective&nbsp;and ii. submit a CVPR 2025 Photorealistic Avatar Challenge paper via&nbsp;<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/cmt3.research.microsoft.com\/AvatarChallenge2025\">Microsoft Conference Management Toolkit (opens in new tab)<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>&nbsp;which reports the computational complexity of the model in terms of the number of parameters. For the real-time track also include the time it takes to capture, process, and render a frame on a NVIDIA 4090 or equivalent GPU. To develop your model, use any available dataset for training data. The final evaluation will be conducted on a blind test set that is similar to the open sourced test set. On February 21, 2025 the blind test dataset will be made available. You will have until 11:59 PM PT on February 28, 2025 to test your model against this dataset and create a set of enhanced clips to submit for judging (your \u201cEntry\u201d). The rules of the challenge are as follows:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>For real-time track, the frame processing must take less than the 33 ms on a NVIDIA 4090 or equivalent GPU. In addition, no future video frames can be used during processing for both encoding and decoding. The time constraint is for encoding + decoding (i.e., end-to-end time). For example, if capture+encoding takes 16ms and decoding (rendering) takes 16ms then the end-to-end time is 32ms which meets the real-time requirements.<\/li>\n\n\n\n<li>For non-real-time tracks, there are no constraints on computation time, but no future video frames can be used during processing (for both encoding and decoding).<\/li>\n\n\n\n<li>Winners will be picked based on the subjective MOS evaluated on the blind test set using the test framework described <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/arxiv.org\/pdf\/2411.09066\">here<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>.<\/li>\n\n\n\n<li>The blind test set will be made available to the participants on February 21, 2025. Participants must send the results achieved by their developed models to the organizers. We will use the submitted clips to conduct subjective evaluation and pick the winners based on the results. Participants are forbidden from using the blind test set to retrain or tune their models. Failing to adhere to these rules will lead to disqualification from the challenge.<\/li>\n\n\n\n<li>Participants should report the computational complexity of their model in terms of the number of parameters and the time it takes to infer a frame on a NVIDIA 4090 or equivalent GPU.<\/li>\n\n\n\n<li>Submitted papers will undergo the standard peer-review process of CVPR 2025. The paper needs to be accepted to the workshop for the participants to be eligible for the challenge.\n<ul class=\"wp-block-list\">\n<li><strong>New<\/strong>: Papers are optional for teams that already have a CVPR 2025 paper describing their avatar<\/li>\n<\/ul>\n<\/li>\n<\/ul>\n\n\n\n<p>CVPR 2025 Photorealistic Avatar Challange Paper Submission and Judging Period: March 15, 2025 11:59 PM PT to April 1, 2025. To submit a paper, visit&nbsp;<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/cmt3.research.microsoft.com\/AvatarChallenge2025\">Microsoft Conference Management Toolkit<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/cmt3.research.microsoft.com\/AvatarChallenge2025\">(opens in new tab)<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>. The entry limit is one per person during the Entry Period. Any attempt to obtain more than the stated number of entries by using multiple\/different accounts, identities, registrations, logins, or any other methods will void your entries and you may be disqualified. Use of any automated system to participate is prohibited. We are not responsible for excess, lost, late, or incomplete entries. If disputed, entries will be deemed submitted by the \u201cauthorized account holder\u201d of the email address, social media account, or other method used to enter. The \u201cauthorized account holder\u201d is the natural person assigned to an email address by an internet or online service provider, or other organization responsible for assigning email addresses.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"heading-eligible-entry\">Eligible entry<\/h3>\n\n\n\n<p>To be eligible, an entry must meet the following content\/technical requirements:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Your Entry must be the method described in a paper accepted as CVPR 2025 Photorealistic Avatar Challange paper. Papers should use the&nbsp;<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/github.com\/cvpr-org\/author-kit\/releases\">official CVPR 2025 template<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>. <\/li>\n\n\n\n<li>Your entry must be your own original work; and<\/li>\n\n\n\n<li>You must have obtained any and all consents, approvals, or licenses required for you to submit your entry; and<\/li>\n\n\n\n<li>To the extent that entry requires the submission of user-generated content such as software, photos, videos, music, artwork, essays, etc., entrants warrant that their entry is their original work, has not been copied from others without permission or apparent rights, and does not violate the privacy, intellectual property rights, or other rights of any other person or entity. You may include Microsoft trademarks, logos, and designs, for which Microsoft grants you a limited license to use for the sole purposes of submitting an entry into this Contest; and<\/li>\n\n\n\n<li>Your entry may NOT contain, as determined by us in our sole and absolute discretion, any content that is obscene or offensive, violent, defamatory, disparaging or illegal, or that promotes alcohol, illegal drugs, tobacco or a particular political agenda, or that communicates messages that may reflect negatively on the goodwill of Microsoft.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"heading-use-of-entries\">Use of entries<\/h3>\n\n\n\n<p>We are not claiming ownership rights to your Submission. However, by submitting an entry, you grant us an irrevocable, royalty-free, worldwide right and license to use, review, assess, test and otherwise analyze your entry and all its content in connection with this Contest and use your entry in any media whatsoever now known or later invented for any non-commercial or commercial purpose, including, but not limited to, the marketing, sale or promotion of Microsoft products or services, without further permission from you. You will not receive any compensation or credit for use of your entry, other than what is described in these Official Rules. By entering you acknowledge that we may have developed or commissioned materials similar or identical to your entry and you waive any claims resulting from any similarities to your entry. Further you understand that we will not restrict work assignments of representatives who have had access to your entry and you agree that use of information in our representatives\u2019 unaided memories in the development or deployment of our products or services does not create liability for us under this agreement or copyright or trade secret law. Your entry may be posted on a public website. We are not responsible for any unauthorized use of your entry by visitors to this website. We are not obligated to use your entry for any purpose, even if it has been selected as a winning entry.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"heading-winner-selection-and-notification\">Winner selection and notification<\/h3>\n\n\n\n<p>Pending confirmation of eligibility, potential winners will be selected by Microsoft or their Agent or a qualified judging panel from among all eligible entries received based on the following judging criteria: 99% \u2013 The subjective quality is evaluated on the blind test set using the subjective test framework. We will use the submitted clips with no alteration to conduct subjective evaluation and pick the winners based on the results.&nbsp;The challenge metric is an average of Realism, Gesture Accuracy, Expression Accuracy, and Resemblance. 1% \u2013 The Entry was described in an accepted CVPR 2025 Photorealistic Avatar Challenge paper. Winners will be selected within 7 days following the event. Winners will be notified within 7 days following the Event. In the event of a tie between any eligible entries, an additional judge will break the tie based on the judging criteria described above. The decisions of the judges are final and binding. If public vote determines winners, it is prohibited for any person to obtain votes by any fraudulent or inappropriate means, including offering prizes or other inducements in exchange for votes, automated programs or fraudulent ID\u2019s. Microsoft will void any questionable votes.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"heading-odds\">Odds<\/h3>\n\n\n\n<p>The odds of winning are based on the number and quality of eligible entries received.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"heading-general-conditions-and-release-of-liability\">General conditions and release of liability<\/h3>\n\n\n\n<p>To the extent allowed by law, by entering you agree to release and hold harmless Microsoft and its respective parents, partners, subsidiaries, affiliates, employees, and agents from any and all liability or any injury, loss, or damage of any kind arising in connection with this. All local laws apply. The decisions of Microsoft are final and binding. We reserve the right to cancel, change, or suspend this Contest for any reason, including cheating, technology failure, catastrophe, war, or any other unforeseen or unexpected event that affects the integrity of this Contest, whether human or mechanical. If the integrity of the Contest cannot be restored, we may select winners from among all eligible entries received before we had to cancel, change or suspend the Contest. If you attempt or we have strong reason to believe that you have compromised the integrity or the legitimate operation of this Contest by cheating, hacking, creating a bot or other automated program, or by committing fraud in any way, we may seek damages from you to the full extent of the law and you may be banned from participation in future Microsoft promotions.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"heading-governing-law\">Governing law<\/h3>\n\n\n\n<p>This Contest will be governed by the laws of the State of Washington, and you consent to the exclusive jurisdiction and venue of the courts of the State of Washington for any disputes arising out of this Contest.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"heading-privacy\">Privacy<\/h3>\n\n\n\n<p>At Microsoft, we are committed to protecting your privacy. Microsoft uses the information you provide on this form to notify you of important information about our products, upgrades and enhancements, and to send you information about other Microsoft products and services. Microsoft will not share the information you provide with third parties without your permission except where necessary to complete the services or transactions you have requested, or as required by law. Microsoft is committed to protecting the security of your personal information. We use a variety of security technologies and procedures to help protect your personal information from unauthorized access, use, or disclosure. Your personal information is never shared outside the company without your permission, except under conditions explained above. If you believe that Microsoft has not adhered to this statement, please contact Microsoft by sending an email to&nbsp;<a href=\"mailto:privrc@microsoft.com\">privrc@microsoft.com<\/a>&nbsp;or postal mail to Microsoft Privacy Response Center, Microsoft Corporation, One Microsoft Way, Redmond, WA 98052<\/p>\n\n\n\n<p><\/p>\n\n\n\n<p><\/p>\n\n\n\n\n\n<h3 class=\"wp-block-heading\" id=\"program-timeline\">Program timeline<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Start of challenge (release initial test set): January 15, 2025<\/strong><\/li>\n\n\n\n<li>Release final test set: March 15, 2025<\/li>\n\n\n\n<li>Results submitted for evaluation: March 20, 2025<\/li>\n\n\n\n<li>Results of subjective evaluation: March 25, 2025<\/li>\n\n\n\n<li>Submit papers for review: March 27, 2025<\/li>\n\n\n\n<li>Results of paper review: April 3, 2025<\/li>\n\n\n\n<li>Workshop camera-ready submission: April 7,2025<\/li>\n\n\n\n<li>CVPR 2025 Photorealistic Avatar Challenge workshop: June 10, 2025<\/li>\n<\/ul>\n\n\n\n<p><\/p>\n\n\n\n\n\n<h3 class=\"wp-block-heading\" id=\"organizers\">Organizers<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/rosscutler.github.io\/\">Ross Cutler<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, Microsoft, ross.cutler@microsoft.com<\/li>\n\n\n\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/ch.linkedin.com\/in\/valentinjulien\">Julien Valentin<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, Meta, JulienValentin@meta.com<\/li>\n\n\n\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/justusthies.github.io\/\">Justus Thies<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, TU Darmstadt, justus.thies@tu-darmstadt.de<\/li>\n\n\n\n<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/babaknaderi\/\">Babak Naderi<\/a>, Microsoft, babaknaderi@microsoft.com<\/li>\n\n\n\n<li>Vishak Gopal, Microsoft, vishak.gopal@microsoft.com<\/li>\n\n\n\n<li>Ilya Gurvich, Microsoft, ilyagu@microsoft.com<\/li>\n\n\n\n<li>Oliver Graham, Microsoft, grahamoliver@microsoft.com<\/li>\n<\/ul>\n\n\n\n\n\n<h3 class=\"wp-block-heading\" id=\"related-links\">Related links<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/github.com\/microsoft\/PhotorealisticAvatarChallenge\">CVPR 2025 Photorealistic Avatar Challenge test set<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<\/ul>\n\n\n\n\n\n<p>TBD<\/p>\n\n\n\n\n\n<p>The workshop is on June 11, 2025. The schedule is:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>1:00PM: Photorealistic Avatar Challenge Workshop: Ross Cutler (Distinguished Engineer, Microsoft)<\/li>\n\n\n\n<li>1:40PM: <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/arxiv.org\/pdf\/2502.19739\">LUCAS: Layered Universal Codec Avatars<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>: Di Liu (Meta)<\/li>\n\n\n\n<li>2:00PM: <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/arxiv.org\/pdf\/2403.11453\">Hybrid Explicit Representation for Ultra-Realistic Head Avatars<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>: Juyong Zhang (USTC)<\/li>\n\n\n\n<li>2:20PM: Invited speaker: Jason Saragih (Director Research Scientist, Meta): Metrics, Perception and The Expert Eye: Developing Codec Avatars for Telepresence<\/li>\n\n\n\n<li>3:00PM: <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/arxiv.org\/pdf\/2503.08224\">HRAvatar: High-Quality and Relightable Gaussian Head Avatar<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> Yu Li (Tsinghua University)<\/li>\n\n\n\n<li>3:20PM: <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/conallwang.github.io\/MeGA_Pages\/\">MeGA: Hybrid Mesh-Gaussian Head Avatar for High-Fidelity Rendering and Head Editing<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>: <span data-teams=\"true\">Cong Wang<\/span> (Tsinghua University)<\/li>\n\n\n\n<li>3:40PM: <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/arxiv.org\/pdf\/2407.05712\">MobilePortrait: Real-Time One-Shot Neural Head Avatars on Mobile Devices<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>: Jianwen Jiang (ByteDance)<\/li>\n<\/ul>\n\n\n","protected":false},"featured_media":0,"template":"","meta":{"msr-url-field":"","msr-podcast-episode":"","msrModifiedDate":"","msrModifiedDateEnabled":false,"ep_exclude_from_search":false,"_classifai_error":"","msr_hide_image_in_river":null,"footnotes":""},"msr-opportunity-type":[187426],"msr-region":[],"msr-locale":[268875],"msr-program-audience":[],"msr-post-option":[269142,269151],"msr-impact-theme":[],"class_list":["post-1114599","msr-academic-program","type-msr-academic-program","status-publish","hentry","msr-opportunity-type-challenges","msr-locale-en_us","msr-post-option-include-in-river","msr-post-option-reject-for-river"],"msr_description":"","msr_social_media":[],"related-researchers":[{"type":"user_nicename","display_name":"Ross Cutler","user_id":40660,"people_section":"Section name 0","alias":"rcutler"},{"type":"user_nicename","display_name":"Vishak Gopal","user_id":39624,"people_section":"Section name 0","alias":"vigopal"},{"type":"user_nicename","display_name":"Babak Naderi","user_id":42525,"people_section":"Section name 0","alias":"babaknaderi"}],"tab-content":[],"msr_impact_theme":[],"_links":{"self":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-academic-program\/1114599","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-academic-program"}],"about":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/types\/msr-academic-program"}],"version-history":[{"count":69,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-academic-program\/1114599\/revisions"}],"predecessor-version":[{"id":1141798,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-academic-program\/1114599\/revisions\/1141798"}],"wp:attachment":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media?parent=1114599"}],"wp:term":[{"taxonomy":"msr-opportunity-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-opportunity-type?post=1114599"},{"taxonomy":"msr-region","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-region?post=1114599"},{"taxonomy":"msr-locale","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-locale?post=1114599"},{"taxonomy":"msr-program-audience","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-program-audience?post=1114599"},{"taxonomy":"msr-post-option","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-post-option?post=1114599"},{"taxonomy":"msr-impact-theme","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-impact-theme?post=1114599"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}