{"id":969735,"date":"2023-10-18T20:52:04","date_gmt":"2023-10-19T03:52:04","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/?post_type=msr-academic-program&#038;p=969735"},"modified":"2024-03-20T09:11:40","modified_gmt":"2024-03-20T16:11:40","slug":"bandwidth-estimation-challenge","status":"publish","type":"msr-academic-program","link":"https:\/\/www.microsoft.com\/en-us\/research\/academic-program\/bandwidth-estimation-challenge\/","title":{"rendered":"Bandwidth Estimation Challenge 2024"},"content":{"rendered":"\n\n<p>Offline Reinforcement Learning for Bandwidth Estimation in Real Time Communications<\/p>\n\n\n\n\n\n\n<p>Video conferencing systems have recently emerged as indispensable tools to sustain global business operations and enable accessible education by revolutionizing the way people connect, collaborate, and communicate despite physical barriers and geographical divides. The quality of experience (QoE) delivered by these systems to the end user depends on <em>bandwidth estimation<\/em>, which is the problem of estimating the variable capacity of the bottleneck link between the sender and the receiver over time. In real time communication systems (RTC), the bandwidth estimate serves as a target bit rate for the audio\/video encoder, controlling the send rate from the client. Overestimating the capacity of the bottleneck link causes network congestion as the client sends data at a rate higher than what the network can handle. Network congestion is characterized by increased delays in packet delivery, jitter, and potential packet losses. In terms of user\u2019s experience, users will typically experience many resolution switches, frequent video freezes, garbled speech, and audio\/video desynchronization, to name a few. Underestimating the available bandwidth on the other hand causes the client to encode and transmit the audio\/video streams in a lower rate signal than what the network can handle, which leads to underutilization and degraded QoE. Estimating the available bandwidth accurately is therefore critical to providing the best possible QoE to users in RTC systems. Nonetheless, bandwidth estimation is faced with a multitude of challenges such as dynamic network paths between senders and receivers with fluctuating traffic loads, existence of diverse wired and wireless access network technologies with distinct characteristics, existence of different transmission protocols fighting for bandwidth to carry side and cross traffic, and partial observability of the network as only local packet statistics are available at the client side to base the estimate on.&nbsp;&nbsp;<\/p>\n\n\n\n<p>To improve QoE for users in RTC systems, the ACM MMSys 2024 grand challenge focuses on the development of a deep learning-based bandwidth estimator using offline reinforcement learning (RL) techniques. A real-world dataset of observed network dynamics with objective metrics that reflect user-perceived audio\/video quality in Microsoft Teams is released to train the deep RL policies for bandwidth estimation. &nbsp;<\/p>\n\n\n\n<p><em><strong>Please NOTE<\/strong>&nbsp;that the intellectual property (IP) is not transferred to the challenge organizers, i.e., participants remain the owners of their code (when the code is made publicly available, an appropriate license should be added).<\/em><\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"challenge-task\">Challenge task&nbsp;<\/h3>\n\n\n\n<p>Offline RL is a variant of RL where the agent learns from a fixed dataset of previously collected experiences, without interacting with the environment during training. In offline RL, the goal is to learn a policy that maximizes the expected cumulative reward based on the dataset. Offline RL is different from online RL where the agent can interact with the environment using its updated policy and learn from the feedback it receives online.&nbsp;&nbsp;<\/p>\n\n\n\n<p>In this challenge, participants are provided with a dataset of real-world trajectories for Microsoft Teams audio\/video calls. Each trajectory corresponds to the sequence of high-dimensional observation vector (<em>o<sub>n<\/sub><\/em>) computed based on packet information received by the client in one audio\/video call, along with the bandwidth estimates (<em>b<sub>n<\/sub><\/em>) predicted by different estimators (behavior policies). Objective signals which capture the user-perceived audio\/video quality during the call are provided. These objective signals are predicted by ML models whose predictions have high correlation with subjective audio and video quality scores as determined by ITU-T&#8217;s P.808 and P.910, respectively.<\/p>\n\n\n\n<p>The goal of the challenge is to improve QoE for RTC system users as measured by objective audio\/video quality scores by developing a deep learning-based policy model (receiver-side bandwidth estimator, \u03c0) with offline RL techniques, such as conservative Q-learning, inverse reinforcement learning, and constrained policy optimization, to name a few. To this end, participants are free to specify an appropriate reward function based on the provided dataset of observed network dynamics and objective metrics, the model architecture, and the training algorithm, given that the developed model adheres to the below requirements.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"challenge-requirements\">Challenge requirements<\/h3>\n\n\n\n<p>Failing to adhere to challenge rules will lead to disqualification from the challenge.<\/p>\n\n\n\n<ol class=\"wp-block-list\">\n<li>The policy model (\u03c0) can be a state-less or a stateful model that outputs the bandwidth estimate (<em>b<sub>n<\/sub><\/em>) in bits per second (bps). The input to a stateless model is the observation vector (<em>o<sub>n<\/sub><\/em>), hence, \u03c0<sub>stateless<\/sub> : <em>o<sub>n<\/sub><\/em> \u2192 <em>b<sub>n<\/sub><\/em>. On the other hand, the inputs to a stateful model are the observation vector (<em>o<sub>n<\/sub><\/em>), as well as hidden (<em>h<sub>n-1<\/sub><\/em>) and cell (<em>c<sub>n-1<\/sub><\/em>)  states which are representations learned by the model to capture the underlying structure and temporal dependencies in the sequence of observation vectors, hence, \u03c0<sub>stateful<\/sub>: <em>o<sub>n<\/sub><\/em>, <em>h<sub>n-1<\/sub><\/em>, <em>c<sub>n-1<\/sub><\/em> \u2192 <em>b<sub>n<\/sub><\/em>. Please refer to the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/github.com\/microsoft\/RL4BandwidthEstimationChallenge\/blob\/main\/tf_policy.py\" target=\"_blank\" rel=\"noopener noreferrer\">Tensorflow model class<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> or <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/github.com\/microsoft\/RL4BandwidthEstimationChallenge\/blob\/main\/torch_policy.py\" target=\"_blank\" rel=\"noopener noreferrer\">Pytorch_model_class<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> in the repository which shows the required inputs and outputs. <strong>Any policy model that does not adhere to this input\/output signature will be disqualified from the competition.<\/strong><\/li>\n\n\n\n<li>Feature transformation and\/or feature selection should be performed in a processing block within the model. For instance, the first layer (<em>l<sub>0<\/sub><\/em>) of the model can map the observation vector (<em>o<sub>n<\/sub><\/em>) to a desired agent state (<em>s<sub>n<\/sub><\/em>), <em>l<sub>0<\/sub><\/em>: <em>o<sub>n<\/sub><\/em> \u2192 <em>s<sub>n<\/sub><\/em>.<\/li>\n\n\n\n<li>Participants can specify an appropriate action space, e.g. a<sub>n<\/sub> ~ [0,1], however, the transformation from the action space to the bps space should be performed by the last layer (<em>l<sub>N<\/sub><\/em>) of the model such that the model predicts the bandwidth estimates in bps, <em>l<sub>N<\/sub><\/em>: a<sub>n<\/sub> \u2192 <em>b<sub>n<\/sub>.<\/em><\/li>\n\n\n\n<li>Participants can specify an appropriate reward function for training the RL agent based on the provided signals: audio quality signal, video quality signal, and network metrics in the observation vector.<\/li>\n\n\n\n<li>To reduce the hardware requirements when the policy model is used for inference at the client side of the video conferencing system, the model size must be smaller than 10 MB and inference latency should be no more than 5ms on an Intel Core i5 Quadcore clocked at 2.4 GHz using a single thread. <\/li>\n\n\n\n<li>In offline RL it is typical to use an actor-critic architecture. As long as the inputs to the actor\/policy model adhere to the aforementioned requirements, any set of features can be used as inputs for the critic.<\/li>\n\n\n\n<li>Participants can train the model using PyTorch or TensorFlow, and the model should be exported to <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/onnx.ai\/\" target=\"_blank\" rel=\"noopener noreferrer\">ONNX<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>. To ensure that organizers can run the model correctly, participants are required to share a small subset of their validation data along with their model outputs to be used for verification. We provide sample scripts to convert&nbsp;<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/github.com\/microsoft\/RL4BandwidthEstimationChallenge\/blob\/main\/torch_policy.py\" target=\"_blank\" rel=\"noopener noreferrer\">PyTorch<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>&nbsp;and&nbsp;<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/github.com\/microsoft\/RL4BandwidthEstimationChallenge\/blob\/main\/tf_policy.py\" target=\"_blank\" rel=\"noopener noreferrer\">TF models<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>&nbsp;in the repository. We have also released a&nbsp;<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/github.com\/microsoft\/RL4BandwidthEstimationChallenge\/blob\/main\/onnx_models\/\" target=\"_blank\" rel=\"noopener noreferrer\">baseline stateless model (MLP)<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>&nbsp;as a reference, with an&nbsp;<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/github.com\/microsoft\/RL4BandwidthEstimationChallenge\/blob\/main\/run_baseline_model.py\" target=\"_blank\" rel=\"noopener noreferrer\">example script<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>&nbsp;to run this model.<\/li>\n\n\n\n<li>Participants should submit their training code to the Open-source Software and Datasets track of the conference to receive a reproducibility badge.&nbsp;&nbsp;<\/li>\n<\/ol>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"evaluation-criteria-and-methodology\">Evaluation criteria and methodology&nbsp;<\/h3>\n\n\n\n<p>Evaluation of submitted models will be conducted in a 2-stage evaluation process:&nbsp;<\/p>\n\n\n\n<ol class=\"wp-block-list\" start=\"1\">\n<li>In the first stage, <strong>all the submitted models<\/strong> <strong>that adhere to the challenge requirements<\/strong> will be evaluated in our emulation platform on a set of validation network traces with multiple repetitions per network trace. Evaluating the submitted models in our emulation platform, which is a controlled lab environment of connected nodes with software to emulate different network links, enables us to estimate performance statistics and establish statistical significance between models.<\/li>\n\n\n\n<li>The <strong>top 3 models from stage 1<\/strong> will be evaluated in calls conducted between nodes in our geographically distributed testbed. Each model will be used to make several calls, and performance will be reported based on those calls. <strong>The final ranking of those three models will be determined based on the results of the second evaluation stage. &nbsp;<\/strong><\/li>\n<\/ol>\n\n\n\n<p>In either evaluation stage, the following scoring function will be used to rank the models:&nbsp;<\/p>\n\n\n\n<p class=\"has-text-align-center\"><strong><em>E<sub>calls<\/sub>[E<sub>n<\/sub>[objective video quality + objective audio quality]]<\/em><\/strong><\/p>\n\n\n\n<p>The outer expectation is across all calls and across repetitions per call, while the inner expectation is the temporal average of the objective metrics throughout a call.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"registration-procedure\">Registration procedure<\/h3>\n\n\n\n<p>There are two steps in registering for the challenge: <\/p>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Participants are required to fill <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/forms.office.com\/r\/LE5PR9EL5q\" target=\"_blank\" rel=\"noopener noreferrer\">this form<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> with the list of all the participants, affiliation of each participant (include country name), contact information for participants and name of your team. A confirmation email will be sent once we receive your registration information.<\/li>\n\n\n\n<li>Participants need to register on the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/mmsys24ods.hotcrp.com\/\" target=\"_blank\" rel=\"noopener noreferrer\">Open-source Software and Datasets submission system<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> and follow the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/2024.acmmmsys.org\/participation\/cfp\/\" target=\"_blank\" rel=\"noopener noreferrer\">submission guidelines<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> of the ODS track when submitting their work.<\/li>\n<\/ol>\n\n\n\n<p>Organizers will communicate updates regarding the grand challenge and announce the availability of data, and evaluation results etc. via email. <strong>Challenge leaderboard for each evaluation stage will be posted on the challenge website.<\/strong><\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"awards-1\">Awards<\/h3>\n\n\n\n<p>If accepted, paper submissions to the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/2024.acmmmsys.org\/participation\/cfp\/\">Open-source Software and Datasets<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> will be included in the ACM MMSys 2024 conference proceedings, and code submissions will be given the appropriate <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/www.acm.org\/publications\/policies\/artifact-review-and-badging-current\" target=\"_blank\" rel=\"noopener noreferrer\">reproducibility badge<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>.&nbsp;Authors of accepted papers will also have a chance to present their work during the conference. Moreover, the winner and the runner-up, <strong>based on the results of the second evaluation stage<\/strong>, will be awarded cash prizes as described in the <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/academic-program\/bandwidth-estimation-challenge\/rules\/\" target=\"_blank\" rel=\"noreferrer noopener\">rules section<\/a>.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"contact-us\">Contact us<\/h3>\n\n\n\n<p>Participants with queries related to this grand challenge can either contact Sami Khairy by <a href=\"mailto:samikhairy@microsoft.com\">email<\/a> or create an issue on the Github repository.<\/p>\n\n\n\n<p>&nbsp;<\/p>\n\n\n\n<p><\/p>\n\n\n\n<p><\/p>\n\n\n\n\n\n<h2 class=\"wp-block-heading\" id=\"heading-official-rules\">Official rules<\/h2>\n\n\n\n<p>SPONSOR<\/p>\n\n\n\n<p>These Official Rules (\u201cRules\u201d) govern the operation of the&nbsp;ACM MMSys 2024 Bandwidth Estimation&nbsp;Challenge&nbsp;Event&nbsp;Contest (\u201cContest\u201d). Microsoft Corporation, One Microsoft Way, Redmond, WA, 98052, USA, is the Contest sponsor (\u201cSponsor\u201d).<\/p>\n\n\n\n<p>DEFINITIONS<\/p>\n\n\n\n<p>In these Rules, \u201cMicrosoft\u201d, \u201cwe\u201d, \u201cour\u201d, and \u201cus\u201d, refer to Sponsor, and \u201cyou\u201d and \u201cyourself\u201d refers to a Contest participant or the parent\/legal guardian of any Contest entrant who has not reached the age of majority to contractually obligate themselves in their legal place of residence. \u201cEvent\u201d refers to the&nbsp;ACM MMSys 2024 Bandwidth Estimation Challenge&nbsp;event held in&nbsp;Bari, Italy&nbsp;(the \u201cEvent\u201d). By entering you (your parent\/legal guardian if you are not the age of majority in your legal place of residence) agree to be bound by these Rules.<\/p>\n\n\n\n<p>ENTRY PERIOD<\/p>\n\n\n\n<p>The Contest will operate from&nbsp;October 9,&nbsp;2023&nbsp;to April 18, 2024&nbsp;(\u201cEntry Period\u201d). The Entry Period is divided into several periods as described in Section 5 How to Enter.<\/p>\n\n\n\n<p>ELIGIBILITY<\/p>\n\n\n\n<p>Open to any registered Event attendee 18 years of age or older. If you are 18 years of age or older but have not reached the age of majority in your legal place of residence, then you must have the consent of a parent\/legal guardian. Employees and directors of Microsoft Corporation and its subsidiaries, affiliates, advertising agencies, and Contest Parties are not eligible, nor are persons involved in the execution or administration of this promotion, or the family members of each above (parents, children, siblings, spouse\/domestic partners, or individuals residing in the same household). Void in Cuba, Iran, North Korea, Sudan, Syria, Region of Crimea, and where prohibited. For business\/tradeshow events: If you are attending the Event in your capacity as an employee, it is your sole responsibility to comply with your employer\u2019s gift policies. Microsoft will not be a party to any disputes or actions related to this matter. PLEASE NOTE: If you are a public sector employee (government and education), all prize awards will be awarded directly to your public sector organization and subject to receipt of a gift letter signed by your agency\/institution\u2019s ethics officer, attorney, or designated executive\/officer responsible for your organization\u2019s gifts\/ethics policy. Microsoft seeks to&nbsp;ensure that by offering items of value at no charge in promotional settings it does not create any violation of the letter or spirit of the entrant\u2019s applicable gifts and ethics rules.<\/p>\n\n\n\n<p>HOW TO ENTER<\/p>\n\n\n\n<p>The Contest Objective is to improve QoE for video conferencing system users as measured by objective audio\/video quality scores by developing a deep learning-based model (receiver-side bandwidth estimator, \u03c0) with offline RL techniques. To this end, participants should specify an appropriate reward function based on the provided dataset, the model architecture, and the training algorithm, given that the developed model adheres to the challenge requirements described in the homepage. Prizes will be awarded to the top 2 performing models based on the evaluation criteria and methodology described in the homepage. You may participate as an individual or a team. If forming a team, you must designate a \u201cTeam Captain\u201d who will submit all entry materials on behalf of the team. Once you register as part of a Team, you cannot change Teams or alter your current team (either by adding or removing members) after the submission of your Entry. Limit one Entry per person and per team. You may not compete&nbsp;on&nbsp;multiple&nbsp;teams,&nbsp;and you may not enter individually and on a team. We are not responsible for Entries that we do not receive for any reason, or for Entries that we receive but are not decipherable or not functional for any reason. Each Team is solely responsible for its own cooperation and teamwork. In no event will Microsoft officiate in any dispute regarding the conduct or cooperation of any Team or its members. The Contest will operate as follows:<\/p>\n\n\n\n<p><strong>Registration \/ Development Period:&nbsp;October 9, 2023 \u2013 January 5, 2024<\/strong><\/p>\n\n\n\n<p>There are two steps in registering for the challenge: <\/p>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Participants are required to fill <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/forms.office.com\/r\/LE5PR9EL5q\" target=\"_blank\" rel=\"noopener noreferrer\">this form<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> with the list of all the participants, affiliation of each participant (include country name), contact information for the team and name of your team. A confirmation email will be sent once we receive your registration information.<\/li>\n\n\n\n<li>Participants need to register on the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/mmsys24ods.hotcrp.com\/\" target=\"_blank\" rel=\"noopener noreferrer\">Open-source Software and Datasets submission system<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> and follow the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/2024.acmmmsys.org\/participation\/cfp\/\" target=\"_blank\" rel=\"noopener noreferrer\">submission guidelines<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> of the ODS track when submitting their work.<\/li>\n<\/ol>\n\n\n\n<p>Organizers will communicate updates regarding the grand challenge and announce the availability of data, and evaluation results etc. via email. Challenge leadership board and final results will be posted on the challenge website.<\/p>\n\n\n\n<p>After registration, participants will&nbsp;develop a bandwidth estimation model that best meets the contest objective and the challenge requirements as described in the homepage.<\/p>\n\n\n\n<p>The final evaluation will be conducted in a 2-stage process as described in the homepage during the judging period: January 27, 2024,&nbsp;12:00 AM AoE\u2013&nbsp;March 1, 2024, 11:59 PM AoE. <\/p>\n\n\n\n<p>To submit your entry, submit your model, code, and paper through the grand challenge conference submission system. You must satisfy all the challenge requirements as described on the homepage. ACM MMSys 2024&nbsp;grand challenge submissions are due on January 5, 2024,&nbsp;11:59 PM AoE. The entry limit is one per person during the Entry Period. Any attempt by any you to obtain more than the stated number of entries by using multiple\/different accounts, identities, registrations, logins, or any other methods will void your entries and you may be disqualified. Use of any automated system to participate is prohibited.<\/p>\n\n\n\n<p>We are not responsible for excess, lost, late, or incomplete entries. If disputed, entries will be deemed submitted by the \u201cauthorized account holder\u201d of the email address, social media account, or other method used to enter. The \u201cauthorized account holder\u201d is the natural person assigned to an email address by internet or online service provider, or other organization responsible for assigning email addresses.<\/p>\n\n\n\n<p>PAPER FORMAT<\/p>\n\n\n\n<p>Accepted challenge papers are published in the proceedings of the ACM MMSys 2024. Paper format should adhere to <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/www.acm.org\/publications\/proceedings-template\" target=\"_blank\" rel=\"noopener noreferrer\">ACM&#8217;s author guidelines<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>.<\/p>\n\n\n\n<p>ELIGIBLE ENTRY<\/p>\n\n\n\n<p>To be eligible, an entry must meet the following content\/technical requirements:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Your paper must be accepted by&nbsp;ACM MMSys 2024 Grand Challenge review process.<\/li>\n\n\n\n<li>Your entry must be your own original work; and<\/li>\n\n\n\n<li>Your entry cannot have been selected as a winner in any other contest; and<\/li>\n\n\n\n<li>You must have obtained&nbsp;any and all&nbsp;consents, approvals, or licenses required for you to submit your entry; and<\/li>\n\n\n\n<li>To the extent that entry requires the submission of user-generated content such as software, photos, videos, music, artwork, essays, etc., entrants warrant that their entry is their original work, has not been copied from others without permission or apparent rights, and does not violate the privacy, intellectual property rights, or other rights of any other person or entity. You may include Microsoft trademarks, logos, and designs, for which Microsoft grants you a limited license to use for the sole purposes of submitting an entry into this Contest; and<\/li>\n\n\n\n<li>Your entry may NOT contain, as determined by us in our sole and absolute discretion, any content that is obscene or offensive, violent, defamatory, disparaging or illegal, or that promotes alcohol, illegal drugs, tobacco or a particular political&nbsp;agenda, or that communicates messages that may reflect negatively on the goodwill of Microsoft.<\/li>\n<\/ul>\n\n\n\n<p>USE OF ENTRIES<\/p>\n\n\n\n<p>We are not claiming ownership rights to your Submission. However, by submitting an entry, you grant us an irrevocable, royalty-free, worldwide right and license to use, review, assess, test, and otherwise analyze your entry and all its content in connection with this Contest and use your entry in any media whatsoever now known or later invented for any non-commercial or commercial purpose, including, but not limited to, the marketing, sale or promotion of Microsoft products or services, without further permission from you. You will not receive any compensation or credit for use of your entry, other than what is described in these Official Rules.<\/p>\n\n\n\n<p>By entering you acknowledge that we may have developed or commissioned materials similar or identical to your entry and you waive any claims resulting from any similarities to your entry. Further, you understand that we will not restrict work assignments of representatives who have had access to your&nbsp;entry&nbsp;and you agree that the use of information in our representatives\u2019 unaided memories in the development or deployment of our products or services does not create liability for us under this agreement or copyright or trade secret law.<\/p>\n\n\n\n<p>Your entry may be posted on a public website. We are not responsible for any unauthorized use of your entry by visitors to this website. We are not obligated to use your entry for any purpose, even if it has been selected as a winning entry.<\/p>\n\n\n\n<p>WINNER SELECTION AND NOTIFICATION<\/p>\n\n\n\n<p>Pending confirmation of eligibility, potential prize winners will be selected by Microsoft or their Agent or a qualified judging panel from among all eligible entries received based on the following 2-stage judging process: <\/p>\n\n\n\n<ol class=\"wp-block-list\" start=\"1\">\n<li>In the first stage, <strong>all the submitted models<\/strong> will be evaluated in our emulation platform on a set of validation network traces with multiple repetitions per network trace. Evaluating the submitted models in our emulation platform, which is a controlled lab environment, enables us to estimate performance statistics and establish statistical significance between models.<\/li>\n\n\n\n<li>The <strong>top 3 models from stage 1<\/strong> will be evaluated in calls conducted between nodes in our testbed. Each model will be used to make several calls, and performance will be reported based on those calls.&nbsp;&nbsp;<\/li>\n<\/ol>\n\n\n\n<p>In either judging stage, the following scoring function will be used to rank the models:&nbsp;<\/p>\n\n\n\n<p class=\"has-text-align-center\"><strong><em>E<sub>calls<\/sub>[E<sub>n<\/sub>[objective video quality + objective audio quality]]<\/em><\/strong><\/p>\n\n\n\n<p>The outer expectation is across all calls and across repetitions per call, while the inner expectation is the temporal average of the objective metrics throughout a call.<\/p>\n\n\n\n<p>This challenge metric gives an equal weighting between objective audio and video quality.&nbsp;Among the submitted proposals, if the&nbsp;difference between&nbsp;overall&nbsp;evaluation metric between the models is not statistically significant, the model with a lower number of operations per second will be given a higher ranking.&nbsp;Winners will be notified after the end of the judging period on&nbsp;March 1, 2024.<\/p>\n\n\n\n<p>In the event of a tie between any eligible entries, an additional judge will break the tie based on the judging criteria described above. The decisions of the judges are final and binding. If we do not receive enough entries meeting the entry requirements, we may, at our discretion, select fewer winners. If public vote determines winners, it is prohibited for any person to obtain votes by any fraudulent or inappropriate means, including offering prizes or other inducements in exchange for votes, automated programs, or fraudulent IDs. Microsoft will void any questionable votes.<\/p>\n\n\n\n<p>PRIZES<\/p>\n\n\n\n<p>The following prizes will be awarded:<\/p>\n\n\n\n<p>First Prize: the winner will receive: $2000, paid in the form of a check from the sponsor or donation to the Individual Entrant or Team Captain\u2019s Educational institution. Approximate Retail Value (ARV) $2,000.00.<br>Second Prize: the runner-up will receive: $1000, paid in the form of a check from the sponsor or donation to the Individual Entrant or Team Captain\u2019s Educational Institution. Approximate Retail Value (ARV) $1,000.00.<\/p>\n\n\n\n<p>All prizes are stated in US dollars.<\/p>\n\n\n\n<p>The total Approximate Retail Value (ARV) of all prizes: $3,000.00.<\/p>\n\n\n\n<p><mark style=\"background-color:#ffb900\" class=\"has-inline-color\">PLEASE NOTE: If you are a public sector employee (government and education), your prize will be awarded directly to your public sector organization as a donation, and subject to receipt of a gift letter signed by your agency\/institution\u2019s ethics officer, attorney, or designated executive\/officer responsible for your organization\u2019s gifts\/ethics policy.<\/mark><\/p>\n\n\n\n<p>We will only award one prize per person. No more than the stated number of prizes will be awarded. No substitution, transfer, or assignment of prize permitted, except that Microsoft reserves the right to substitute a prize of equal or greater value in the event the offered prize is unavailable. Prizes are awarded \u201cAS IS\u201d with no warranty of any kind, either express or implied, including but not limited to, the implied warranties or merchantability, fitness for a particular purpose, or non-infringement. Prize winners may be required to complete and return prize claim and \/ or tax forms (\u201cForms\u201d) within the deadline stated in the winner notification. Taxes on the prize, if any, are the sole responsibility of the winner, who is advised to seek independent counsel regarding the tax implications of accepting a prize. By accepting a prize, you agree that Microsoft may use your entry, name, image and hometown online and in print, or in any other media, in connection with this Contest without payment or compensation to you, except where prohibited by law.<\/p>\n\n\n\n<p>ODDS<\/p>\n\n\n\n<p>The odds of winning are based on the number and quality of eligible entries received.<\/p>\n\n\n\n<p>GENERAL CONDITIONS AND RELEASE OF LIABILITY<\/p>\n\n\n\n<p>To the extent allowed by law, by entering you agree to release and hold harmless Microsoft and its respective parents, partners, subsidiaries, affiliates, employees, and agents from&nbsp;any and all&nbsp;liability or any injury, loss, or damage of any kind arising in connection with this&nbsp;Contest&nbsp;or any prize won.<\/p>\n\n\n\n<p>All local laws apply. The decisions of Microsoft are final and binding.<\/p>\n\n\n\n<p>We reserve the right to cancel, change, or suspend this Contest for any reason, including cheating, technology failure, catastrophe, war, or any other unforeseen or unexpected event that affects the integrity of this Contest, whether human or mechanical. If the integrity of the Contest cannot be restored, we may select winners from among all eligible entries received before we had to cancel, change or suspend the Contest.<\/p>\n\n\n\n<p>If you attempt or we have strong reason to believe that you have compromised the integrity or the legitimate operation of this Contest by cheating, hacking, creating a&nbsp;bot&nbsp;or other automated program, or by committing fraud in any way, we may seek damages from you to the full extent of the law and you may be banned from participation in future Microsoft promotions.<\/p>\n\n\n\n<p>GOVERNING LAW<\/p>\n\n\n\n<p>This Contest will be governed by the laws of the State of Washington, and you consent to the exclusive jurisdiction and venue of the courts of the State of Washington for any disputes arising out of this Contest.<\/p>\n\n\n\n<p>PRIVACY<\/p>\n\n\n\n<p>At Microsoft, we are committed to protecting your privacy. Microsoft uses the information you provide on this form to notify you of important information about our products, upgrades and enhancements, and to send you information about other Microsoft products and services. Microsoft will not share the information you provide with third parties without&nbsp;your permission except where necessary to complete the services or transactions you have requested, or as required by law. Microsoft is committed to protecting the security of your personal information. We use a variety of security technologies and procedures to help protect your personal information from unauthorized access, use, or disclosure. Your personal information is never shared outside the company without your permission, except under&nbsp;conditions&nbsp;explained above.<\/p>\n\n\n\n<p>If you believe that Microsoft has not adhered to this statement, please contact Microsoft by sending an email to\u202f<a href=\"mailto:privrc@microsoft.com\">privrc@microsoft.com<\/a>\u202for postal mail to Microsoft Privacy Response Center, Microsoft Corporation, One Microsoft Way, Redmond, WA.<\/p>\n\n\n\n\n\n<h2 class=\"wp-block-heading\" id=\"dataset-description\">Dataset Description<\/h2>\n\n\n\n<p>We will release a dataset of trajectories for Microsoft Teams audio\/video calls along with two objective signals that measure user-perceived audio\/video quality in audio\/video calls. The data is collected from audio\/video peer-to-peer Microsoft Teams calls conducted between testbed nodes that are geographically distributed across many countries and continents. Nodes are connected to the internet through various Internet Service Providers (ISPs) using either wired or wireless connections. Calls have been conducted with different bandwidth estimators, i.e., behavior policies, including traditional methods such as Kalman-filtering-based estimators and WebRTC (Web Real Time Communications), as well as different ML policies. <\/p>\n\n\n\n<p>Each trajectory corresponds to one audio\/video call leg and consists of a sequence of:<\/p>\n\n\n\n<ol class=\"wp-block-list\">\n<li>150-dimensional observation vector (<em>o<sub>n<\/sub><\/em>): computed based on packet information received by the client in one audio\/video call,<\/li>\n\n\n\n<li>Bandwidth estimates (<em>b<sub>n<\/sub><\/em>): predicted by a behavior policy<\/li>\n\n\n\n<li>Objective audio quality (r<sub>n<\/sub><sup>audio<\/sup>): objective received audio quality on a scale of [0,5], with a score of 5 being the highest.<\/li>\n\n\n\n<li>Objective video quality (r<sub>n<\/sub><sup>video<\/sup>): objective received video quality on a scale of [0,5], with a score of 5 being the highest.<\/li>\n<\/ol>\n\n\n\n<p>The observation vector at a time step <em>n<\/em> encapsulates observed network statistics that characterize the state of the bottleneck link between the sender and receiver over the <strong>5<\/strong> most recent short term monitor intervals (MI) of <em>60ms<\/em> and the <strong>5<\/strong> most recent long-term MIs of <em>600ms<\/em>. Specifically, the observation vector tracks <strong>15<\/strong> different network features over <strong>5<\/strong> short and <strong>5<\/strong> long term MIs<strong> (15 features x  (5 short term MIs + 5 long term MIs) = 150)<\/strong>. The 15 features and their description are as follows. Features are based on packets received during the short and long term monitor intervals.<\/p>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Receiving rate: rate at which the client receives data from the sender during a MI, unit: <em>bps.<\/em><\/li>\n\n\n\n<li>Number of received packets: total number of packets received in a MI, unit: <em>packet<\/em>.<\/li>\n\n\n\n<li>Received bytes: total number of bytes received in a MI, unit: <em>Bytes<\/em>.<\/li>\n\n\n\n<li>Queuing delay: average delay of packets received in a MI minus the minimum packet delay observed so far, unit: <em>ms<\/em>.<\/li>\n\n\n\n<li>Delay: average delay of packets received in a MI minus a fixed base delay of 200ms, unit: <em>ms<\/em>.<\/li>\n\n\n\n<li>Minimum seen delay: minimum packet delay observed so far, unit: <em>ms.<\/em><\/li>\n\n\n\n<li>Delay ratio: average delay of packets received in a MI divided by the minimum delay of packets received in the same MI, unit: <em>ms\/ms<\/em>.<\/li>\n\n\n\n<li>Delay average minimum difference: average delay of packets received in a MI minus the minimum delay of packets received in the same MI, unit: <em>ms<\/em>.<\/li>\n\n\n\n<li>Packet interarrival time: mean interarrival time of packets received in a MI, unit: <em>ms<\/em>.<\/li>\n\n\n\n<li>Packet jitter: standard deviation of interarrival time of packets received in a MI, unit: <em>ms<\/em>.<\/li>\n\n\n\n<li>Packet loss ratio: probability of packet loss in a MI, unit: <em>packet\/packet<\/em>.<\/li>\n\n\n\n<li>Average number of lost packets: average number of lost packets given a loss occurs, unit: <em>packet<\/em>.<\/li>\n\n\n\n<li>Video packets probability: proportion of video packets in the packets received in a MI, unit: <em>packet\/packet<\/em>.<\/li>\n\n\n\n<li>Audio packets probability: proportion of audio packets in the packets received in a MI, unit: <em>packet\/packet<\/em>.<\/li>\n\n\n\n<li>Probing packets probability: proportion of probing packets in the packets received in a MI, unit: <em>packet\/packet<\/em>.<\/li>\n<\/ol>\n\n\n\n<p>&nbsp;The indices (zero-indexed) of features over the 5 short term MIs are <strong>{(feature number &#8211; 1)*10, ., ., ., feature number * 10 &#8211; 5 &#8211; 1}<\/strong>.<\/p>\n\n\n\n<p>The indices (zero-indexed) of features over the 5 long term MIs are <strong>{feature number * 10 &#8211; 5, ., ., ., feature number * 10 &#8211; 1}<\/strong>.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"training-and-validation-datasets\">Training and Validation Datasets&nbsp;<\/h3>\n\n\n\n<p>We are going to release data from <strong>18859<\/strong> calls conducted between testbed nodes to be used as training and validation data by the participants. Participants are free to split the data into train\/validation sets as deemed necessary.&nbsp;<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"emulated-dataset\">Emulated Dataset&nbsp;<\/h3>\n\n\n\n<p>We will also release a dataset from <strong>9405<\/strong> emulated test calls which contains ground truth information about the bottleneck link between the sender and receiver, namely, bottleneck capacity and loss rate, in addition to the aforementioned data. In this dataset, the characteristics of the bottleneck, namely ground truth capacity and loss rate, are randomly varied throughout the duration of the test call to generate a diverse set of trajectories with network dynamics that may not occur in the real world but are nevertheless important to enhance state-action space coverage and aid in learning generalizable policies.<\/p>\n\n\n\n<p>Participants are free to use this dataset in conjunction with the real-world testbed dataset to train the policy, given that ground truth information <strong>is not provided as input to the model and the true environment remains partially observable (only through the observation vector).<\/strong>&nbsp;<strong>Any model which has ground truth information in the inputs will be disqualified from the contest.<\/strong><\/p>\n\n\n\n\n\n<p><em>Time zone for below dates is Anywhere on Earth<\/em>&nbsp;(AoE)<\/p>\n\n\n\n<figure class=\"wp-block-table is-style-stripes\"><table><thead><tr><th>Milestone<\/th><th>Date<\/th><\/tr><\/thead><tbody><tr><td>Challenge announcement<\/td><td>October 9, 2023<\/td><\/tr><tr><td>Dataset release<\/td><td>October 20, 2023<\/td><\/tr><tr><td>Registration on the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/mmsys24ods.hotcrp.com\/\">hotcrp submission website<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/td><td>January 26, 2024<\/td><\/tr><tr><td>Model, code, and paper submission deadline<\/td><td>February 2nd, 2024<\/td><\/tr><tr><td>Grand challenge paper acceptance notification<\/td><td>March 1, 2024<\/td><\/tr><tr><td>Camera-ready paper due<\/td><td>March 8, 2024<\/td><\/tr><\/tbody><\/table><\/figure>\n\n\n\n<p><\/p>\n\n\n\n\n\n<h2 class=\"wp-block-heading\" id=\"organizers\">Organizers<\/h2>\n\n\n\n<ul class=\"wp-block-list\">\n<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/samikhairy\/\">Sami Khairy<\/a>, Microsoft, Canada<\/li>\n\n\n\n<li>Gabriel Mittag, Microsoft, USA<\/li>\n\n\n\n<li>Ezra Ameri, Microsoft, Canada<\/li>\n\n\n\n<li>Scott Inglis, Microsoft, USA<\/li>\n\n\n\n<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/vigopal\/\">Vishak Gopal<\/a>, Microsoft, USA<\/li>\n\n\n\n<li>Mehrsa Golestaneh, Microsoft, Canada<\/li>\n\n\n\n<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/rcutler\/\">Ross Cutler<\/a>, Microsoft, USA<\/li>\n\n\n\n<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/francisy\/\">Francis Yan<\/a>, Microsoft Research, USA<\/li>\n\n\n\n<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/zhniu\/\">Zhixiong Niu<\/a>, Microsoft Research, China<\/li>\n<\/ul>\n\n\n\n\n\n<h3 class=\"wp-block-heading\" id=\"objective-metrics\">Objective Metrics<\/h3>\n\n\n\n<p>In our audio\/video test calls, video transmission starts few seconds after the beginning of the call. During this initial period, only audio packets and probing packets flow from the sender to the receiver. Consequently, video quality is not defined during this period, which is the reason why the provided objective video quality signal for each test call begins with a sequence of nan values. Another point to keep in mind is that the flow rate of audio packets in this initial period is typically much less than the capacity of the bottleneck link. During this period, different behavior policies adopt different ramp up strategies to estimate the available bandwidth.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"asymmetric-rl\">Asymmetric RL&nbsp;<\/h3>\n\n\n\n<p>In asymmetric RL, the agent has access to extra information during training that is not available at test\/deployment time. Because the emulated dataset contains ground truth information (loss rate, capacity) for the bottleneck link, ideas from asymmetric RL and optimization objectives with auxiliary losses can be leveraged to train the policy. Furthermore, ground truth information can be helpful to design and train representation learning architectures to facilitate bandwidth estimation.&nbsp;That said, ground truth information <strong><em>cannot be used<\/em><\/strong> as additional inputs to the model since in practice, this information is unknown. <strong>Any model which has ground truth information in the inputs will be disqualified from the contest.<\/strong><\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"onnx-model-input-output-shapes\">Onnx model Input\/Output Shapes<\/h2>\n\n\n\n<p>As per the challenge requirements, submitted ONNX models should have the following input shapes:<\/p>\n\n\n\n<ol class=\"wp-block-list\" start=\"1\">\n<li><strong>obs:<\/strong> <strong>float32[1,1,150]<\/strong><\/li>\n\n\n\n<li><strong>hidden_states:<\/strong> <strong>float32[1,N], N can be any positive integer, e.g. float32[1,128] or float32[1,1]<\/strong><\/li>\n\n\n\n<li><strong>cell_states:<\/strong> <strong>float32[1,N], N is the same as for the input hidden states dimension<\/strong><\/li>\n<\/ol>\n\n\n\n<p>Even if the model does not make use of the hidden\/cell states, these inputs should still be passed as inputs to the model and returned as outputs. On the other hand, the outputs should have the following shapes:<\/p>\n\n\n\n<ol class=\"wp-block-list\" start=\"1\">\n<li><strong>output: float32[1,1,2]<\/strong><\/li>\n\n\n\n<li><strong>state_out: float32[1,N], N is the same as for the input hidden states<\/strong> <strong>dimension<\/strong><\/li>\n\n\n\n<li><strong>cell_out:<\/strong> <strong>float32[1,N], N is the same as for the input cell states dimension<\/strong><\/li>\n<\/ol>\n\n\n\n<p>Importantly, the output should be the <strong>mean<\/strong> predicted bandwidth and the <strong>standard deviation<\/strong> of the prediction, <strong>hence the shape is [1x1x2]<\/strong>. If the training method does not learn the standard deviation by design, a constant value can be appended to conform to the required shape ([1x1x2]). When the model is evaluated online, <strong>only the mean predicted bandwidth in bps<\/strong> is used and the standard deviation is ignored. Furthermore, dynamic axes for the batch and sequence length dimensions are not supported. Hence, <strong>batch_size = 1 and seq_len = 1<\/strong>. Input\/output shapes of ONNX models can be inspected using a neural network visualizer like <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/netron.app\/\">Netron<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> which for the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/github.com\/microsoft\/RL4BandwidthEstimationChallenge\/tree\/main\/onnx_models\">released baseline model<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> look like the following:<\/p>\n\n\n\n<figure class=\"wp-block-image aligncenter size-large is-resized\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"889\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2023\/12\/input_output_shapes-65863a95ddb6a-1024x889.png\" alt=\"Onnx Model Input\/Output Shapes\" class=\"wp-image-995235\" style=\"width:570px;height:auto\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2023\/12\/input_output_shapes-65863a95ddb6a-1024x889.png 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2023\/12\/input_output_shapes-65863a95ddb6a-300x260.png 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2023\/12\/input_output_shapes-65863a95ddb6a-768x666.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2023\/12\/input_output_shapes-65863a95ddb6a-207x180.png 207w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2023\/12\/input_output_shapes-65863a95ddb6a.png 1445w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure>\n\n\n\n<p><\/p>\n\n\n\n<p><\/p>\n\n\n\n\n\n<h3 class=\"wp-block-heading\" id=\"related-links\">Related Links<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/forms.office.com\/r\/LE5PR9EL5q\">Registration form<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n\n\n\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/github.com\/microsoft\/RL4BandwidthEstimationChallenge\/blob\/main\/download-testbed-dataset.sh\">Testbed dataset<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n\n\n\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/github.com\/microsoft\/RL4BandwidthEstimationChallenge\/blob\/main\/download-emulated-dataset.sh\">Emulated dataset<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n\n\n\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/github.com\/microsoft\/RL4BandwidthEstimationChallenge\">GitHub repository<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n\n\n\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/github.com\/microsoft\/RL4BandwidthEstimationChallenge\/blob\/main\/tf_policy.py\">Custom Tensorflow model class and ONNX conversion script<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n\n\n\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/github.com\/microsoft\/RL4BandwidthEstimationChallenge\/blob\/main\/torch_policy.py\">Custom Pytorch model class and ONNX conversion script<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n\n\n\n<li>Offline RL <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/github.com\/microsoft\/RL4BandwidthEstimationChallenge\/tree\/main\/onnx_models\">baseline model<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> and <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/github.com\/microsoft\/RL4BandwidthEstimationChallenge\/blob\/main\/run_baseline_model.py\">inference script<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n\n\n\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/ieeexplore.ieee.org\/document\/10095711\">ML model used to predict objective video quality<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"previous-challenge\">Previous challenge&nbsp;<\/h3>\n\n\n\n<p>In the <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/2021.acmmmsys.org\/rtc_challenge.php\" target=\"_blank\" rel=\"noopener noreferrer\">previous bandwidth estimation challenge<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> that was hosted at ACM MMSys 2021, participants were provided with a \u201cgym\u201d simulation environment based on network simulator 3 (NS-3) and the challenge focused on learning a bandwidth estimator using online reinforcement learning (RL). Policies trained in simulation may not be optimal when deployed in the real world because of many challenges, including, sim-to-real gap, and the misalignment between rewards computed based on network statistics and actual user perceived quality of experience. This challenge fills this gap by providing signals that measure the quality of the received audio\/video streams and focusing on training the bandwidth estimator using traces from audio\/video calls conducted over the internet. Nevertheless, participants are encouraged to leverage the resources from the previous challenge and to test their proposed models on the simulation environment to ensure generalization.&nbsp;<\/p>\n\n\n\n\n\n<h2 class=\"wp-block-heading\" id=\"first-preliminary-evaluation-stage\">First preliminary evaluation stage<\/h2>\n\n\n\n<p>To support the research efforts of challenge participants and enhance the overall quality of submissions ahead of the final deadline, we offer an <strong>optional<\/strong> <strong>small-scale preliminary evaluation opportunity<\/strong> for all participating teams. This will provide participants with the chance to submit <strong>up to three models per registered team<\/strong> for online testing. Each submitted model will be evaluated in our emulation platform by conducting <strong>24 peer-2-peer test calls<\/strong> <strong>with 8 different network traces<\/strong> <strong>and the average objective audio\/video quality scores for these models will be posted in a leaderboard on the challenge website<\/strong>. This initiative aims to assist participants with refining their designs, identifying potential flaws early in the process, and ultimately enhancing the robustness of the solutions.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"instructions\">Instructions<\/h3>\n\n\n\n<p>Registered teams who wish to leverage the preliminary evaluation opportunity should send an <a href=\"mailto:samikhairy@microsoft.com\">email<\/a> to Sami Khairy with:<\/p>\n\n\n\n<ol class=\"wp-block-list\">\n<li>a zip file containing <strong>up to three ONNX models<\/strong>. Each ONNX model should adhere to the <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/academic-program\/bandwidth-estimation-challenge\/#:~:text=the%20below%20requirements.-,Challenge%20Requirements,-Failing%20to%20adhere\">challenge requirements<\/a>. Specifically, the inputs and outputs for the model should be consistent with the provided <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/github.com\/microsoft\/RL4BandwidthEstimationChallenge\/blob\/main\/tf_policy.py\" target=\"_blank\" rel=\"noopener noreferrer\">TF model class<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> or <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/github.com\/microsoft\/RL4BandwidthEstimationChallenge\/blob\/main\/torch_policy.py\" target=\"_blank\" rel=\"noopener noreferrer\">PyTorch model<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> and ONNX conversion therein. We have released a&nbsp;<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/github.com\/microsoft\/RL4BandwidthEstimationChallenge\/blob\/main\/onnx_models\/\" target=\"_blank\" rel=\"noopener noreferrer\">baseline stateless model (MLP)<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>&nbsp;trained with offline RL on the emulated dataset as a reference, with an&nbsp;<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/github.com\/microsoft\/RL4BandwidthEstimationChallenge\/blob\/main\/run_baseline_model.py\">example script<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>&nbsp;to run this model on the offline data. The preliminary evaluation results for this baseline model are available in the leaderboard <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/github.com\/microsoft\/RL4BandwidthEstimationChallenge\/blob\/main\/run_baseline_model.py\" target=\"_blank\" rel=\"noopener noreferrer\">here<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>.<\/li>\n\n\n\n<li>The name of each ONNX model should be prefixed with the registered team name.<\/li>\n\n\n\n<li>Please include &#8220;<strong>TEAMNAME_PRELIMINARY_EVALUATION<\/strong>&#8221; in the subject line.<\/li>\n\n\n\n<li>The deadline to submit the models via email is <strong>December 6, 2023, 11:59 PM AoE<\/strong>. <strong>Submissions after the deadline will not be accepted and the models will not be evaluated.<\/strong><\/li>\n\n\n\n<li><strong>Each team has the chance to send exactly one email with up to three models in a zip file. Subsequent emails will be discarded<\/strong>,<strong> so please make sure to include the correct models that you wish to be evaluated.<\/strong><\/li>\n\n\n\n<li>Participation in the preliminary evaluation stage is completely optional but is highly recommended.<\/li>\n\n\n\n<li>We aim to conclude the preliminary evaluation and announce the results by December 15<sup>th<\/sup>, 2023.<\/li>\n<\/ol>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"announcement\">Announcement<\/h3>\n\n\n\n<p>[December 6th, 2023]: preliminary evaluation submission deadline extended to <strong>December 8, 2023, 11:59 PM AoE<\/strong>. torch_policy.py&nbsp;<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/github.com\/microsoft\/RL4BandwidthEstimationChallenge\/commit\/7258bda0e87d2f7f1ad8a8380806e45bb88c29d7\" target=\"_blank\" rel=\"noopener noreferrer\">bug fixed<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"second-preliminary-evaluation-stage\">Second preliminary evaluation stage<\/h2>\n\n\n\n<p>In order to further support the research endeavors of participating teams in the challenge, we are pleased to announce a second optional  small-scale preliminary evaluation opportunity. The details and operational aspects of this second preliminary evaluation stage mirror those of the first stage, with the following exceptions:<\/p>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Each registered team can submit up to <strong>two models<\/strong> for online testing.<\/li>\n\n\n\n<li>The deadline for model submissions via email is <strong>January 5, 2024<\/strong>, <strong>11:59 PM AoE<\/strong>. This is a hard deadline and there will not be any extension. Submissions received after this deadline will not be accepted, and the models will not undergo evaluation.<\/li>\n\n\n\n<li>Our aim is to conclude the preliminary evaluation process and announce the results by <strong>January 15, 2024<\/strong>.<\/li>\n<\/ol>\n\n\n\n\n\n<h2 class=\"wp-block-heading\" id=\"first-preliminary-evaluation-stage-1\">First Preliminary Evaluation Stage<\/h2>\n\n\n\n<figure class=\"wp-block-image aligncenter size-full is-resized\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"598\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/03\/first-prelim-stage.png\" alt=\"table\" class=\"wp-image-1014039\" style=\"width:530px;height:auto\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/03\/first-prelim-stage.png 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/03\/first-prelim-stage-300x175.png 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/03\/first-prelim-stage-768x449.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/03\/first-prelim-stage-480x280.png 480w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/03\/first-prelim-stage-240x140.png 240w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure>\n\n\n\n<p>Leaderboard for participating models in the first preliminary evaluation stage<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"second-preliminary-evaluation-stage-1\">Second Preliminary Evaluation Stage<\/h2>\n\n\n\n<figure class=\"wp-block-image aligncenter size-large is-resized\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"465\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/01\/Ranking-1024x465.png\" alt=\"table\" class=\"wp-image-998841\" style=\"width:556px;height:auto\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/01\/Ranking-1024x465.png 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/01\/Ranking-300x136.png 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/01\/Ranking-768x349.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/01\/Ranking-1536x698.png 1536w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/01\/Ranking-2048x930.png 2048w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/01\/Ranking-240x109.png 240w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure>\n\n\n\n<p>Leaderboard for participating models in the second preliminary evaluation stage<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"first-evaluation-stage\">First Evaluation Stage<\/h2>\n\n\n\n<figure class=\"wp-block-image aligncenter size-large is-resized\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"503\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/03\/FirstStage-1024x503.png\" alt=\"table\" class=\"wp-image-1014003\" style=\"width:543px;height:auto\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/03\/FirstStage-1024x503.png 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/03\/FirstStage-300x147.png 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/03\/FirstStage-768x377.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/03\/FirstStage-1536x754.png 1536w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/03\/FirstStage-240x118.png 240w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/03\/FirstStage.png 2022w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure>\n\n\n\n<p>First evaluation stage rankings: Schaferct, Paramecium, SJTU Medialab, and Fast and furious models were all advanced to the final evaluation stage on the testbed.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"final-evaluation-stage\">Final Evaluation Stage<\/h2>\n\n\n\n<figure class=\"wp-block-image aligncenter size-large is-resized\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"338\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/03\/FinalStage-65ef9636c3ade-1024x338.png\" alt=\"table\" class=\"wp-image-1014015\" style=\"width:523px;height:auto\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/03\/FinalStage-65ef9636c3ade-1024x338.png 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/03\/FinalStage-65ef9636c3ade-300x99.png 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/03\/FinalStage-65ef9636c3ade-768x254.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/03\/FinalStage-65ef9636c3ade-1536x508.png 1536w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/03\/FinalStage-65ef9636c3ade-240x79.png 240w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2024\/03\/FinalStage-65ef9636c3ade.png 1940w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure>\n\n\n\n<p><strong>Final evaluation stage rankings: <\/strong><\/p>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Winning team: <strong>Schaferct<\/strong>.<\/li>\n\n\n\n<li>Runner up: <strong>Fast and furious<\/strong>.<\/li>\n<\/ol>\n\n\n\n<p><\/p>\n\n\n","protected":false},"featured_media":0,"template":"","meta":{"msr-url-field":"","msr-podcast-episode":"","msrModifiedDate":"","msrModifiedDateEnabled":false,"ep_exclude_from_search":false,"_classifai_error":"","msr_hide_image_in_river":0,"footnotes":""},"msr-opportunity-type":[187426],"msr-region":[256048],"msr-locale":[268875],"msr-program-audience":[],"msr-post-option":[],"msr-impact-theme":[264846],"class_list":["post-969735","msr-academic-program","type-msr-academic-program","status-publish","hentry","msr-opportunity-type-challenges","msr-region-global","msr-locale-en_us"],"msr_description":"","msr_social_media":[],"related-researchers":[{"type":"user_nicename","display_name":"Sami Khairy","user_id":42906,"people_section":"Section name 0","alias":"samikhairy"},{"type":"guest","display_name":"Gabriel Mittag","user_id":972807,"people_section":"Section name 0","alias":""},{"type":"guest","display_name":"Ezra Ameri","user_id":972801,"people_section":"Section name 0","alias":""},{"type":"guest","display_name":"Scott Inglis","user_id":972813,"people_section":"Section name 0","alias":""},{"type":"user_nicename","display_name":"Vishak Gopal","user_id":39624,"people_section":"Section name 0","alias":"vigopal"},{"type":"guest","display_name":"Mehrsa Golestaneh","user_id":972816,"people_section":"Section name 0","alias":""},{"type":"user_nicename","display_name":"Ross Cutler","user_id":40660,"people_section":"Section name 0","alias":"rcutler"},{"type":"user_nicename","display_name":"Zhixiong Niu","user_id":38118,"people_section":"Section name 0","alias":"zhniu"}],"tab-content":[],"msr_impact_theme":["Computing foundations"],"_links":{"self":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-academic-program\/969735","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-academic-program"}],"about":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/types\/msr-academic-program"}],"version-history":[{"count":149,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-academic-program\/969735\/revisions"}],"predecessor-version":[{"id":1016892,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-academic-program\/969735\/revisions\/1016892"}],"wp:attachment":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media?parent=969735"}],"wp:term":[{"taxonomy":"msr-opportunity-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-opportunity-type?post=969735"},{"taxonomy":"msr-region","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-region?post=969735"},{"taxonomy":"msr-locale","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-locale?post=969735"},{"taxonomy":"msr-program-audience","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-program-audience?post=969735"},{"taxonomy":"msr-post-option","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-post-option?post=969735"},{"taxonomy":"msr-impact-theme","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-impact-theme?post=969735"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}