{"id":717037,"date":"2021-01-14T10:24:02","date_gmt":"2021-01-14T18:24:02","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/?post_type=msr-academic-program&#038;p=717037"},"modified":"2024-11-21T17:43:51","modified_gmt":"2024-11-22T01:43:51","slug":"deep-noise-suppression-challenge-interspeech-2021","status":"publish","type":"msr-academic-program","link":"https:\/\/www.microsoft.com\/en-us\/research\/academic-program\/deep-noise-suppression-challenge-interspeech-2021\/","title":{"rendered":"Deep Noise Suppression Challenge \u2013 INTERSPEECH 2021"},"content":{"rendered":"\n\n<p><\/p>\n\n\n\n\n\n\n<p><strong>Program dates:<\/strong> January 2020 &#8211; March 2021<\/p>\n<p>The Deep Noise Suppression (DNS) challenge is designed to foster innovation in the area of noise suppression to achieve superior perceptual speech quality. We recently organized a DNS challenge special session at INTERSPEECH 2020 and ICASSP 2020. We open-sourced training and test datasets for the wideband scenario. We also open-sourced a subjective evaluation framework based on ITU-T standard P.808, which was used to evaluate challenge submissions. Many researchers from academia and industry made significant contributions to push the field forward, yet even the best noise suppressor was far from achieving superior speech quality in challenging scenarios. In this version of the challenge organized at INTERSPEECH 2021, we are expanding both our training and test datasets to accommodate full band scenarios. The two tracks in this challenge will focus on real-time denoising for (i) wide band, and (ii) full band scenarios. We are also making available a reliable non-intrusive objective speech quality metric for wide band called DNSMOS for the participants to use during their development phase.<\/p>\n<p><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/interspeech-2021-deep-noise-suppression-challenge\/\" target=\"_blank\" rel=\"noopener\">Challenge description (PDF)<\/a><\/p>\n<p>We will have two tracks in this challenge:<\/p>\n<ul>\n<li><strong>Track 1: Real-Time Denoising track for wide band scenario<\/strong><br \/>\nThe noise suppressor must take less than the stride time Ts (in ms) to process a frame of size T (in ms) on an Intel Core i5 quad-core machine clocked at 2.4 GHz or equivalent processor. For example, Ts = T\/2 for 50% overlap between frames. The total algorithmic latency allowed including the frame size T, stride time Ts, and any look ahead must be less than or equal to 40ms. For example, for a real-time system that receives 20ms audio chunks, if you use a frame length of 20ms with a stride of 10ms resulting in an algorithmic latency of 30ms, then you satisfy the latency requirements. If you use a frame of size 32ms with a stride of 16ms resulting in an algorithmic latency of 48ms, then your method does not satisfy the latency requirements as the total algorithmic latency exceeds 40ms. If your frame size plus stride T1=T+Ts is less than 40ms, then you can use up to (40-T1) ms future information.<\/li>\n<li><strong>Track 2: Real-Time Denoising track for full band scenario<\/strong><br \/>\nSatisfy Track 1 requirements but at 48 kHz.<\/li>\n<\/ul>\n<p><strong>Participants are forbidden from using the blind test set to retrain or tweak their models. Participants must submit results only if they intend to submit a paper to INTERSPEECH 2021. Failing to adhere to these rules will lead to disqualification from the challenge.<\/strong><\/p>\n<h3>Registration<\/h3>\n<p>Please send an email to <a href=\"mailto:dns_challenge@microsoft.com?Subject=Challenge%20Query\" target=\"_blank\" rel=\"noopener\">dns_challenge@microsoft.com<\/a> stating that you are interested to participate in the challenge. Please include the following details in your email:<\/p>\n<ul>\n<li>List of participants<\/li>\n<li>Affiliation of each participant<\/li>\n<li>Email ID of each participant<\/li>\n<\/ul>\n<p>Also, please create a new submission at <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/cmt3.research.microsoft.com\/3rdDNSChallenge\" target=\"_blank\" rel=\"noopener noreferrer\">https:\/\/cmt3.research.microsoft.com\/3rdDNSChallenge<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> and fill out all the details. This will help us to easily send out any announcements.<\/p>\n<p><strong>Contact us:<\/strong> If you have questions about this program, email us at <a href=\"mailto:dns_challenge@microsoft.com\">dns_challenge@microsoft.com<\/a>.<\/p>\n\n\n\n\n\n<h2>Official rules<\/h2>\n<p>SPONSOR<br \/>\nThese Official Rules (\u201cRules\u201d) govern the operation of the Microsoft M365 Core INTERSPEECH 2021 Deep Noise Suppression Event Contest (\u201cContest\u201d). Microsoft Corporation, One Microsoft Way, Redmond, WA, 98052, USA, is the Contest sponsor (\u201cSponsor\u201d).<\/p>\n<p>DEFINITIONS<br \/>\nIn these Rules, &#8220;Microsoft&#8221;, &#8220;we&#8221;, &#8220;our&#8221;, and &#8220;us&#8221;, refer to Sponsor, and &#8220;you&#8221; and &#8220;yourself&#8221; refers to a Contest participant or the parent\/legal guardian of any Contest entrant who has not reached the age of majority to contractually obligate themselves in their legal place of residence. &#8220;Event&#8221; refers to the INTERSPEECH 2021 Deep Noise Suppression event held in Brno, Czechia (the &#8220;Event&#8221;). By entering you (your parent\/legal guardian if you are not the age of majority in your legal place of residence) agree to be bound by these Rules.<\/p>\n<p>ENTRY PERIOD<br \/>\nThe Contest will operate from January 8, 2021 to March 8, 2021 (&#8220;Entry Period&#8221;). The Entry Period is divided into several periods as described in Section 5 How to Enter.<\/p>\n<p>ELIGIBILITY<br \/>\nOpen to any registered Event attendee 18 years of age or older. If you are 18 years of age or older but have not reached the age of majority in your legal place of residence, then you must have the consent of a parent\/legal guardian. Employees and directors of Microsoft Corporation and its subsidiaries, affiliates, advertising agencies, and Contest Parties are not eligible, nor are persons involved in the execution or administration of this promotion, or the family members of each above (parents, children, siblings, spouse\/domestic partners, or individuals residing in the same household). Void in Cuba, Iran, North Korea, Sudan, Syria, Region of Crimea, and where prohibited. For business\/tradeshow events: If you are attending the Event in your capacity as an employee, it is your sole responsibility to comply with your employer&#8217;s gift policies. Microsoft will not be a party to any disputes or actions related to this matter. PLEASE NOTE: If you are a public sector employee (government and education), all prize awards will be awarded directly to your public sector organization and subject to receipt of a gift letter signed by your agency\/institution&#8217;s ethics officer, attorney, or designated executive\/officer responsible for your organization&#8217;s gifts\/ethics policy. Microsoft seeks to ensure that by offering items of value at no charge in promotional settings it does not create any violation of the letter or spirit of the entrant&#8217;s applicable gifts and ethics rules.<\/p>\n<p>HOW TO ENTER<br \/>\nThe Contest Objective is to promote collaborative research in real-time single-channel Speech Enhancement aimed to maximize the subjective (perceptual) quality of the enhanced speech. Prizes will be awarded based on the speech quality of deep noise suppression models using the online subjective evaluation framework ITU-T P.835. Only methods described in accepted INTERSPEECH 2021 papers will be eligible for the contest. You may participate as an individual or a team. If forming a team, you must designate a &#8220;Team Captain&#8221; who will submit all entry materials on behalf of the team. Once you register as part of a Team, you cannot change Teams or alter your current team (either by adding or removing members) after the submission of your Entry. Limit one Entry per person and per team. You may not compete on multiple teams and you may not enter individually and on a team. We are not responsible for Entries that we do not receive for any reason, or for Entries that we receive but are not decipherable or not functional for any reason. Each Team is solely responsible for its own cooperation and teamwork. In no event will Microsoft officiate in any dispute regarding the conduct or cooperation of any Team or its members. The Contest will operate as follows:<\/p>\n<p>Registration \/ Development Period: January 8, 2021 &#8211; March 8, 2021<br \/>\nTo register, please send an email to <a href=\"mailto:dns_challenge@microsoft.com\">dns_challenge@microsoft.com<\/a> stating that you are interested to participate in the challenge. Please include the following details in your email:<\/p>\n<ol>\n<li>Names of the participants and name of the team captain<\/li>\n<li>Institution\/Company<\/li>\n<li>Email<\/li>\n<\/ol>\n<p>Create a submission by registering at <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/cmt3.research.microsoft.com\/3rdDNSChallenge\/\" target=\"_blank\" rel=\"noopener noreferrer\">Conference Management Toolkit &#8211; 3rd DNS Challenge<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> and fill in all your details. You will be using this tool for final submission and to receive any email announcements from organizers.<\/p>\n<p>Then i. develop a speech enhancement model that best meets the Contest Objective as described in the base paper and ii. submit a paper to INTERSPEECH 2021 which reports the computational complexity of the model in terms of the number of parameters and the time it takes to infer a frame on a particular CPU (preferably Intel Core i5 quad core machine clocked at 2.4 GHz). To develop your model, use any publicly available clean speech and noise datasets, including the contest datasets provided for training and developing models. You may augment your datasets with the contest dataset. You may mix clean speech and noise in any way that improves the performance of your model.<\/p>\n<p>The final evaluation will be conducted on a blind test set that is similar to the open sourced development stage test set. You may use scripts for a baseline noise suppressor that was recently published here. Testing \/ Entry Period: January 8, 2021 &#8211; March 8, 2021. On March 8, the blind test dataset will be released. You will have until 11:59 PM PT on March 15, 2021 to test your model against this dataset and create a set of enhanced clips to submit for judging (your &#8220;Entry&#8221;) via Conference Management Tool.<\/p>\n<p>You may not use the blind test set to retrain or tweak your model. To submit your entry, submit your processed clips via conference management tool. Each Entry will fall in one of two tracks based on the sampling rate. You must satisfy all the requirements of each track in terms of algorithmic latency. You must also specify the Number of operations per second in your paper submission. INTERSPEECH 2021 Paper Submission and Judging Period: March 15 &#8211; 11:59 PM PT June 2, 2021. Your Entry must be described in a paper accepted by INTERSPEECH 2021. To submit a paper, visit <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/www.interspeech2021.org\/call-for-special-sessions-and-challenges\" target=\"_blank\" rel=\"noopener noreferrer\">Call for Special Sessions & Challenges | INTERSPEECH 2021<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> and follow instructions to submit a paper.<\/p>\n<p>The entry limit is one per person during the Entry Period. Any attempt by any you to obtain more than the stated number of entries by using multiple\/different accounts, identities, registrations, logins, or any other methods will void your entries and you may be disqualified. Use of any automated system to participate is prohibited.<\/p>\n<p>We are not responsible for excess, lost, late, or incomplete entries. If disputed, entries will be deemed submitted by the &#8220;authorized account holder&#8221; of the email address, social media account, or other method used to enter. The &#8220;authorized account holder&#8221; is the natural person assigned to an email address by internet or online service provider, or other organization responsible for assigning email addresses.<\/p>\n<p>ELIGIBLE ENTRY<br \/>\nTo be eligible, an entry must meet the following content\/technical requirements:<\/p>\n<ul>\n<li>Your Entry must be the method described in a paper accepted by INTERSPEECH 2021<\/li>\n<li>Your entry must be your own original work; and<\/li>\n<li>Your entry cannot have been selected as a winner in any other contest; and<\/li>\n<li>You must have obtained any and all consents, approvals, or licenses required for you to submit your entry; and<\/li>\n<li>To the extent that entry requires the submission of user-generated content such as software, photos, videos, music, artwork, essays, etc., entrants warrant that their entry is their original work, has not been copied from others without permission or apparent rights, and does not violate the privacy, intellectual property rights, or other rights of any other person or entity. You may include Microsoft trademarks, logos, and designs, for which Microsoft grants you a limited license to use for the sole purposes of submitting an entry into this Contest; and<\/li>\n<li>Your entry may NOT contain, as determined by us in our sole and absolute discretion, any content that is obscene or offensive, violent, defamatory, disparaging or illegal, or that promotes alcohol, illegal drugs, tobacco or a particular political agenda, or that communicates messages that may reflect negatively on the goodwill of Microsoft.<\/li>\n<li>Your entry must NOT include enhanced clips using other noise suppression methods that you are not submitting to INTERSPEECH 2021.<\/li>\n<\/ul>\n<p>USE OF ENTRIES<br \/>\nWe are not claiming ownership rights to your Submission. However, by submitting an entry, you grant us an irrevocable, royalty-free, worldwide right and license to use, review, assess, test, and otherwise analyze your entry and all its content in connection with this Contest and use your entry in any media whatsoever now known or later invented for any non-commercial or commercial purpose, including, but not limited to, the marketing, sale or promotion of Microsoft products or services, without further permission from you. You will not receive any compensation or credit for use of your entry, other than what is described in these Official Rules.<\/p>\n<p>By entering you acknowledge that we may have developed or commissioned materials similar or identical to your entry and you waive any claims resulting from any similarities to your entry. Further, you understand that we will not restrict work assignments of representatives who have had access to your entry and you agree that the use of information in our representatives&#8217; unaided memories in the development or deployment of our products or services does not create liability for us under this agreement or copyright or trade secret law.<\/p>\n<p>Your entry may be posted on a public website. We are not responsible for any unauthorized use of your entry by visitors to this website. We are not obligated to use your entry for any purpose, even if it has been selected as a winning entry.<\/p>\n<p>WINNER SELECTION AND NOTIFICATION<br \/>\nPending confirmation of eligibility, potential prize winners will be selected by Microsoft or their Agent or a qualified judging panel from among all eligible entries received based on the following judging criteria: 99% &#8211; The subjective speech quality evaluated on the blind test set using ITU-T P.835 framework. We will use the submitted clips with no alteration to conduct ITU-T P.835 subjective evaluation and pick the winners based on the results. Among the submitted proposals, if the difference between the models is not statistically significant, the model with a lower number of operations per second be given a higher ranking. 1% &#8211; The Entry was described in an accepted INTERSPEECH 2021 paper. Winners will be selected within 7 days following the event. Winners will be notified within 7 days following the Event.<\/p>\n<p>In the event of a tie between any eligible entries, an additional judge will break the tie based on the judging criteria described above. The decisions of the judges are final and binding. If we do not receive enough entries meeting the entry requirements, we may, at our discretion, select fewer winners. If public vote determines winners, it is prohibited for any person to obtain votes by any fraudulent or inappropriate means, including offering prizes or other inducements in exchange for votes, automated programs, or fraudulent IDs. Microsoft will void any questionable votes.<\/p>\n<p>ODDS<br \/>\nThe odds of winning are based on the number and quality of eligible entries received.<\/p>\n<p>GENERAL CONDITIONS AND RELEASE OF LIABILITY<br \/>\nTo the extent allowed by law, by entering you agree to release and hold harmless Microsoft and its respective parents, partners, subsidiaries, affiliates, employees, and agents from any and all liability or any injury, loss, or damage of any kind arising in connection with this Contest or any prize won.<\/p>\n<p>All local laws apply. The decisions of Microsoft are final and binding.<\/p>\n<p>We reserve the right to cancel, change, or suspend this Contest for any reason, including cheating, technology failure, catastrophe, war, or any other unforeseen or unexpected event that affects the integrity of this Contest, whether human or mechanical. If the integrity of the Contest cannot be restored, we may select winners from among all eligible entries received before we had to cancel, change or suspend the Contest.<\/p>\n<p>If you attempt or we have strong reason to believe that you have compromised the integrity or the legitimate operation of this Contest by cheating, hacking, creating a bot or other automated program, or by committing fraud in any way, we may seek damages from you to the full extent of the law and you may be banned from participation in future Microsoft promotions.<\/p>\n<p>GOVERNING LAW<br \/>\nThis Contest will be governed by the laws of the State of Washington, and you consent to the exclusive jurisdiction and venue of the courts of the State of Washington for any disputes arising out of this Contest.<\/p>\n<p>PRIVACY<br \/>\nAt Microsoft, we are committed to protecting your privacy. Microsoft uses the information you provide on this form to notify you of important information about our products, upgrades and enhancements, and to send you information about other Microsoft products and services. Microsoft will not share the information you provide with third parties without your permission except where necessary to complete the services or transactions you have requested, or as required by law. Microsoft is committed to protecting the security of your personal information. We use a variety of security technologies and procedures to help protect your personal information from unauthorized access, use, or disclosure. Your personal information is never shared outside the company without your permission, except under conditions explained above.<\/p>\n<p>If you believe that Microsoft has not adhered to this statement, please contact Microsoft by sending an email to <a href=\"mailto:privrc@microsoft.com\">privrc@microsoft.com<\/a> or postal mail to Microsoft Privacy Response Center, Microsoft Corporation, One Microsoft Way, Redmond, WA<\/p>\n\n\n\n\n\n<h2>Program timeline<\/h2>\n<ul>\n<li><strong>January 8, 2021<\/strong>: Release of the datasets and scripts for training and testing<\/li>\n<li><strong>March 8, 2021<\/strong>: Blind test set released to participants<\/li>\n<li><strong>March 15, 2021<\/strong>: Deadline for participants to submit their results for P.835 subjective evaluation on the blind test set<\/li>\n<li><strong>March 22, 2021<\/strong>: Organizers will notify the participants about the results<\/li>\n<li><strong>March 26, 2021<\/strong>: Regular paper submission deadline for INTERSPEECH 2021<\/li>\n<li><strong>June 2, 2021<\/strong>: Paper acceptance\/rejection notification<\/li>\n<li><strong>June 4, 2021<\/strong>: Notification of the winners<\/li>\n<\/ul>\n\n\n\n\n\n<h2>Organizers<\/h2>\n<ul>\n<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/chkarada\/\">Chandan K A Reddy<\/a>, Microsoft, USA<\/li>\n<li>Hari Dubey, Microsoft, USA<\/li>\n<li>Kazuhito Koishida, Microsoft, USA<\/li>\n<li>Arun Nair, Johns Hopkins University, USA<\/li>\n<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/vigopal\/\">Vishak Gopal<\/a>, Microsoft, USA<\/li>\n<li>Ross Cutler, Microsoft, USA<\/li>\n<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/raichner\/\">Robert Aichner<\/a>, Microsoft, USA<\/li>\n<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/sebraun\/\">Sebastian Braun<\/a>, Microsoft Research, Germany<\/li>\n<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/hagamper\/\">Hannes Gamper<\/a>, Microsoft Research, USA<\/li>\n<li>Sriram Srinivasan, Microsoft, USA<\/li>\n<\/ul>\n\n\n\n\n\n<h2>Related links<\/h2>\n<ul>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/github.com\/microsoft\/DNS-Challenge\/tree\/master\/datasets\" target=\"_blank\" rel=\"noopener noreferrer\">Training and test datasets<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/github.com\/microsoft\/DNS-Challenge\" target=\"_blank\" rel=\"noopener noreferrer\">Data synthesizer and unit tests scripts<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/arxiv.org\/pdf\/2101.01902.pdf\" target=\"_blank\" rel=\"noopener noreferrer\">Base paper describing the challenge<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<li><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/github.com\/microsoft\/DNS-Challenge\/tree\/master\/DNSMOS\" target=\"_blank\" rel=\"noopener noreferrer\">DNSMOS Azure service<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/li>\n<\/ul>\n<h3>Other challenges<\/h3>\n<ul>\n<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/academic-program\/deep-noise-suppression-challenge-icassp-2021\/\">Deep Noise Suppression Challenge \u2013 ICASSP 2021<\/a><\/li>\n<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/academic-program\/acoustic-echo-cancellation-challenge-icassp-2021\/\">Acoustic Echo Cancellation Challenge &#8211; ICASSP 2021<\/a><\/li>\n<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/academic-program\/acoustic-echo-cancellation-challenge-interspeech-2021\/\">Acoustic Echo Cancellation Challenge \u2013 INTERSPEECH 2021<\/a><\/li>\n<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/academic-program\/deep-noise-suppression-challenge-interspeech-2020\/\">Deep Noise Suppression Challenge &#8211; INTERSPEECH 2020<\/a><\/li>\n<\/ul>\n\n\n\n\n\n<p><span style=\"text-decoration: underline\"><strong>Track 1 results:<\/strong><\/span><\/p>\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">The accepted papers for Track 1 are given below. Papers were accepted through the normal INTERSPEECH peer-review process.<\/p>\n<div style=\"direction: ltr\">\n<table style=\"direction: ltr;border-collapse: separate;border: 1pt solid #a3a3a3;border-spacing: 0px;width: 1369px\" title=\"\" border=\"1\" summary=\"\" cellspacing=\"0\" cellpadding=\"0\">\n<thead>\n<tr>\n<th style=\"background-color: #0070c0;vertical-align: top;width: 36px;padding: 0px;border: 1px solid\" scope=\"col\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt;color: white\">Place<\/p>\n<\/th>\n<th style=\"background-color: #0070c0;vertical-align: top;width: 84px;padding: 0px;border: 1px solid\" scope=\"col\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt;color: white\">Performance Rank<\/p>\n<\/th>\n<th style=\"background-color: #0070c0;vertical-align: top;width: 265px;padding: 0px;border: 1px solid\" scope=\"col\">\n<p style=\"margin: 0in;font-family: inherit;font-size: 11.0pt;color: white\">Team<\/p>\n<\/th>\n<th style=\"background-color: #0070c0;vertical-align: top;width: 378px;padding: 0px;border: 1px solid\" scope=\"col\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt;color: white\">Authors<\/p>\n<\/th>\n<th style=\"background-color: #0070c0;vertical-align: top;width: 606px;padding: 0px;border: 1px solid\" scope=\"col\">\n<p style=\"margin: 0in;font-family: inherit;font-size: 11.0pt;color: white\">Title<\/p>\n<\/th>\n<\/tr>\n<\/thead>\n<tbody>\n<tr>\n<td style=\"vertical-align: top;width: 36px;padding: 0px;border: 1px solid\" scope=\"row\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">1<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 84px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">1<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 265px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Chinese Academy of Sciences<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 378px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Andong Li, Wenzhe Liu, Xiaoxue Luo, Guochen Yu, Chengshi Zheng, Xiaodong Li<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 606px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\"><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/arxiv.org\/abs\/2106.12743\">A simultaneous denoising and dereverberation framework with target decoupling<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/p>\n<\/td>\n<\/tr>\n<tr>\n<td style=\"vertical-align: top;width: 36px;padding: 0px;border: 1px solid\" scope=\"row\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">2<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 84px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">2<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 265px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Nanjing University,<\/p>\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Horizon Robotics, Nanjing Institute of Advanced Artificial Intelligence<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 378px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Xiaohuai Le, Hongsheng Chen, Kai Chen, Jing Lu<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 606px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\"><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/arxiv.org\/abs\/2107.05429\">DPCRN: Dual-Path Convolution Recurrent Network for Single Channel Speech Enhancement<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/p>\n<\/td>\n<\/tr>\n<tr>\n<td style=\"vertical-align: top;width: 36px;padding: 0px;border: 1px solid\" scope=\"row\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">3<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 84px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">3<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 265px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Northwestern Polytechnical University<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 378px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Shubo Lv , Yanxin Hu , Shimin Zhang, Lei Xie<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 606px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\"><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/arxiv.org\/abs\/2106.08672\">Dccrn+: Channel-wise subband dccrn with snr estimation for speech enhancement<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/p>\n<\/td>\n<\/tr>\n<tr>\n<td style=\"vertical-align: top;width: 36px;padding: 0px;border: 1px solid\" scope=\"row\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">4<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 84px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">5<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 265px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Inner Mongolia University<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 378px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Kanghao Zhang, Shulin He, Hao Li, Xueliang Zhang<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 606px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\"><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/arxiv.org\/abs\/2105.02436\">DBNet: A Dual-branch Network Architecture Processing on Spectrum and Waveform for Single-channel Speech Enhancement<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/p>\n<\/td>\n<\/tr>\n<tr>\n<td style=\"vertical-align: top;width: 36px;padding: 0px;border: 1px solid\" scope=\"row\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">5<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 84px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">6<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 265px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Kuaishou Technology Co<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 378px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Xu Zhang, Xinlei Ren, Xiguang Zheng, Lianwu Chen, Chen Zhang, Liang Guo, Bing Yu<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 606px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\"><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/www.isca-speech.org\/archive\/pdfs\/interspeech_2021\/zhang21t_interspeech.pdf\">Low-Delay Speech Enhancement Using Perceptually Motivated Target and Loss<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/p>\n<\/td>\n<\/tr>\n<tr>\n<td style=\"vertical-align: top;width: 36px;padding: 0px;border: 1px solid\" scope=\"row\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">6<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 84px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">9<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 265px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Technische Universit\u00a8at Braunschwe<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 378px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Ziyi Xu, Maximilian Strake, Tim Fingscheid<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 606px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\"><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/arxiv.org\/abs\/2103.17088\">Deep noise suppression with non-intrusive pesqnet supervision enabling the use of real training data<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/p>\n<\/td>\n<\/tr>\n<tr>\n<td style=\"vertical-align: top;width: 36px;padding: 0px;border: 1px solid\" scope=\"row\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">7<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 84px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">10<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 265px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">University of Science and Technology of China<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 378px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Koen Oostermeijer, Qing Wang, Jun Du<\/p>\n<\/td>\n<td style=\"vertical-align: top;width: 606px;padding: 0px;border: 1px solid\">\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\"><a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"http:\/\/staff.ustc.edu.cn\/~jundu\/Publications\/publications\/oostermeijer21_interspeech.pdf\">Lightweight Causal Transformer with Local Self-Attention for Real-Time Speech Enhancement<span class=\"sr-only\"> (opens in new tab)<\/span><\/a><\/p>\n<\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<\/div>\n<p><strong>Speech MOS<\/strong><\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-738544 size-large\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track1_speechMOS-1024x374.png\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track1_speechMOS-1024x374.png 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track1_speechMOS-300x110.png 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track1_speechMOS-768x280.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track1_speechMOS-1536x561.png 1536w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track1_speechMOS-2048x748.png 2048w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track1_speechMOS-16x6.png 16w\" alt=\"Track 1 Speech MOS\" width=\"1024\" height=\"374\" \/><\/p>\n<p><strong>Background Noise MOS<\/strong><\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-738538 size-large\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track1_background_MOS-1024x370.png\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track1_background_MOS-1024x370.png 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track1_background_MOS-300x108.png 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track1_background_MOS-768x277.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track1_background_MOS-1536x555.png 1536w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track1_background_MOS-2048x740.png 2048w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track1_background_MOS-16x6.png 16w\" alt=\"Track 1 Background Noise MOS\" width=\"1024\" height=\"370\" \/><\/p>\n<p><strong>Overall MOS<\/strong><\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-738541 size-large\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track1_overall_MOS-1024x406.png\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track1_overall_MOS-1024x406.png 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track1_overall_MOS-300x119.png 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track1_overall_MOS-768x305.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track1_overall_MOS-1536x609.png 1536w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track1_overall_MOS-2048x812.png 2048w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track1_overall_MOS-16x6.png 16w\" alt=\"Track 1 Background Noise MOS\" width=\"1024\" height=\"406\" \/><\/p>\n<p><span style=\"text-decoration: underline\"><strong>Track 2 Results<\/strong><\/span><\/p>\n<p><strong>Speech MOS<\/strong><\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-738553 size-large\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track2_speechMOS-1024x121.png\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track2_speechMOS-1024x121.png 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track2_speechMOS-300x35.png 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track2_speechMOS-768x90.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track2_speechMOS-1536x181.png 1536w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track2_speechMOS-2048x241.png 2048w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track2_speechMOS-16x2.png 16w\" alt=\"track 2 speech MOS\" width=\"1024\" height=\"121\" \/><\/p>\n<p><strong>Background MOS<\/strong><\/p>\n<p><span style=\"text-decoration: underline\"><strong><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-738547 size-large\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track2_background_MOS-1024x122.png\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track2_background_MOS-1024x122.png 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track2_background_MOS-300x36.png 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track2_background_MOS-768x92.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track2_background_MOS-1536x184.png 1536w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track2_background_MOS-2048x245.png 2048w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track2_background_MOS-16x2.png 16w\" alt=\"track 2 Background MOS\" width=\"1024\" height=\"122\" \/><\/strong><\/span><\/p>\n<p><strong>Overall MOS<\/strong><\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-738550 size-large\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track2_overall_MOS-1024x120.png\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track2_overall_MOS-1024x120.png 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track2_overall_MOS-300x35.png 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track2_overall_MOS-768x90.png 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track2_overall_MOS-1536x179.png 1536w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track2_overall_MOS-2048x239.png 2048w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track2_overall_MOS-16x2.png 16w\" alt=\"track 2 Overall MOS\" width=\"1024\" height=\"120\" \/><\/p>\n\n\n","protected":false},"featured_media":0,"template":"","meta":{"msr-url-field":"","msr-podcast-episode":"","msrModifiedDate":"","msrModifiedDateEnabled":false,"ep_exclude_from_search":false,"_classifai_error":"","msr_hide_image_in_river":null,"footnotes":""},"msr-opportunity-type":[187426],"msr-region":[256048],"msr-locale":[268875],"msr-program-audience":[],"msr-post-option":[],"msr-impact-theme":[],"class_list":["post-717037","msr-academic-program","type-msr-academic-program","status-publish","hentry","msr-opportunity-type-challenges","msr-region-global","msr-locale-en_us"],"msr_description":"The Deep Noise Suppression (DNS) challenge is designed to foster innovation in the area of noise suppression to achieve superior perceptual speech quality.","msr_social_media":[],"related-researchers":[{"type":"user_nicename","display_name":"Robert Aichner","user_id":39781,"people_section":"Section name 0","alias":"raichner"},{"type":"user_nicename","display_name":"Sebastian Braun","user_id":37688,"people_section":"Section name 0","alias":"sebraun"},{"type":"user_nicename","display_name":"Ross Cutler","user_id":40660,"people_section":"Section name 0","alias":"rcutler"},{"type":"user_nicename","display_name":"Hannes Gamper","user_id":31943,"people_section":"Section name 0","alias":"hagamper"},{"type":"user_nicename","display_name":"Vishak Gopal","user_id":39624,"people_section":"Section name 0","alias":"vigopal"}],"tab-content":[{"id":0,"name":"About","content":"<strong>Program dates:<\/strong> January 2020 - March 2021\r\n\r\nThe Deep Noise Suppression (DNS) challenge is designed to foster innovation in the area of noise suppression to achieve superior perceptual speech quality. We recently organized a DNS challenge special session at INTERSPEECH 2020 and ICASSP 2020. We open-sourced training and test datasets for the wideband scenario. We also open-sourced a subjective evaluation framework based on ITU-T standard P.808, which was used to evaluate challenge submissions. Many researchers from academia and industry made significant contributions to push the field forward, yet even the best noise suppressor was far from achieving superior speech quality in challenging scenarios. In this version of the challenge organized at INTERSPEECH 2021, we are expanding both our training and test datasets to accommodate full band scenarios. The two tracks in this challenge will focus on real-time denoising for (i) wide band, and (ii) full band scenarios. We are also making available a reliable non-intrusive objective speech quality metric for wide band called DNSMOS for the participants to use during their development phase.\r\n\r\n<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/interspeech-2021-deep-noise-suppression-challenge\/\" target=\"_blank\" rel=\"noopener\">Challenge description (PDF)<\/a>\r\n\r\nWe will have two tracks in this challenge:\r\n<ul>\r\n \t<li><strong>Track 1: Real-Time Denoising track for wide band scenario<\/strong>\r\nThe noise suppressor must take less than the stride time Ts (in ms) to process a frame of size T (in ms) on an Intel Core i5 quad-core machine clocked at 2.4 GHz or equivalent processor. For example, Ts = T\/2 for 50% overlap between frames. The total algorithmic latency allowed including the frame size T, stride time Ts, and any look ahead must be less than or equal to 40ms. For example, for a real-time system that receives 20ms audio chunks, if you use a frame length of 20ms with a stride of 10ms resulting in an algorithmic latency of 30ms, then you satisfy the latency requirements. If you use a frame of size 32ms with a stride of 16ms resulting in an algorithmic latency of 48ms, then your method does not satisfy the latency requirements as the total algorithmic latency exceeds 40ms. If your frame size plus stride T1=T+Ts is less than 40ms, then you can use up to (40-T1) ms future information.<\/li>\r\n \t<li><strong>Track 2: Real-Time Denoising track for full band scenario<\/strong>\r\nSatisfy Track 1 requirements but at 48 kHz.<\/li>\r\n<\/ul>\r\n<strong>Participants are forbidden from using the blind test set to retrain or tweak their models. Participants must submit results only if they intend to submit a paper to INTERSPEECH 2021. Failing to adhere to these rules will lead to disqualification from the challenge.<\/strong>\r\n<h3>Registration<\/h3>\r\nPlease send an email to <a href=\"mailto:dns_challenge@microsoft.com?Subject=Challenge%20Query\" target=\"_blank\" rel=\"noopener\">dns_challenge@microsoft.com<\/a> stating that you are interested to participate in the challenge. Please include the following details in your email:\r\n<ul>\r\n \t<li>List of participants<\/li>\r\n \t<li>Affiliation of each participant<\/li>\r\n \t<li>Email ID of each participant<\/li>\r\n<\/ul>\r\nAlso, please create a new submission at <a href=\"https:\/\/cmt3.research.microsoft.com\/3rdDNSChallenge\" target=\"_blank\" rel=\"noopener\">https:\/\/cmt3.research.microsoft.com\/3rdDNSChallenge<\/a> and fill out all the details. This will help us to easily send out any announcements.\r\n\r\n<strong>Contact us:<\/strong> If you have questions about this program, email us at <a href=\"mailto:dns_challenge@microsoft.com\">dns_challenge@microsoft.com<\/a>."},{"id":1,"name":"Rules","content":"<h2>Official rules<\/h2>\r\nSPONSOR\r\nThese Official Rules (\u201cRules\u201d) govern the operation of the Microsoft M365 Core INTERSPEECH 2021 Deep Noise Suppression Event Contest (\u201cContest\u201d). Microsoft Corporation, One Microsoft Way, Redmond, WA, 98052, USA, is the Contest sponsor (\u201cSponsor\u201d).\r\n\r\nDEFINITIONS\r\nIn these Rules, \"Microsoft\", \"we\", \"our\", and \"us\", refer to Sponsor, and \"you\" and \"yourself\" refers to a Contest participant or the parent\/legal guardian of any Contest entrant who has not reached the age of majority to contractually obligate themselves in their legal place of residence. \"Event\" refers to the INTERSPEECH 2021 Deep Noise Suppression event held in Brno, Czechia (the \"Event\"). By entering you (your parent\/legal guardian if you are not the age of majority in your legal place of residence) agree to be bound by these Rules.\r\n\r\nENTRY PERIOD\r\nThe Contest will operate from January 8, 2021 to March 8, 2021 (\"Entry Period\"). The Entry Period is divided into several periods as described in Section 5 How to Enter.\r\n\r\nELIGIBILITY\r\nOpen to any registered Event attendee 18 years of age or older. If you are 18 years of age or older but have not reached the age of majority in your legal place of residence, then you must have the consent of a parent\/legal guardian. Employees and directors of Microsoft Corporation and its subsidiaries, affiliates, advertising agencies, and Contest Parties are not eligible, nor are persons involved in the execution or administration of this promotion, or the family members of each above (parents, children, siblings, spouse\/domestic partners, or individuals residing in the same household). Void in Cuba, Iran, North Korea, Sudan, Syria, Region of Crimea, and where prohibited. For business\/tradeshow events: If you are attending the Event in your capacity as an employee, it is your sole responsibility to comply with your employer's gift policies. Microsoft will not be a party to any disputes or actions related to this matter. PLEASE NOTE: If you are a public sector employee (government and education), all prize awards will be awarded directly to your public sector organization and subject to receipt of a gift letter signed by your agency\/institution's ethics officer, attorney, or designated executive\/officer responsible for your organization's gifts\/ethics policy. Microsoft seeks to ensure that by offering items of value at no charge in promotional settings it does not create any violation of the letter or spirit of the entrant's applicable gifts and ethics rules.\r\n\r\nHOW TO ENTER\r\nThe Contest Objective is to promote collaborative research in real-time single-channel Speech Enhancement aimed to maximize the subjective (perceptual) quality of the enhanced speech. Prizes will be awarded based on the speech quality of deep noise suppression models using the online subjective evaluation framework ITU-T P.835. Only methods described in accepted INTERSPEECH 2021 papers will be eligible for the contest. You may participate as an individual or a team. If forming a team, you must designate a \"Team Captain\" who will submit all entry materials on behalf of the team. Once you register as part of a Team, you cannot change Teams or alter your current team (either by adding or removing members) after the submission of your Entry. Limit one Entry per person and per team. You may not compete on multiple teams and you may not enter individually and on a team. We are not responsible for Entries that we do not receive for any reason, or for Entries that we receive but are not decipherable or not functional for any reason. Each Team is solely responsible for its own cooperation and teamwork. In no event will Microsoft officiate in any dispute regarding the conduct or cooperation of any Team or its members. The Contest will operate as follows:\r\n\r\nRegistration \/ Development Period: January 8, 2021 - March 8, 2021\r\nTo register, please send an email to <a href=\"mailto:dns_challenge@microsoft.com\">dns_challenge@microsoft.com<\/a> stating that you are interested to participate in the challenge. Please include the following details in your email:\r\n<ol>\r\n \t<li>Names of the participants and name of the team captain<\/li>\r\n \t<li>Institution\/Company<\/li>\r\n \t<li>Email<\/li>\r\n<\/ol>\r\nCreate a submission by registering at <a href=\"https:\/\/cmt3.research.microsoft.com\/3rdDNSChallenge\/\" target=\"_blank\" rel=\"noopener\">Conference Management Toolkit - 3rd DNS Challenge<\/a> and fill in all your details. You will be using this tool for final submission and to receive any email announcements from organizers.\r\n\r\nThen i. develop a speech enhancement model that best meets the Contest Objective as described in the base paper and ii. submit a paper to INTERSPEECH 2021 which reports the computational complexity of the model in terms of the number of parameters and the time it takes to infer a frame on a particular CPU (preferably Intel Core i5 quad core machine clocked at 2.4 GHz). To develop your model, use any publicly available clean speech and noise datasets, including the contest datasets provided for training and developing models. You may augment your datasets with the contest dataset. You may mix clean speech and noise in any way that improves the performance of your model.\r\n\r\nThe final evaluation will be conducted on a blind test set that is similar to the open sourced development stage test set. You may use scripts for a baseline noise suppressor that was recently published here. Testing \/ Entry Period: January 8, 2021 - March 8, 2021. On March 8, the blind test dataset will be released. You will have until 11:59 PM PT on March 15, 2021 to test your model against this dataset and create a set of enhanced clips to submit for judging (your \"Entry\") via Conference Management Tool.\r\n\r\nYou may not use the blind test set to retrain or tweak your model. To submit your entry, submit your processed clips via conference management tool. Each Entry will fall in one of two tracks based on the sampling rate. You must satisfy all the requirements of each track in terms of algorithmic latency. You must also specify the Number of operations per second in your paper submission. INTERSPEECH 2021 Paper Submission and Judging Period: March 15 - 11:59 PM PT June 2, 2021. Your Entry must be described in a paper accepted by INTERSPEECH 2021. To submit a paper, visit <a href=\"https:\/\/www.interspeech2021.org\/call-for-special-sessions-and-challenges\" target=\"_blank\" rel=\"noopener\">Call for Special Sessions &amp; Challenges | INTERSPEECH 2021<\/a> and follow instructions to submit a paper.\r\n\r\nThe entry limit is one per person during the Entry Period. Any attempt by any you to obtain more than the stated number of entries by using multiple\/different accounts, identities, registrations, logins, or any other methods will void your entries and you may be disqualified. Use of any automated system to participate is prohibited.\r\n\r\nWe are not responsible for excess, lost, late, or incomplete entries. If disputed, entries will be deemed submitted by the \"authorized account holder\" of the email address, social media account, or other method used to enter. The \"authorized account holder\" is the natural person assigned to an email address by internet or online service provider, or other organization responsible for assigning email addresses.\r\n\r\nELIGIBLE ENTRY\r\nTo be eligible, an entry must meet the following content\/technical requirements:\r\n<ul>\r\n \t<li>Your Entry must be the method described in a paper accepted by INTERSPEECH 2021<\/li>\r\n \t<li>Your entry must be your own original work; and<\/li>\r\n \t<li>Your entry cannot have been selected as a winner in any other contest; and<\/li>\r\n \t<li>You must have obtained any and all consents, approvals, or licenses required for you to submit your entry; and<\/li>\r\n \t<li>To the extent that entry requires the submission of user-generated content such as software, photos, videos, music, artwork, essays, etc., entrants warrant that their entry is their original work, has not been copied from others without permission or apparent rights, and does not violate the privacy, intellectual property rights, or other rights of any other person or entity. You may include Microsoft trademarks, logos, and designs, for which Microsoft grants you a limited license to use for the sole purposes of submitting an entry into this Contest; and<\/li>\r\n \t<li>Your entry may NOT contain, as determined by us in our sole and absolute discretion, any content that is obscene or offensive, violent, defamatory, disparaging or illegal, or that promotes alcohol, illegal drugs, tobacco or a particular political agenda, or that communicates messages that may reflect negatively on the goodwill of Microsoft.<\/li>\r\n \t<li>Your entry must NOT include enhanced clips using other noise suppression methods that you are not submitting to INTERSPEECH 2021.<\/li>\r\n<\/ul>\r\nUSE OF ENTRIES\r\nWe are not claiming ownership rights to your Submission. However, by submitting an entry, you grant us an irrevocable, royalty-free, worldwide right and license to use, review, assess, test, and otherwise analyze your entry and all its content in connection with this Contest and use your entry in any media whatsoever now known or later invented for any non-commercial or commercial purpose, including, but not limited to, the marketing, sale or promotion of Microsoft products or services, without further permission from you. You will not receive any compensation or credit for use of your entry, other than what is described in these Official Rules.\r\n\r\nBy entering you acknowledge that we may have developed or commissioned materials similar or identical to your entry and you waive any claims resulting from any similarities to your entry. Further, you understand that we will not restrict work assignments of representatives who have had access to your entry and you agree that the use of information in our representatives' unaided memories in the development or deployment of our products or services does not create liability for us under this agreement or copyright or trade secret law.\r\n\r\nYour entry may be posted on a public website. We are not responsible for any unauthorized use of your entry by visitors to this website. We are not obligated to use your entry for any purpose, even if it has been selected as a winning entry.\r\n\r\nWINNER SELECTION AND NOTIFICATION\r\nPending confirmation of eligibility, potential prize winners will be selected by Microsoft or their Agent or a qualified judging panel from among all eligible entries received based on the following judging criteria: 99% - The subjective speech quality evaluated on the blind test set using ITU-T P.835 framework. We will use the submitted clips with no alteration to conduct ITU-T P.835 subjective evaluation and pick the winners based on the results. Among the submitted proposals, if the difference between the models is not statistically significant, the model with a lower number of operations per second be given a higher ranking. 1% - The Entry was described in an accepted INTERSPEECH 2021 paper. Winners will be selected within 7 days following the event. Winners will be notified within 7 days following the Event.\r\n\r\nIn the event of a tie between any eligible entries, an additional judge will break the tie based on the judging criteria described above. The decisions of the judges are final and binding. If we do not receive enough entries meeting the entry requirements, we may, at our discretion, select fewer winners. If public vote determines winners, it is prohibited for any person to obtain votes by any fraudulent or inappropriate means, including offering prizes or other inducements in exchange for votes, automated programs, or fraudulent IDs. Microsoft will void any questionable votes.\r\n\r\nODDS\r\nThe odds of winning are based on the number and quality of eligible entries received.\r\n\r\nGENERAL CONDITIONS AND RELEASE OF LIABILITY\r\nTo the extent allowed by law, by entering you agree to release and hold harmless Microsoft and its respective parents, partners, subsidiaries, affiliates, employees, and agents from any and all liability or any injury, loss, or damage of any kind arising in connection with this Contest or any prize won.\r\n\r\nAll local laws apply. The decisions of Microsoft are final and binding.\r\n\r\nWe reserve the right to cancel, change, or suspend this Contest for any reason, including cheating, technology failure, catastrophe, war, or any other unforeseen or unexpected event that affects the integrity of this Contest, whether human or mechanical. If the integrity of the Contest cannot be restored, we may select winners from among all eligible entries received before we had to cancel, change or suspend the Contest.\r\n\r\nIf you attempt or we have strong reason to believe that you have compromised the integrity or the legitimate operation of this Contest by cheating, hacking, creating a bot or other automated program, or by committing fraud in any way, we may seek damages from you to the full extent of the law and you may be banned from participation in future Microsoft promotions.\r\n\r\nGOVERNING LAW\r\nThis Contest will be governed by the laws of the State of Washington, and you consent to the exclusive jurisdiction and venue of the courts of the State of Washington for any disputes arising out of this Contest.\r\n\r\nPRIVACY\r\nAt Microsoft, we are committed to protecting your privacy. Microsoft uses the information you provide on this form to notify you of important information about our products, upgrades and enhancements, and to send you information about other Microsoft products and services. Microsoft will not share the information you provide with third parties without your permission except where necessary to complete the services or transactions you have requested, or as required by law. Microsoft is committed to protecting the security of your personal information. We use a variety of security technologies and procedures to help protect your personal information from unauthorized access, use, or disclosure. Your personal information is never shared outside the company without your permission, except under conditions explained above.\r\n\r\nIf you believe that Microsoft has not adhered to this statement, please contact Microsoft by sending an email to <a href=\"mailto:privrc@microsoft.com\">privrc@microsoft.com<\/a> or postal mail to Microsoft Privacy Response Center, Microsoft Corporation, One Microsoft Way, Redmond, WA"},{"id":2,"name":"Timeline","content":"<h2>Program timeline<\/h2>\r\n<ul>\r\n \t<li><strong>January 8, 2021<\/strong>: Release of the datasets and scripts for training and testing<\/li>\r\n \t<li><strong>March 8, 2021<\/strong>: Blind test set released to participants<\/li>\r\n \t<li><strong>March 15, 2021<\/strong>: Deadline for participants to submit their results for P.835 subjective evaluation on the blind test set<\/li>\r\n \t<li><strong>March 22, 2021<\/strong>: Organizers will notify the participants about the results<\/li>\r\n \t<li><strong>March 26, 2021<\/strong>: Regular paper submission deadline for INTERSPEECH 2021<\/li>\r\n \t<li><strong>June 2, 2021<\/strong>: Paper acceptance\/rejection notification<\/li>\r\n \t<li><strong>June 4, 2021<\/strong>: Notification of the winners<\/li>\r\n<\/ul>"},{"id":3,"name":"Organizers","content":"<h2>Organizers<\/h2>\r\n<ul>\r\n \t<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/chkarada\/\">Chandan K A Reddy<\/a>, Microsoft, USA<\/li>\r\n \t<li>Hari Dubey, Microsoft, USA<\/li>\r\n \t<li>Kazuhito Koishida, Microsoft, USA<\/li>\r\n \t<li>Arun Nair, Johns Hopkins University, USA<\/li>\r\n \t<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/vigopal\/\">Vishak Gopal<\/a>, Microsoft, USA<\/li>\r\n \t<li>Ross Cutler, Microsoft, USA<\/li>\r\n \t<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/raichner\/\">Robert Aichner<\/a>, Microsoft, USA<\/li>\r\n \t<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/sebraun\/\">Sebastian Braun<\/a>, Microsoft Research, Germany<\/li>\r\n \t<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/people\/hagamper\/\">Hannes Gamper<\/a>, Microsoft Research, USA<\/li>\r\n \t<li>Sriram Srinivasan, Microsoft, USA<\/li>\r\n<\/ul>"},{"id":4,"name":"Links","content":"<h2>Related links<\/h2>\r\n<ul>\r\n \t<li><a href=\"https:\/\/github.com\/microsoft\/DNS-Challenge\/tree\/master\/datasets\" target=\"_blank\" rel=\"noopener\">Training and test datasets<\/a><\/li>\r\n \t<li><a href=\"https:\/\/github.com\/microsoft\/DNS-Challenge\" target=\"_blank\" rel=\"noopener\">Data synthesizer and unit tests scripts<\/a><\/li>\r\n \t<li><a href=\"https:\/\/arxiv.org\/pdf\/2101.01902.pdf\" target=\"_blank\" rel=\"noopener\">Base paper describing the challenge<\/a><\/li>\r\n \t<li><a href=\"https:\/\/github.com\/microsoft\/DNS-Challenge\/tree\/master\/DNSMOS\" target=\"_blank\" rel=\"noopener\">DNSMOS Azure service<\/a><\/li>\r\n<\/ul>\r\n<h3>Other challenges<\/h3>\r\n<ul>\r\n \t<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/academic-program\/deep-noise-suppression-challenge-icassp-2021\/\">Deep Noise Suppression Challenge \u2013 ICASSP 2021<\/a><\/li>\r\n \t<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/academic-program\/acoustic-echo-cancellation-challenge-icassp-2021\/\">Acoustic Echo Cancellation Challenge - ICASSP 2021<\/a><\/li>\r\n \t<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/academic-program\/acoustic-echo-cancellation-challenge-interspeech-2021\/\">Acoustic Echo Cancellation Challenge \u2013 INTERSPEECH 2021<\/a><\/li>\r\n \t<li><a href=\"https:\/\/www.microsoft.com\/en-us\/research\/academic-program\/deep-noise-suppression-challenge-interspeech-2020\/\">Deep Noise Suppression Challenge - INTERSPEECH 2020<\/a><\/li>\r\n<\/ul>"},{"id":5,"name":"Results","content":"<span style=\"text-decoration: underline\"><strong>Track 1 results:<\/strong><\/span>\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">The accepted papers for Track 1 are given below. Papers were accepted through the normal INTERSPEECH peer-review process.<\/p>\r\n\r\n<div style=\"direction: ltr\">\r\n<table style=\"direction: ltr;border-collapse: separate;border: 1pt solid #a3a3a3;border-spacing: 0px;width: 1369px\" title=\"\" border=\"1\" summary=\"\" cellspacing=\"0\" cellpadding=\"0\">\r\n<tbody>\r\n<tr>\r\n<td style=\"background-color: #0070c0;vertical-align: top;width: 36px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt;color: white\">Place<\/p>\r\n<\/td>\r\n<td style=\"background-color: #0070c0;vertical-align: top;width: 84px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt;color: white\">Performance Rank<\/p>\r\n<\/td>\r\n<td style=\"background-color: #0070c0;vertical-align: top;width: 265px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: inherit;font-size: 11.0pt;color: white\">Team<\/p>\r\n<\/td>\r\n<td style=\"background-color: #0070c0;vertical-align: top;width: 378px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt;color: white\">Authors<\/p>\r\n<\/td>\r\n<td style=\"background-color: #0070c0;vertical-align: top;width: 606px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: inherit;font-size: 11.0pt;color: white\">Title<\/p>\r\n<\/td>\r\n<\/tr>\r\n<tr>\r\n<td style=\"vertical-align: top;width: 36px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">1<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 84px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">1<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 265px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Chinese Academy of Sciences<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 378px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Andong Li, Wenzhe Liu, Xiaoxue Luo, Guochen Yu, Chengshi Zheng, Xiaodong Li<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 606px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\"><a href=\"https:\/\/arxiv.org\/abs\/2106.12743\">A simultaneous denoising and dereverberation framework with target decoupling<\/a><\/p>\r\n<\/td>\r\n<\/tr>\r\n<tr>\r\n<td style=\"vertical-align: top;width: 36px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">2<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 84px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">2<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 265px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Nanjing University,<\/p>\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Horizon Robotics, Nanjing Institute of Advanced Artificial Intelligence<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 378px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Xiaohuai Le, Hongsheng Chen, Kai Chen, Jing Lu<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 606px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\"><a href=\"https:\/\/arxiv.org\/abs\/2107.05429\">DPCRN: Dual-Path Convolution Recurrent Network for Single Channel Speech Enhancement<\/a><\/p>\r\n<\/td>\r\n<\/tr>\r\n<tr>\r\n<td style=\"vertical-align: top;width: 36px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">3<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 84px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">3<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 265px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Northwestern Polytechnical University<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 378px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Shubo Lv , Yanxin Hu , Shimin Zhang, Lei Xie<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 606px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\"><a href=\"https:\/\/arxiv.org\/abs\/2106.08672\">Dccrn+: Channel-wise subband dccrn with snr estimation for speech enhancement<\/a><\/p>\r\n<\/td>\r\n<\/tr>\r\n<tr>\r\n<td style=\"vertical-align: top;width: 36px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">4<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 84px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">5<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 265px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Inner Mongolia University<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 378px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Kanghao Zhang, Shulin He, Hao Li, Xueliang Zhang<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 606px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\"><a href=\"https:\/\/arxiv.org\/abs\/2105.02436\">DBNet: A Dual-branch Network Architecture Processing on Spectrum and Waveform for Single-channel Speech Enhancement<\/a><\/p>\r\n<\/td>\r\n<\/tr>\r\n<tr>\r\n<td style=\"vertical-align: top;width: 36px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">5<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 84px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">6<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 265px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Kuaishou Technology Co<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 378px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Xu Zhang, Xinlei Ren, Xiguang Zheng, Lianwu Chen, Chen Zhang, Liang Guo, Bing Yu<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 606px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\"><a href=\"https:\/\/www.isca-speech.org\/archive\/pdfs\/interspeech_2021\/zhang21t_interspeech.pdf\">Low-Delay Speech Enhancement Using Perceptually Motivated Target and Loss<\/a><\/p>\r\n<\/td>\r\n<\/tr>\r\n<tr>\r\n<td style=\"vertical-align: top;width: 36px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">6<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 84px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">9<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 265px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Technische Universit\u00a8at Braunschwe<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 378px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Ziyi Xu, Maximilian Strake, Tim Fingscheid<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 606px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\"><a href=\"https:\/\/arxiv.org\/abs\/2103.17088\">Deep noise suppression with non-intrusive pesqnet supervision enabling the use of real training data<\/a><\/p>\r\n<\/td>\r\n<\/tr>\r\n<tr>\r\n<td style=\"vertical-align: top;width: 36px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">7<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 84px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">10<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 265px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">University of Science and Technology of China<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 378px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\">Koen Oostermeijer, Qing Wang, Jun Du<\/p>\r\n<\/td>\r\n<td style=\"vertical-align: top;width: 606px;padding: 0px;border: 1px solid\">\r\n<p style=\"margin: 0in;font-family: Calibri;font-size: 11.0pt\"><a href=\"http:\/\/staff.ustc.edu.cn\/~jundu\/Publications\/publications\/oostermeijer21_interspeech.pdf\">Lightweight Causal Transformer with Local Self-Attention for Real-Time Speech Enhancement<\/a><\/p>\r\n<\/td>\r\n<\/tr>\r\n<\/tbody>\r\n<\/table>\r\n<\/div>\r\n<strong>Speech MOS<\/strong>\r\n\r\n<img class=\"aligncenter wp-image-738544 size-large\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track1_speechMOS-1024x374.png\" alt=\"Track 1 Speech MOS\" width=\"1024\" height=\"374\" \/>\r\n\r\n<strong>Background Noise MOS<\/strong>\r\n\r\n<img class=\"aligncenter wp-image-738538 size-large\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track1_background_MOS-1024x370.png\" alt=\"Track 1 Background Noise MOS\" width=\"1024\" height=\"370\" \/>\r\n\r\n<strong>Overall MOS<\/strong>\r\n\r\n<img class=\"aligncenter wp-image-738541 size-large\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track1_overall_MOS-1024x406.png\" alt=\"Track 1 Background Noise MOS\" width=\"1024\" height=\"406\" \/>\r\n\r\n<span style=\"text-decoration: underline\"><strong>Track 2 Results<\/strong><\/span>\r\n\r\n<strong>Speech MOS<\/strong>\r\n\r\n<img class=\"aligncenter wp-image-738553 size-large\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track2_speechMOS-1024x121.png\" alt=\"track 2 speech MOS\" width=\"1024\" height=\"121\" \/>\r\n\r\n<strong>Background MOS<\/strong>\r\n\r\n<span style=\"text-decoration: underline\"><strong><img class=\"aligncenter wp-image-738547 size-large\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track2_background_MOS-1024x122.png\" alt=\"track 2 Background MOS\" width=\"1024\" height=\"122\" \/><\/strong><\/span>\r\n\r\n<strong>Overall MOS<\/strong>\r\n\r\n<img class=\"aligncenter wp-image-738550 size-large\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2021\/01\/track2_overall_MOS-1024x120.png\" alt=\"track 2 Overall MOS\" width=\"1024\" height=\"120\" \/>"}],"msr_impact_theme":[],"_links":{"self":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-academic-program\/717037","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-academic-program"}],"about":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/types\/msr-academic-program"}],"version-history":[{"count":11,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-academic-program\/717037\/revisions"}],"predecessor-version":[{"id":1106439,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-academic-program\/717037\/revisions\/1106439"}],"wp:attachment":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media?parent=717037"}],"wp:term":[{"taxonomy":"msr-opportunity-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-opportunity-type?post=717037"},{"taxonomy":"msr-region","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-region?post=717037"},{"taxonomy":"msr-locale","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-locale?post=717037"},{"taxonomy":"msr-program-audience","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-program-audience?post=717037"},{"taxonomy":"msr-post-option","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-post-option?post=717037"},{"taxonomy":"msr-impact-theme","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-impact-theme?post=717037"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}