{"id":698635,"date":"2020-10-19T08:03:01","date_gmt":"2020-10-19T15:03:01","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/?p=698635"},"modified":"2021-10-12T12:26:14","modified_gmt":"2021-10-12T19:26:14","slug":"microsoft-turing-universal-language-representation-model-t-ulrv2-tops-xtreme-leaderboard","status":"publish","type":"post","link":"https:\/\/www.microsoft.com\/en-us\/research\/blog\/microsoft-turing-universal-language-representation-model-t-ulrv2-tops-xtreme-leaderboard\/","title":{"rendered":"Microsoft Turing Universal Language Representation model, T-ULRv2, tops XTREME leaderboard"},"content":{"rendered":"\n<p>Today, <strong>we are happy to announce that Turing multilingual language model (T-ULRv2) is the state of the art at the top of the Google <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/sites.research.google\/xtreme\">XTREME public leaderboard<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>.<\/strong> Created by the Microsoft Turing team in collaboration with Microsoft Research, the model beat the previous best from Alibaba (VECO) by 3.5 points in average score. To achieve this, in addition to the pretrained model, we leveraged \u201cStableTune,\u201d a novel multilingual fine-tuning technique based on stability training. Other models on the leaderboard include XLM-R, mBERT, XLM and more. One of the previous best submissions is also from Microsoft using <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/filter-an-enhanced-fusion-method-for-cross-lingual-language-understanding\/\">FILTER<\/a>.<\/p>\n\n\n\n<div class=\"annotations \" data-bi-aN=\"margin-callout\">\n\t<article class=\"annotations__list card depth-16 bg-body p-4 annotations__list--left\">\n\t\t<div class=\"annotations__list-item\">\n\t\t\t\t\t\t<span class=\"annotations__type d-block text-uppercase font-weight-semibold text-neutral-300 small\">PUBLICATION<\/span>\n\t\t\t<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/filter-an-enhanced-fusion-method-for-cross-lingual-language-understanding\/\" data-bi-cN=\"FILTER: An Enhanced Fusion Method for Cross-lingual Language Understanding\" data-external-link=\"false\" data-bi-aN=\"margin-callout\" data-bi-type=\"annotated-link\" class=\"annotations__link font-weight-semibold text-decoration-none\"><span>FILTER: An Enhanced Fusion Method for Cross-lingual Language Understanding<\/span>&nbsp;<span class=\"glyph-in-link glyph-append glyph-append-chevron-right\" aria-hidden=\"true\"><\/span><\/a>\t\t\t\t\t<\/div>\n\t<\/article>\n<\/div>\n\n\n\n<figure class=\"wp-block-image size-large\"><img decoding=\"async\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/Tuirng_Extreme_Leaderboard-Graphic-1024x638.jpg\" alt=\"FILTER: An Enhanced Fusion Method for Cross-lingual Language Understanding\"\/><\/figure>\n\n\n\n<h2 id=\"universal-language-representation\">Universal Language Representation<\/h2>\n\n\n\n<div class=\"annotations \" data-bi-aN=\"margin-callout\">\n\t<article class=\"annotations__list card depth-16 bg-body p-4 annotations__list--right\">\n\t\t<div class=\"annotations__list-item\">\n\t\t\t\t\t\t<span class=\"annotations__type d-block text-uppercase font-weight-semibold text-neutral-300 small\">PUBLICATION<\/span>\n\t\t\t<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/towards-language-agnostic-universal-representations\/\" data-bi-cN=\"Towards Language Agnostic Universal Representations\" data-external-link=\"false\" data-bi-aN=\"margin-callout\" data-bi-type=\"annotated-link\" class=\"annotations__link font-weight-semibold text-decoration-none\"><span>Towards Language Agnostic Universal Representations<\/span>&nbsp;<span class=\"glyph-in-link glyph-append glyph-append-chevron-right\" aria-hidden=\"true\"><\/span><\/a>\t\t\t\t\t<\/div>\n\t<\/article>\n<\/div>\n\n\n\n<p>The Microsoft Turing team has long believed that language representation should be universal. In this <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/towards-language-agnostic-universal-representations\/\">paper<\/a>, published in 2018, we presented a method to train language-agnostic representation in an unsupervised fashion. This kind of approach would allow for the trained model to be fine-tuned in one language and applied to a different one in a zero-shot fashion. This would overcome the challenge of requiring labeled data to train the model in every language. Since the publication of that paper, unsupervised pretrained language modeling has become the backbone of all NLP models, with transformer-based models at the heart of all such innovation.<\/p>\n\n\n\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"576\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/figure-turing-blog-1024x576.jpg\" alt=\"Towards Language Agnostic Universal Representations\" class=\"wp-image-698764\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/figure-turing-blog-1024x576.jpg 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/figure-turing-blog-300x169.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/figure-turing-blog-768x432.jpg 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/figure-turing-blog-1066x600.jpg 1066w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/figure-turing-blog-655x368.jpg 655w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/figure-turing-blog-343x193.jpg 343w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/figure-turing-blog-640x360.jpg 640w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/figure-turing-blog-960x540.jpg 960w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/figure-turing-blog.jpg 1247w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure>\n\n\n\n<div class=\"annotations \" data-bi-aN=\"margin-callout\">\n\t<article class=\"annotations__list card depth-16 bg-body p-4 annotations__list--left\">\n\t\t<div class=\"annotations__list-item\">\n\t\t\t\t\t\t<span class=\"annotations__type d-block text-uppercase font-weight-semibold text-neutral-300 small\">PUBLICATION<\/span>\n\t\t\t<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/infoxlm-an-information-theoretic-framework-for-cross-lingual-language-model-pre-training\/\" data-bi-cN=\"INFOXLM: An Information-Theoretic Framework for Cross-Lingual Language Model Pre-Training\" data-external-link=\"false\" data-bi-aN=\"margin-callout\" data-bi-type=\"annotated-link\" class=\"annotations__link font-weight-semibold text-decoration-none\"><span>INFOXLM: An Information-Theoretic Framework for Cross-Lingual Language Model Pre-Training<\/span>&nbsp;<span class=\"glyph-in-link glyph-append glyph-append-chevron-right\" aria-hidden=\"true\"><\/span><\/a>\t\t\t\t\t<\/div>\n\t<\/article>\n<\/div>\n\n\n\n<p>As part of <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/innovation.microsoft.com\/en-us\/ai-at-scale\">Microsoft AI at Scale<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, the Turing family of NLP models have been powering the next generation of AI experiences in Microsoft products. The Turing Universal Language Representation (T-ULRv2) model is our latest cross-lingual innovation, which incorporates our recent innovation of <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/publication\/infoxlm-an-information-theoretic-framework-for-cross-lingual-language-model-pre-training\/\">InfoXLM<\/a>, to create a universal model that represents 94 languages in the same vector space. In a recent <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/blogs.bing.com\/search-quality-insights\/september-2020\/Introducing-the-next-wave-of-AI-at-Scale-innovations-in-Bing\">blog post<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>, we discussed how we used T-ULR to scale Microsoft Bing intelligent answers to all supported languages and regions. The same model is being used to extend Microsoft Word <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/insider.office.com\/en-us\/blog\/microsoft-search-search-your-document-like-you-search-the-web\">Semantic Search<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> functionality beyond the English language and to power <a href=\"https:\/\/www.microsoft.com\/en-us\/research\/group\/msai\/articles\/assistive-ai-makes-replying-easier-2\/\">Suggested Replies<\/a> for Microsoft Outlook and Microsoft Teams universally. We will have these universal experiences coming to our users soon.<\/p>\n\n\n\n<div class=\"wp-block-image\"><figure class=\"aligncenter size-large\"><img decoding=\"async\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/Figure-2_turing-Updated-1024x286.jpg\" alt=\"Examples of Microsoft Bing intelligent answers in Spanish and Arabic, powered by T-ULR\"\/><figcaption>Examples of Microsoft Bing intelligent answers in Spanish and Arabic, powered by T-ULR<\/figcaption><\/figure><\/div>\n\n\n\n<div class=\"wp-block-image\"><figure class=\"aligncenter size-large\"><a href=\"https:\/\/insider.office.com\/en-us\/blog\/microsoft-search-search-your-document-like-you-search-the-web\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"616\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/Turing-Fig-3-1024x616.jpg\" alt=\"Example of Microsoft Word Semantic Search in French, powered by T-ULR\" class=\"wp-image-698674\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/Turing-Fig-3-1024x616.jpg 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/Turing-Fig-3-300x180.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/Turing-Fig-3-768x462.jpg 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/Turing-Fig-3.jpg 1172w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/a><figcaption>Example of Microsoft Word Semantic Search in French, powered by T-ULR<\/figcaption><\/figure><\/div>\n\n\n\n<p>These real products scenarios require extremely high quality and therefore provide the perfect test bed for our AI models. As a result, most of our models are near state of the art in accuracy and performance on NLP tasks.<\/p>\n\n\n\n\n\t<div class=\"border-bottom border-top border-gray-300 mt-5 mb-5 msr-promo text-center text-md-left alignwide\" data-bi-aN=\"promo\" data-bi-id=\"1160910\">\n\t\t\n\n\t\t<p class=\"msr-promo__label text-gray-800 text-center text-uppercase\">\n\t\t<span class=\"px-4 bg-white display-inline-block font-weight-semibold small\">video series<\/span>\n\t<\/p>\n\t\n\t<div class=\"row pt-3 pb-4 align-items-center\">\n\t\t\t\t\t\t<div class=\"msr-promo__media col-12 col-md-5\">\n\t\t\t\t<a class=\"bg-gray-300 display-block\" href=\"https:\/\/www.microsoft.com\/en-us\/research\/story\/on-second-thought\/\" aria-label=\"On Second Thought\" data-bi-cN=\"On Second Thought\" target=\"_blank\">\n\t\t\t\t\t<img decoding=\"async\" class=\"w-100 display-block\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2026\/01\/MFST_feature_SecondThought_1400x788.jpg\" alt=\"On Second Thought with Sinead Bovell\" \/>\n\t\t\t\t<\/a>\n\t\t\t<\/div>\n\t\t\t\n\t\t\t<div class=\"msr-promo__content p-3 px-5 col-12 col-md\">\n\n\t\t\t\t\t\t\t\t\t<h2 class=\"h4\">On Second Thought<\/h2>\n\t\t\t\t\n\t\t\t\t\t\t\t\t<p id=\"on-second-thought\" class=\"large\">A video series with Sinead Bovell built around the questions everyone\u2019s asking about AI. With expert voices from across Microsoft, we break down the tension and promise of this rapidly changing technology, exploring what\u2019s evolving and what\u2019s possible.<\/p>\n\t\t\t\t\n\t\t\t\t\t\t\t\t<div class=\"wp-block-buttons justify-content-center justify-content-md-start\">\n\t\t\t\t\t<div class=\"wp-block-button\">\n\t\t\t\t\t\t<a href=\"https:\/\/www.microsoft.com\/en-us\/research\/story\/on-second-thought\/\" aria-describedby=\"on-second-thought\" class=\"btn btn-brand glyph-append glyph-append-chevron-right\" data-bi-cN=\"On Second Thought\" target=\"_blank\">\n\t\t\t\t\t\t\tExplore the series\t\t\t\t\t\t<\/a>\n\t\t\t\t\t<\/div>\n\t\t\t\t<\/div>\n\t\t\t\t\t\t\t<\/div><!--\/.msr-promo__content-->\n\t<\/div><!--\/.msr-promo__inner-wrap-->\n\t<\/div><!--\/.msr-promo-->\n\t\n\n\n\n<h2 id=\"xtreme-benchmark\">XTREME Benchmark<\/h2>\n\n\n\n<p>The <strong>C<\/strong>ross-lingual <strong>TR<\/strong>ansfer <strong>E<\/strong>valuation of <strong>M<\/strong>ultilingual <strong>E<\/strong>ncoders (<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/sites.research.google\/xtreme\">XTREME<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>) benchmark covers 40 typologically diverse languages that span 12 language families, and it includes 9 tasks that require reasoning about different levels of syntax or semantics. The languages in XTREME are selected to maximize language diversity, coverage in existing tasks, and availability of training data.<\/p>\n\n\n\n<p>The tasks included in XTREME cover a range of paradigms, including sentence text classification, structured prediction, sentence retrieval and cross-lingual question answering. Consequently, for models to be successful on the XTREME benchmarks, they must learn representations that generalize to many standard cross-lingual transfer settings.<\/p>\n\n\n\n<p>For a full description of the benchmark, languages, and tasks, please see&nbsp;<a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" href=\"https:\/\/arxiv.org\/abs\/2003.11080\" target=\"_blank\" rel=\"noopener noreferrer\">XTREME: A Massively Multilingual Multi-task Benchmark for Evaluating Cross-lingual Generalization<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>.<\/p>\n\n\n\n<h2 id=\"t-ulrv2-data-architecture-and-pretraining\">T-ULRv2: Data, Architecture, and Pretraining<\/h2>\n\n\n\n<p>Turing Universal Language Representation (T-ULRv2) is a transformer architecture with 24 layers and 1,024 hidden states, with a total of 550 million parameters. T-ULRv2 pretraining has three different tasks: multilingual masked language modeling (MMLM), translation language modeling (TLM) and cross-lingual contrast (XLCo). The objective of the MMLM task, also known as Cloze task, is to predict masked tokens from inputs in different languages. T-ULRv2 uses a multilingual data corpus from web that consists of 94 languages for MMLM task training. Like MMLM, TLM task is also to predict masked tokens, but the prediction is conditioned on concatenated translation pairs. For example, given a pair of sentences in English and French, the model can predict the masked English token by either attending to surrounding English tokens or to its French translation. This helps the model align representations in different languages.<\/p>\n\n\n\n<div class=\"wp-block-image\"><figure class=\"aligncenter size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"584\" height=\"322\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/Turing_figure-4.jpg\" alt=\"Cross-lingual pretraining Masked Language Modeling (MLM) and TLM tasks (source: XLM)\" class=\"wp-image-698686\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/Turing_figure-4.jpg 584w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/Turing_figure-4-300x165.jpg 300w\" sizes=\"auto, (max-width: 584px) 100vw, 584px\" \/><figcaption>Cross-lingual pretraining Masked Language Modeling (MLM) and TLM tasks (source: <a href=\"https:\/\/arxiv.org\/pdf\/1901.07291.pdf\">XLM<\/a>)<\/figcaption><\/figure><\/div>\n\n\n\n<p>XLCo also uses parallel training data. The objective of the task is to maximize the mutual information between the representations of parallel sentences. Unlike maximizing token-sequence mutual information as in MMLM and TLM, XLCo targets cross-lingual sequence-level mutual information. T-ULRv2 uses translation parallel data with 14 language pairs for both TLM and XLCo tasks.<\/p>\n\n\n\n<p>The loss function for XLCo is as follows:<\/p>\n\n\n\n<div class=\"wp-block-image\"><figure class=\"aligncenter size-large is-resized\"><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/Turing-Latex-1.jpg\" alt=\"Loss function for XLCo\" class=\"wp-image-698776\" width=\"492\" height=\"114\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/Turing-Latex-1.jpg 914w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/Turing-Latex-1-300x70.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/Turing-Latex-1-768x178.jpg 768w\" sizes=\"auto, (max-width: 492px) 100vw, 492px\" \/><\/figure><\/div>\n\n\n\n<p>This is subsequently added to the MMLM and TLM loss to get the overall loss for the cross-lingual pretraining:<\/p>\n\n\n\n<div class=\"wp-block-image\"><figure class=\"aligncenter size-large is-resized\"><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/Turing-latex-2.jpg\" alt=\"Function to get get the overall loss for the cross-lingual pretraining\" class=\"wp-image-698782\" width=\"439\" height=\"83\"\/><\/figure><\/div>\n\n\n\n<h2 id=\"t-ulrv2-release-information\">T-ULRv2: Release Information<\/h2>\n\n\n\n<p>At Microsoft Ignite 2020, we <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/blogs.microsoft.com\/ai-for-business\/ai-at-scale-ignite\/\">announced<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> that Turing models will be made available for building custom applications as part of a private preview. T-ULRv2 will also be part of this program. If you are interested in learning more about this and other Turing models, you can submit a request <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"http:\/\/aka.ms\/turing-earlyaccess\">here<span class=\"sr-only\"> (opens in new tab)<\/span><\/a>. We are closely collaborating with <a class=\"msr-external-link glyph-append glyph-append-open-in-new-tab glyph-append-xsmall\" rel=\"noopener noreferrer\" target=\"_blank\" href=\"https:\/\/azure.microsoft.com\/en-us\/services\/cognitive-services\/\">Azure Cognitive Services<span class=\"sr-only\"> (opens in new tab)<\/span><\/a> to power current and future language services with Turing models. Existing Azure Cognitive Services customers will automatically benefit from these improvements through the APIs.<\/p>\n\n\n\n<h2 id=\"democratizing-our-ai-experiences\">Democratizing Our AI Experiences<\/h2>\n\n\n\n<p>At Microsoft, globalization is not just a research problem. It is a product challenge that we must face head on. Windows ships everywhere in the world. Microsoft Office and Microsoft Bing are available in over 100 languages across 200 regions. We have customers in every corner of the planet, and they use our products in their native languages. To truly democratize our product experience to empower <strong>all users<\/strong> and efficiently scale globally, we are pushing the boundaries of multilingual models. The result is language-agnostic representations like T-ULRv2 that improve product experiences across all languages.<\/p>\n\n\n\n<p>The Microsoft Turing team welcomes your feedback and comments and looks forward to sharing more developments in the future.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Today, we are happy to announce that Turing multilingual language model (T-ULRv2) is the state of the art at the top of the Google XTREME public leaderboard (opens in new tab). Created by the Microsoft Turing team in collaboration with Microsoft Research, the model beat the previous best from Alibaba (VECO) by 3.5 points in [&hellip;]<\/p>\n","protected":false},"author":38838,"featured_media":699064,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"msr-url-field":"","msr-podcast-episode":"","msrModifiedDate":"","msrModifiedDateEnabled":false,"ep_exclude_from_search":false,"_classifai_error":"","msr-author-ordering":[{"type":"user_nicename","value":"Saurabh Tiwary","user_id":"39603"},{"type":"user_nicename","value":"Ming Zhou","user_id":"32942"}],"msr_hide_image_in_river":0,"footnotes":""},"categories":[1],"tags":[],"research-area":[13556,13545],"msr-region":[],"msr-event-type":[],"msr-locale":[268875],"msr-post-option":[],"msr-impact-theme":[],"msr-promo-type":[],"msr-podcast-series":[],"class_list":["post-698635","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-research-blog","msr-research-area-artificial-intelligence","msr-research-area-human-language-technologies","msr-locale-en_us"],"msr_event_details":{"start":"","end":"","location":""},"podcast_url":"","podcast_episode":"","msr_research_lab":[199560],"msr_impact_theme":[],"related-publications":[],"related-downloads":[],"related-videos":[],"related-academic-programs":[],"related-groups":[],"related-projects":[649749],"related-events":[],"related-researchers":[],"msr_type":"Post","featured_image_thumbnail":"<img width=\"960\" height=\"540\" src=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/1400x788_Leaderboard_NoLogo-1-960x540.jpg\" class=\"img-object-cover\" alt=\"\" decoding=\"async\" loading=\"lazy\" srcset=\"https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/1400x788_Leaderboard_NoLogo-1-960x540.jpg 960w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/1400x788_Leaderboard_NoLogo-1-300x169.jpg 300w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/1400x788_Leaderboard_NoLogo-1-1024x576.jpg 1024w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/1400x788_Leaderboard_NoLogo-1-768x432.jpg 768w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/1400x788_Leaderboard_NoLogo-1-1536x865.jpg 1536w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/1400x788_Leaderboard_NoLogo-1-2048x1153.jpg 2048w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/1400x788_Leaderboard_NoLogo-1-1066x600.jpg 1066w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/1400x788_Leaderboard_NoLogo-1-655x368.jpg 655w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/1400x788_Leaderboard_NoLogo-1-343x193.jpg 343w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/1400x788_Leaderboard_NoLogo-1-640x360.jpg 640w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/1400x788_Leaderboard_NoLogo-1-1280x720.jpg 1280w, https:\/\/www.microsoft.com\/en-us\/research\/wp-content\/uploads\/2020\/10\/1400x788_Leaderboard_NoLogo-1-1920x1080.jpg 1920w\" sizes=\"auto, (max-width: 960px) 100vw, 960px\" \/>","byline":"Saurabh Tiwary and Ming Zhou","formattedDate":"October 19, 2020","formattedExcerpt":"Today, we are happy to announce that Turing multilingual language model (T-ULRv2) is the state of the art at the top of the Google XTREME public leaderboard (opens in new tab). Created by the Microsoft Turing team in collaboration with Microsoft Research, the model beat&hellip;","locale":{"slug":"en_us","name":"English","native":"","english":"English"},"_links":{"self":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts\/698635","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/users\/38838"}],"replies":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/comments?post=698635"}],"version-history":[{"count":13,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts\/698635\/revisions"}],"predecessor-version":[{"id":784426,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/posts\/698635\/revisions\/784426"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media\/699064"}],"wp:attachment":[{"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/media?parent=698635"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/categories?post=698635"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/tags?post=698635"},{"taxonomy":"msr-research-area","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/research-area?post=698635"},{"taxonomy":"msr-region","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-region?post=698635"},{"taxonomy":"msr-event-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-event-type?post=698635"},{"taxonomy":"msr-locale","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-locale?post=698635"},{"taxonomy":"msr-post-option","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-post-option?post=698635"},{"taxonomy":"msr-impact-theme","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-impact-theme?post=698635"},{"taxonomy":"msr-promo-type","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-promo-type?post=698635"},{"taxonomy":"msr-podcast-series","embeddable":true,"href":"https:\/\/www.microsoft.com\/en-us\/research\/wp-json\/wp\/v2\/msr-podcast-series?post=698635"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}