Microsoft Research Blog

Artificial intelligence

  1. Analyzing the Nuances of Transformers’ Polynomial Simplification Abilities 

    May 7, 2021 | Vishesh Agarwal, Somak Aditya, and Navin Goyal (navingo)

    Symbolic Mathematical tasks such as integration often require multiple well-defined steps and understanding of sub-tasks to reach a solution. To understand Transformers' abilities in such tasks in a fine-grained manner, we deviate from traditional end-to-end settings, and explore a step-wise polynomial simplification task. Polynomials can…

  2. Training Structured Mechanical Models by Minimizing Discrete Euler-Lagrange Residual 

    May 4, 2021 | Kunal Menda, Jayesh K. Gupta, Zachary Manchester, and Mykel J. Kochenderfer

    Model-based paradigms for decision-making and control are becoming ubiquitous in robotics. They rely on the ability to efficiently learn a model of the system from data. Structured Mechanical Models (SMMs) are a data-efficient black-box parameterization of mechanical systems, typically fit to data by minimizing the…

  3. What Makes Instance Discrimination Good for Transfer Learning 

    May 3, 2021 | Nanxuan Zhao, Zhirong Wu, Rynson W. H. Lau, and Stephen Lin

    Contrastive visual pretraining based on the instance discrimination pretext task has made significant progress. Notably, recent work on unsupervised pretraining has shown to surpass the supervised counterpart for finetuning downstream applications such as object detection and segmentation. It comes as a surprise that image annotations…

  4. Taking Notes on the Fly Helps Language Pre-Training 

    May 3, 2021

    How to make unsupervised language pre-training more efficient and less resource-intensive is an important research direction in NLP. In this paper, we focus on improving the efficiency of language pre-training methods through providing better data utilization. It is well-known that in language data corpus, words…

  5. A Unified Bayesian Framework for Discriminative and Generative Continual Learning 

    May 3, 2021 | Abhishek Kumar, Sunabha Chatterjee, and Piyush Rai

    Continual Learning is a learning paradigm where learning systems are trained on a sequence of tasks. The goal here is to perform well on the current task without suffering from a performance drop on the previous tasks. Two notable directions among the recent advances in…

  6. Taking Notes on the Fly Helps Language Pre-Training 

    May 3, 2021

    How to make unsupervised language pre-training more efficient and less resource-intensive is an important research direction in NLP. In this paper, we focus on improving the efficiency of language pre-training methods through providing better data utilization. It is well-known that in language data corpus, words…