Microsoft Research Blog

Artificial intelligence

  1. SciAI4Industry – Solving PDEs for industry-scale problems with deep learning 

    November 23, 2022

    Solving partial differential equations with deep learning makes it possible to reduce simulation times by multiple orders of magnitude and unlock scientific methods that typically rely on large numbers of sequential simulations, such as optimization and uncertainty quantification. Two of the largest challenges of adopting…

  2. An Empirical Study of Log Analysis at Microsoft 

    November 9, 2022

    Logs are crucial to the management and maintenance of software systems. In recent years, log analysis research has achieved notable progress on various topics such as log parsing and log-based anomaly detection. However, the real voices from front-line practitioners are seldom heard. For example, what…

  3. Principled Offline RL with Rich Exogenous Information 

    October 31, 2022

    Learning to control an agent from data collected offline in a rich pixel-based visual observation space is vital for real-world applications of reinforcement learning (RL). A major challenge in this setting is the presence of input information that is hard to model and irrelevant to…

  4. Contrastive Decoding: Open-ended Text Generation as Optimization 

    October 27, 2022

    Likelihood, although useful as a training loss, is a poor search objective for guiding open-ended generation from language models (LMs). Existing generation algorithms must avoid both unlikely strings, which are incoherent, and highly likely ones, which are short and repetitive. We propose contrastive decoding (CD),…

  5. Effective and Efficient Query-aware Snippet Extraction for Web Search 

    October 27, 2022

    Query-aware webpage snippet extraction is widely used in search engines to help users better understand the content of the returned webpages before clicking. Although important, it is very rarely studied. In this paper, we propose an effective query-aware webpage snippet extraction method named DeepQSE, aiming…

  6. COMPASS: Contrastive Multimodal Pretraining for Autonomous Systems 

    October 1, 2022

    Learning representations that generalize across tasks and domains is challenging yet necessary for autonomous systems. Although task-driven approaches are appealing, designing models specific to each application can be difficult in the face of limited data, especially when dealing with highly variable multimodal input spaces arising…

  7. A new Workflow for Human-AI Collaboration in Citizen Science 

    October 1, 2022

    The unprecedented growth of online citizen science projects provides growing opportunities for the public to participate in scientific discoveries. Nevertheless, volunteers typically make only a few contributions before exiting the system. Thus a significant challenge to such systems is increasing the capacity and efficiency of…

  8. Overwatch: Learning Patterns in Code Edit Sequences 

    October 1, 2022

    Integrated Development Environments (IDEs) provide tool support to automate many source code editing tasks. Traditionally, IDEs use only the spatial context, i.e., the location where the developer is editing, to generate candidate edit recommendations. However, spatial context alone is often not sufficient to confidently predict…

  9. Making the Most of Text Semantics to Improve Biomedical Vision-Language Processing 

    October 1, 2022

    ABSTRACT: Multi-modal data abounds in biomedicine, such as radiology images and reports. Interpreting this data at scale is essential for improving clinical care and accelerating clinical research. Biomedical text with its complex semantics poses additional challenges in vision-language modelling compared to the general domain, and…

  10. PACT: Perception-Action Causal Transformer for Autoregressive Robotics Pre-Training 

    September 22, 2022

    Robotics has long been a field riddled with complex systems architectures whose modules and connections, whether traditional or learning-based, require significant human expertise and prior knowledge. Inspired by large pre-trained language models, this work introduces a paradigm for pre-training a general purpose representation that can…

  11. What is it like to program with artificial intelligence? 

    September 1, 2022

    Large language models, such as OpenAI's codex and Deepmind's AlphaCode, can generate code to solve a variety of problems expressed in natural language. This technology has already been commercialised in at least one widely-used programming editor extension: GitHub Copilot. In this paper, we explore how…