Microsoft Research Blog

Artificial intelligence

  1. Learning Multi-Task Action Abstractions as a Sequence Compression Problem 

    November 6, 2023 | Ruijie Zheng, Ching-An Cheng, Furong Huang, and Andrey Kolobov

    Temporal abstractions, along with belief state representations, have long been recognized as a powerful knowledge sharing mechanism for decision-making scenarios ranging from computer games to robotics. In this work, we propose a novel approach that views inducing temporal action abstractions as sequence compression. In doing…

  2. The MineRL BASALT Competition on Learning from Human Feedback 

    November 1, 2023

    The last decade has seen a significant increase of interest in deep learning research, with many public successes that have demonstrated its potential. As such, these systems are now being incorporated into commercial products. With this comes an additional challenge: how can we build AI…

  3. Interactive Robot Learning from Verbal Correction 

    November 1, 2023

    The ability to learn and refine behavior after deployment has become ever more important for robots as we design them to operate in unstructured environments like households. In this work, we design a new learning system based on large language model (LLM), OLAF, that allows…

  4. TSTR: Target Similarity Tuning Meets the Real World 

    November 1, 2023

    Target similarity tuning (TST) is a method of selecting relevant examples in natural language (NL) to code generation through large language models (LLMs) to improve performance. Its goal is to adapt a sentence embedding model to have the similarity between two NL inputs match the…

  5. Evaluating General-Purpose AI with Psychometrics 

    October 25, 2023

    Comprehensive and accurate evaluation of general-purpose AI systems such as large language models allows for effective mitigation of their risks and deepened understanding of their capabilities. Current evaluation methodology, mostly based on benchmarks of specific tasks, falls short of adequately assessing these versatile AI systems,…

  6. Microscaling Data Formats for Deep Learning 

    October 19, 2023

    Narrow bit-width data formats are key to reducing the computational and storage costs of modern deep learning applications. This paper evaluates Microscaling (MX) data formats that combine a per-block scaling factor with narrow floating-point and integer types for individual elements. MX formats balance the competing…

  7. Joint Prompt Optimization of Stacked LLMs using Variational Inference 

    October 1, 2023

    We view large language models (LLMs) as stochastic language layers in a network, where the learnable parameters are the natural language prompts at each layer. We stack two such layers, feeding the output of one layer to the next. We call the stacked architecture a…

  8. Grounded Copilot: How Programmers Interact with Code-Generating Models 

    October 1, 2023 | Shraddha Barke, Michael James, and Nadia Polikarpova

    Powered by recent advances in code-generating models, AI assistants like Github Copilot promise to change the face of programming forever. But what is this new face of programming? We present the first grounded theory analysis of how programmers interact with Copilot, based on observing 20…