Microsoft Research Blog

English

  1. A summary of insights extracted by using the Eureka framework, shown via two radar charts for multimodal (left) and language (right) capabilities respectively. The radar charts show the best and worst performance observed for each capability.

    Eureka: Evaluating and understanding progress in AI 

    September 17, 2024

    How can we rigorously evaluate and understand state-of-the-art progress in AI? Eureka is an open-source framework for standardizing evaluations of large foundation models, beyond single-score reporting and rankings. Learn more about the extended findings.

  2. AMEGO: Active Memory from long EGOcentric videos 

    September 17, 2024 | Gabriele Goletto, Tushar Nagarajan, Giuseppe Averta, and D. Damen

    Egocentric videos provide a unique perspective into individuals' daily experiences, yet their unstructured nature presents challenges for perception. In this paper, we introduce AMEGO, a novel approach aimed at enhancing the comprehension of very-long egocentric videos. Inspired by the human's ability to maintain information from…

  3. Principal Type Inference under a Prefix (TR) 

    September 17, 2024 | Daan Leijen and Wenjia Ye

    At the heart of the Damas-Hindley-Milner (HM) type system lies the abstraction rule which derives a function type for a lambda expression. This rule allows the type of the parameter to be "guessed", which allows for multiple possible types for functions like the identity function.…

  4. EUREKA: Evaluating and Understanding Large Foundation Models 

    September 17, 2024

    Rigorous and reproducible evaluation of large foundation models is critical for assessing the state of the art, informing next steps in model improvement, and for guiding scientific advances in Artificial Intelligence (AI). Evaluation is also important for informing the increasing number of application developers that…

  5. RetrievalAttention One Page

    RetrievalAttention: Accelerating Long-Context LLM Inference via Vector Retrieval 

    September 16, 2024

    Transformer-based Large Language Models (LLMs) have become increasingly important. However, due to the quadratic time complexity of attention computation, scaling LLMs to longer contexts incurs extremely slow inference speed and high GPU memory consumption for caching key-value (KV) vectors. This paper proposes RetrievalAttention, a training-free…

  6. Research Focus | September 9, 2024

    Research Focus: Week of September 9, 2024 

    September 12, 2024

    Investigating vulnerabilities in LLMs; A novel total-duration-aware (TDA) duration model for text-to-speech (TTS); Generative expert metric system through iterative prompt priming; Integrity protection in 5G fronthaul networks:

  7. Physiological feedback for predictive models 

    September 12, 2024

    US Patent App. 18/118,849. This document relates to employing biosignals to evaluate predictions made by predictive models. For example, user attention can be inferred from a user attention signal such as gaze. When the user directs attention to a prediction output by a given predictive…