Microsoft Research Blog

English

  1. Mitigate Position Bias in Large Language Models via Scaling a Single Dimension 

    June 4, 2024

    Large Language Models (LLMs) are increasingly applied in various real-world scenarios due to their excellent generalization capabilities and robust generative abilities. However, they exhibit position bias, also known as"lost in the middle", a phenomenon that is especially pronounced in long-context scenarios, which indicates the placement…

  2. Measuring and shaping the nutritional environment via food sales logs: case studies of campus-wide food choice and a call to action 

    June 3, 2024

    Although diets influence health and the environment, measuring and changing nutrition is challenging. Traditional measurement methods face challenges, and designing and conducting behavior-changing interventions is conceptually and logistically complicated. Situated local communities such as university campuses offer unique opportunities to shape the nutritional environment and…

  3. Leveraging Visual Tokens for Extended Text Contexts in Multi-Modal Learning 

    June 3, 2024

    Training models with longer in-context lengths is a significant challenge for multimodal model due to substantial GPU memory and computational costs. This exploratory study does not present state-of-the-art models; rather, it introduces an innovative method designed to increase in-context text length in multi-modality large language…

  4. Microsoft at FAccT 2024 event header | abstract background pattern

    FAccT 2024 

    June 3, 2024

    Microsoft is honored to support the ACM Conference on Fairness, Accountability, and Transparency (ACM FAccT (opens in new tab)). This pivotal computer science conference has a multidisciplinary approach, uniting researchers and practitioners who are dedicated to advancing fairness, accountability, and transparency within socio-technical systems. Program…

  5. SLTrain: a sparse plus low-rank approach for parameter and memory efficient pretraining 

    June 3, 2024

    Large language models (LLMs) have shown impressive capabilities across various tasks. However, training LLMs from scratch requires significant computational power and extensive memory capacity. Recent studies have explored low-rank structures on weights for efficient fine-tuning in terms of parameters and memory, either through low-rank adaptation…

  6. MedFuzz: Exploring the Robustness of Large Language Models in Medical Question Answering 

    June 3, 2024

    Large language models (LLM) have achieved impressive performance on medical question-answering benchmarks. However, high benchmark accuracy does not imply that the performance generalizes to real-world clinical settings. Medical question-answering benchmarks rely on assumptions consistent with quantifying LLM performance but that may not hold in the…

  7. MEDIQ: Question-Asking LLMs for Adaptive and Reliable Clinical Reasoning 

    June 2, 2024

    In high-stakes domains like clinical reasoning, AI assistants powered by large language models (LLMs) are yet to be reliable and safe. We identify a key obstacle towards reliability: existing LLMs are trained to answer any question, even with incomplete context in the prompt or insufficient…