Microsoft Research Blog

English

  1. Differentially Private Synthetic Data via Foundation Model APIs 1: Images 

    May 7, 2024

    Generating differentially private (DP) synthetic data that closely resembles the original private data without leaking sensitive user information is a scalable way to mitigate privacy concerns in the current data-driven world. In contrast to current practices that train customized models for this task, we aim…

  2. Microsoft goes from bad boy to top cop in the age of AI 

    May 7, 2024 | Juan M. Lavista Ferres

    Alongside efforts to use artificial intelligence to find new cures for cancer and combat climate change, the Microsoft AI for Good's small engineering team has another job: figuring out how to detect AI-powered deepfake videos, audio clips and images bombarding elections worldwide. So far, Lavista…

  3. Privately Aligning Language Models with Reinforcement Learning 

    May 7, 2024

    Positioned between pre-training and user deployment, aligning large language models (LLMs) through reinforcement learning (RL) has emerged as a prevailing strategy for training instruction following-models such as ChatGPT. In this work, we initiate the study of privacy-preserving alignment of LLMs through Differential Privacy (DP) in…

  4. ICLR conference banner - abstract shapes

    Microsoft at ICLR 2024 

    May 7, 2024

    Microsoft is proud to be a sponsor of The International Conference on Learning Representatives (ICLR) (opens in new tab). This premier gathering of professionals is dedicated to the advancement of the branch of artificial intelligence called representation learning. ICLR is globally renowned for presenting and…

  5. Unifying Feature and Cost Aggregation with Transformers for Dense Correspondence 

    May 7, 2024 | Sunghwan Hong, Seokju Cho, Seungryong Kim, and Stephen Lin

    This paper introduces a Transformer-based integrative feature and cost aggregation network designed for dense matching tasks. In the context of dense matching, many works benefit from one of two forms of aggregation: feature aggregation, which pertains to the alignment of similar features, or cost aggregation,…

  6. You Only Cache Once: Decoder-Decoder Architectures for Language Models 

    May 7, 2024

    We introduce a decoder-decoder architecture, YOCO, for large language models, which only caches key-value pairs once. It consists of two components, i.e., a cross-decoder stacked upon a self-decoder. The self-decoder efficiently encodes global key-value (KV) caches that are reused by the cross-decoder via cross-attention. The…