Pretraining is the preliminary and fundamental step in developing capable language models (LM). Despite this, pretraining data design is critically under-documented and often guided by empirically unsupported intuitions. To address this, we pretrain 28 1.5B…
The Microsoft Search, Assistant, and Intelligence (MSAI) group is looking for a Principal Applied Scientist who can bring deep applied science expertise and a track record of shipping at-scale, AI-enabled, intelligent user-facing experiences to drive…
The Microsoft Azure AI platform team is looking for a Principal Applied Scientist with a background in large language modeling. You will work on adapting and grounding large language models for product-driven scenarios. The focus…