Orca 2: Teaching Small Language Models How to Reason
At Microsoft, we’re expanding AI capabilities by training small language models to achieve the kind of enhanced reasoning and comprehension typically found only in much larger models.
Skeleton-of-Thought: Parallel decoding speeds up and improves LLM output
This research was accepted by the 2024 International Conference on Learning Representations. Large language models (LLMs) such as LLaMA and OpenAI’s GPT-4 are revolutionizing technology. However, one of the common complaints about LLMs is their…
Research Focus: Week of November 8, 2023
Welcome to Research Focus, a series of blog posts that highlights notable publications, events, code/datasets, new hires and other milestones from across the research community at Microsoft. Generating both plausible and accurate full body avatar…
MEGA Multilingual Benchmarking
Official code for the paper published at EMNLP 2023 paper: Multilingual Evaluation of Generative AI (MEGA), a framework to evaluate Large Language Models (LLMs) on various multilingual benchmarks
Research Focus: Week of October 23, 2023
In this issue: Kosmos-2.5: A Multimodal Literate Model; Can vine copulas explain complex relationships of weather variables; New system accelerates the adaptive training process; Structural inequalities and relational labor in the influencer industry.