Microsoft Research Blog
Orca 2: Teaching Small Language Models How to Reason
At Microsoft, we’re expanding AI capabilities by training small language models to achieve the kind of enhanced reasoning and comprehension typically found only in much larger models.
Microsoft Research Blog
Skeleton-of-Thought: Parallel decoding speeds up and improves LLM output
This research was accepted by the 2024 International Conference on Learning Representations. Large language models (LLMs) such as LLaMA and OpenAI’s GPT-4 are revolutionizing technology. However, one of the common complaints about LLMs is their…
Microsoft Research Blog
Research Focus: Week of November 8, 2023
Welcome to Research Focus, a series of blog posts that highlights notable publications, events, code/datasets, new hires and other milestones from across the research community at Microsoft. Generating both plausible and accurate full body avatar…