Loading...

In the news | The Register

Meet Clippy 9000: Microsoft brags about building Earth’s largest AI language model, refuses to let it out of the lab 

February 11, 2020

There’s a new giant AI language model in town: enter Microsoft’s Turing-NLG system, which apparently contains a whopping 17 billion parameters, making it the largest publicly known model of its class yet.

In the news | Neowin

Microsoft builds the world’s largest transformer-based language generation model 

February 11, 2020

Transformer-based language generation models have enabled better conversational applications. Though they still have their shortcomings, which were recently exposed by a team at MIT, researchers continue improving them to build better, larger, and more robust models.

In the news | WinBuzzer

Microsoft’s New Turing NLG is the Largest Transformer Language Model 

February 11, 2020

Microsoft has developed a Transformer-based language generation model that it describes as the largest ever made. This week, Microsoft AI & Research announced Turing NLG, which is twice the size of its nearest competitor.

In the news | WinBuzzer

Microsoft DeepSpeed with Zero Can Train 100 Billion Parameter AI Models 

February 11, 2020

Microsoft has released a new open-source library called DeepSpeed (opens in new tab), which, when combined with its ‘ZeRO’ module can train 100 billion parameter models without using the resources traditionally associated with that.

In the news | ITPro

Microsoft unveils ‘largest ever’ AI natural language model 

February 11, 2020

Microsoft has revealed its largest deep learning language model, the Turing Natural Language Generation (T-NLG), which is claimed to have a record-breaking 17 billion parameters. The T-NLG, according to Microsoft, outperforms the largest deep learning models to date: the University of Washington’s Grover-Mega and Nvidia’s MegatronLM, which…

In the news | MSPoweruser

Meet Microsoft DeepSpeed, a new deep learning library that can train massive 100-billion-parameter models 

February 10, 2020

Microsoft Research today announced DeepSpeed, a new deep learning optimization library that can train massive 100-billion-parameter models. In AI, you need to have larger natural language models for better accuracy. But training larger natural language models is time consuming and…

a man wearing a suit and tie

In the news | Future Decoded Mumbai CEO Summit

Satya talks about HAMS in his keynote during his visit to India 

February 10, 2020

In the news | VentureBeat

Microsoft trains world’s largest Transformer language model 

February 10, 2020

Microsoft AI & Research today shared what it calls the largest Transformer-based language generation model ever and open-sourced a deep learning library named DeepSpeed to make distributed training of large models easier.

In the news | InfoWorld

Microsoft speeds up PyTorch with DeepSpeed 

February 10, 2020

Microsoft has released DeepSpeed, a new deep learning optimization library for PyTorch, that is designed to reduce memory use and train models with better parallelism on existing hardware.

  • Previous
  • 1
  • …
  • 279
  • 280
  • 281
  • 282
  • 283
  • …
  • 576
  • Next