Loading...

In the news | The Register

Meet Clippy 9000: Microsoft brags about building Earth’s largest AI language model, refuses to let it out of the lab 

February 11, 2020

There’s a new giant AI language model in town: enter Microsoft’s Turing-NLG system, which apparently contains a whopping 17 billion parameters, making it the largest publicly known model of its class yet.

In the news | Neowin

Microsoft builds the world’s largest transformer-based language generation model 

February 11, 2020

Transformer-based language generation models have enabled better conversational applications. Though they still have their shortcomings, which were recently exposed by a team at MIT, researchers continue improving them to build better, larger, and more robust models.

In the news | WinBuzzer

Microsoft’s New Turing NLG is the Largest Transformer Language Model 

February 11, 2020

Microsoft has developed a Transformer-based language generation model that it describes as the largest ever made. This week, Microsoft AI & Research announced Turing NLG, which is twice the size of its nearest competitor.

In the news | WinBuzzer

Microsoft DeepSpeed with Zero Can Train 100 Billion Parameter AI Models 

February 11, 2020

Microsoft has released a new open-source library called DeepSpeed, which, when combined with its ‘ZeRO’ module can train 100 billion parameter models without using the resources traditionally associated with that.

In the news | ITPro

Microsoft unveils ‘largest ever’ AI natural language model 

February 11, 2020

Microsoft has revealed its largest deep learning language model, the Turing Natural Language Generation (T-NLG), which is claimed to have a record-breaking 17 billion parameters. The T-NLG, according to Microsoft, outperforms the largest deep learning models to date: the University of Washington’s Grover-Mega and Nvidia’s MegatronLM, which…

In the news | MSPoweruser

Meet Microsoft DeepSpeed, a new deep learning library that can train massive 100-billion-parameter models 

February 10, 2020

Microsoft Research today announced DeepSpeed, a new deep learning optimization library that can train massive 100-billion-parameter models. In AI, you need to have larger natural language models for better accuracy. But training larger natural language models is time consuming and…

a man wearing a suit and tie

In the news | Future Decoded Mumbai CEO Summit

Satya talks about HAMS in his keynote during his visit to India 

February 10, 2020

In the news | VentureBeat

Microsoft trains world’s largest Transformer language model 

February 10, 2020

Microsoft AI & Research today shared what it calls the largest Transformer-based language generation model ever and open-sourced a deep learning library named DeepSpeed to make distributed training of large models easier.

In the news | InfoWorld

Microsoft speeds up PyTorch with DeepSpeed 

February 10, 2020

Microsoft has released DeepSpeed, a new deep learning optimization library for PyTorch, that is designed to reduce memory use and train models with better parallelism on existing hardware.

  • Previous
  • 1
  • …
  • 272
  • 273
  • 274
  • 275
  • 276
  • …
  • 569
  • Next