Megatron gets a tune-up courtesy of Microsoft and NVIDIA

Spread the love

The Visitor's Center at Microsoft Headquarters campus is pictured July 17, 2014 in Redmond, Washington.Source: Stephen Brashear/Getty Images for Microsoft

Megatron-Turing NLG 530B (MT-NLG), the AI language model succeeding the Turing NLG 17B and Megatron-LM, has been described by NVIDIA and Microsoft as the “world’s largest and most powerful generative language model.”

To give a very brief primer on what language models are: They’re tools used to anticipate word choice. This sort of tech can help the services you use to identify spam in an email inbox or figure out what word was spoken in a video, all so that you don’t have to. As you can imagine, crafting these AI language models is no small feat given the complexity of the task at hand. That’s why NVIDIA and Microsoft are proud to present Megatron-Turing NLG 530B. The companies say that MT-NLG has unmatched accuracy in the following areas:

VPN Deals: Lifetime license for $16, monthly plans at $1 & more

  • Completion prediction
  • Reading comprehension
  • Commonsense reasoning
  • Natural language inferences
  • Word sense disambiguation

With that being said, it’s not perfect. The language model isn’t devoid of bias, which is derived from the data it uses to learn how to perform its tasks. Similarly, the model is also influenced by toxicity, according to NVIDIA’s report on the creation.

If you can put those pitfalls aside, it’s still worth noting that MT-NLG remains “the largest and the most powerful monolithic transformer language model trained to date, with 530 billion parameters,” as stated by the NVIDIA report. Think about Megatron the next time you let an email write itself for you. And don’t mistake it with Pegatron.

We may earn a commission for purchases using our links. Learn more.

Leave a Reply