Microsoft unveils 'largest ever' AI natural language model
T-NLG has over twice as many parameters as Nvidia’s MegatronLM
Microsoft has revealed its largest deep learning language model, the Turing Natural Language Generation (T-NLG), which is claimed to have a record-breaking 17 billion parameters.
The T-NLG, according to Microsoft, outperforms the largest deep learning models to date: the University of Washington’s Grover-Mega and Nvidia’s MegatronLM, which possess 1.5 and 8.3 billion parameters, respectively.
According to Microsoft, the T-NLG is capable of completing unfinished sentences, as well as generating direct answers to questions and can create summaries of documents fed into it.
Microsoft also claims that the model has the ability to directly answer the question with a complete sentence.
“This capability is more important outside of web search—for example, this can power AI assistants to intelligently respond when a user asks a question about their personal data such as emails or Word documents,” explained Microsoft’s applied scientist Corby Rosset.
He also thanked the DeepSpeed Library and the ZeRO optimiser for producing “breakthroughs” without which “this work would not be possible”.
Providing an example of how the T-NLG works, the language model introduced itself by generating a summary of its skills:
Get the ITPro. daily newsletter
Receive our latest news, industry updates, featured resources and more. Sign up today to receive our FREE report on AI cyber crime & security - newly updated for 2024.
“Turing Natural Language Generation (T-NLG) is a 17 billion parameter language model by Microsoft that outperforms the state of the art on many downstream NLP tasks. We present a demo of the model, including its freeform generation, question answering, and summarization capabilities, to academics for feedback and research purposes,” said the T-NLG.
Having only graduated from City University in 2019, Sabina has already demonstrated her abilities as a keen writer and effective journalist. Currently a content writer for Drapers, Sabina spent a number of years writing for ITPro, specialising in networking and telecommunications, as well as charting the efforts of technology companies to improve their inclusion and diversity strategies, a topic close to her heart.
Sabina has also held a number of editorial roles at Harper's Bazaar, Cube Collective, and HighClouds.