Microsoft is Reportedly Working on AI Chips to Train LLMs

Microsoft

Microsoft is reportedly working on its own AI chips that can be used to train LLMs

After investing billions in the firm that created ChatGPT, OpenAI, Microsoft is reportedly working on its own AI chips that can be used to train large language models or LLMs. Microsoft is making waves in the artificial intelligence (AI) space.

According to The Information, which cited two sources with firsthand knowledge of the project, the software behemoth has been working on the microprocessor since as early as 2019. A limited number of Microsoft and OpenAI personnel who are evaluating the technology reportedly already have access to it under the code name Athena.

Microsoft hopes that the chip would outperform those it now purchases from other suppliers, saving it time and money on its expensive AI projects. According to the study, other well-known tech firms like Amazon, Google, and Facebook also produce their own AI processors.

For the AI research firm OpenAI, Microsoft has already constructed a supercomputer that can train vast numbers of models. To support ChatGPT and the Bing AI chatbot, the business uses tens of thousands of Nvidia A100 graphics chips connected to a network for the supercomputer. A “massive, cutting-edge supercomputer” was going to be built, thus it invested US$1 billion in OpenAI in 2019.

Microsoft created this supercomputer to have the processing capacity necessary to train and retrain an expanding number of AI models over extended periods using massive amounts of data.

Nidhi Chappell, Microsoft’s head of product for Azure high-performance computing and AI, stated, “One of the things we had learned from research is that the larger the model, the more data you have, and the longer you can train, the greater the accuracy of the model is.

Google’s TPU AI Chip

Google revealed that it has created an AI chip dubbed the Tensor Processing Unit (TPU) that was made exclusively for machine learning activities last year. According to claims, the TPU uses little power and can do billions of processes per second.

TensorFlow, Google’s open-source machine learning software framework, is intended to be used with the tensor processing unit.

The post Microsoft is Reportedly Working on AI Chips to Train LLMs appeared first on Analytics Insight.

Follow us on Twitter, Facebook
0 0 votes
Article Rating
Subscribe
Notify of
guest
0 comments
Oldest
New Most Voted
Inline Feedbacks
View all comments

Latest stories

You might also like...