It would be an understatement to say that AI models have taken the world by storm. They are everywhere—in your phone, the car you drive, the washing machine you use, and even in the video games you play. It all started with the launch of ChatGPT by OpenAI, and since then, multiple other big tech companies, including Meta, Google, and Anthropic, have launched their own AI models and subsequently, bots based on them. But why have only select companies been able to create these AI bots? Is it only a research thing? The answer is no!
AI companies need large capital, and they also need access to massive computing power to train their respective models. Now, where does the “compute” come from? Graphics Processing Units, or GPUs, as you know them. Yes, the same hardware that you may use to power some intensive video games on your PC.
This is exactly why NVIDIA has witnessed exponential growth, making it the number one company by market cap in the world, even beating Apple and Microsoft. It continues to trade places and comfortably sits in the top five now.
So, why do AI companies need GPUs? Let's answer this here in this explainer.
Also Read: Motorola Edge 50 AI features: Know what advanced features smartphone has to offer
The short answer: GPUs are just better at calculating and have better energy efficiency compared to CPUs, and over time, the performance for the money you pay has gradually increased—making them the ideal choice.
The long answer: NVIDIA, which makes GPUs for the likes of OpenAI, Meta, Oracle, Tesla, and more, says that current GPUs can do parallel processing, scale it to supercomputing levels, and the overall software scope for AI is quite broad.
And this is exactly why market leaders like OpenAI trained their Large Language Models using thousands of NVIDIA GPUs. To put it simply, these AI models, at their core, are neural networks, as per NVIDIA, and are made using layers and layers of equations, with one data piece being related to another. This is where
Read more on tech.hindustantimes.com