ChatGPT developer OpenAI has been musing over building its own AI chips for some time now but it looks like the project is definitely going ahead, as United Daily News reports the company is paying TSMC to make the new chips. But rather than using its current N4 or N3 process nodes, OpenAI has booked production slots for the 1.6 nm, so-called A16, process node.
The report from UDN (via Wccftech) doesn't provide any concrete evidence for this claim but the Taiwanese news agency is usually pretty accurate when it comes to tech forecasts like this. At the moment, OpenAI spends vast amounts of money to run ChatGPT, in part due to the very high cost of Nvidia's AI servers.
Nvidia's hardware dominates the industry, with Alphabet, Amazon, Meta, Microsoft, and Tesla spending hundreds of millions of dollars on its Hopper H100 and Blackwell superchips. While the cost of designing and developing a competitive AI chip is just as expensive, once you have a working product, the ongoing costs are much lower.
UDN suggests that OpenAI had originally planned to use TSMC's relatively low-cost N5 process node to manufacture its AI chip but that's apparently been dropped in favour of a system that's still in development—A16 will be the successor to N2, which itself isn't being used to mass produce chips yet.
TSMC states that A16 is a 1.6 nm node but the number itself is fairly meaningless now. It will use the same gate-all-around (GAAFET) nanosheet transistors as N2 but will be the first TSMC node to employ backside power delivery, called Super Power Rail.
But why would OpenAI want to use something that's still a few years away from being ready for bulk orders? UDN's report suggests that OpenAI has approached Broadcom and Marvell to handle the development of the AI chips but neither company has much experience with TSMC's cutting-edge nodes, as far as I know.
One possibility is that the whole project is being done in collaboration with Apple, which uses ChatGPT in its own AI system.
Read more on pcgamer.com