Welcome to WarBulletin - your new best friend in the world of gaming. We're all about bringing you the hottest updates and juicy insights from across the gaming universe. Are you into epic RPG adventures or fast-paced eSports? We've got you covered with the latest scoop on everything from next-level PC gaming rigs to the coolest game releases. But hey, we're more than just news! Ever wondered what goes on behind the scenes of your favorite games? We're talking exclusive interviews with the brains behind the games, fresh off-the-press photos and videos straight from gaming conventions, and, of course, breaking news that you just can't miss. We know you love gaming 24/7, and that's why we're here round the clock, updating you on all things gaming. Whether it's the lowdown on a new patch or the buzz about the next big gaming celeb, we're on it.

Contacts

  • Owner: SNOWLAND s.r.o.
  • Registration certificate 06691200
  • 16200, Na okraji 381/41, Veleslavín, 162 00 Praha 6
  • Czech Republic

OpenAI plans to build its own AI chips on TSMC's forthcoming 1.6 nm A16 process node

ChatGPT developer OpenAI has been musing over building its own AI chips for some time now but it looks like the project is definitely going ahead, as United Daily News reports the company is paying TSMC to make the new chips. But rather than using its current N4 or N3 process nodes, OpenAI has booked production slots for the 1.6 nm, so-called A16, process node.

The report from UDN (via Wccftech) doesn't provide any concrete evidence for this claim but the Taiwanese news agency is usually pretty accurate when it comes to tech forecasts like this. At the moment, OpenAI spends vast amounts of money to run ChatGPT, in part due to the very high cost of Nvidia's AI servers.

Nvidia's hardware dominates the industry, with Alphabet, Amazon, Meta, Microsoft, and Tesla spending hundreds of millions of dollars on its Hopper H100 and Blackwell superchips. While the cost of designing and developing a competitive AI chip is just as expensive, once you have a working product, the ongoing costs are much lower.

UDN suggests that OpenAI had originally planned to use TSMC's relatively low-cost N5 process node to manufacture its AI chip but that's apparently been dropped in favour of a system that's still in development—A16 will be the successor to N2, which itself isn't being used to mass produce chips yet.

TSMC states that A16 is a 1.6 nm node but the number itself is fairly meaningless now. It will use the same gate-all-around (GAAFET) nanosheet transistors as N2 but will be the first TSMC node to employ backside power delivery, called Super Power Rail.

But why would OpenAI want to use something that's still a few years away from being ready for bulk orders? UDN's report suggests that OpenAI has approached Broadcom and Marvell to handle the development of the AI chips but neither company has much experience with TSMC's cutting-edge nodes, as far as I know.

One possibility is that the whole project is being done in collaboration with Apple, which uses ChatGPT in its own AI system.

Read more on pcgamer.com