
Meta Platforms opens new tab unveiled details on Wednesday about the next generation of the company’s in-house artificial intelligence accelerator chip.
“This chip’s architecture is fundamentally focused on providing the right balance of compute, memory bandwidth, and memory capacity for serving ranking and recommendation models,” the company wrote in a blog post.
The new Meta Training and Inference Accelerator (MTIA) chip is part of a broad custom silicon effort at the company that includes looking at other hardware systems too. Beyond building the chips and hardware, Meta has made significant investments in developing the software necessary to harness the power of its infrastructure in the most efficient way.
The company is also spending billions on buying Nvidia and other AI chips: This year CEO Mark Zuckerberg said the company planned to acquire roughly 350,000 flagship H100 chips from Nvidia. Combined with other suppliers, Meta plans to accumulate the equivalent of 600,000 H100 chips this year, he said.
Taiwan Semiconductor Manufacturing Co opens new tab will produce the new chip on its “5nm” process. Meta said it is capable of three times the performance of its first-generation processor.
The chip has been deployed in the data centre and is engaged in serving AI applications. The company said it has several programs underway “aimed at expanding the scope of MTIA, including support of (generative AI) workloads.”