Facebook owner Meta Platforms Inc is testing its first in-house chip for training artificial intelligence (AI) systems, a key milestone as it moves to design more of its own custom silicon and reduce reliance on external suppliers like Nvidia Corp, two sources told Reuters.
The world’s biggest social media company has begun a small deployment of the chip and plans to ramp up production for wide-scale use if the test goes well, the sources said.
The push to develop in-house chips is part of a long-term plan at Meta to bring down its mammoth infrastructure costs, as the company places expensive bets on AI tools to drive growth.
Photo: Reuters
Meta, which also owns Instagram and WhatsApp, has forecast total expenses of US$114 billion to US$119 billion for this year, including up to US$65 billion in capital expenditure largely driven by spending on AI infrastructure.
One of the sources said Meta’s new training chip is a dedicated accelerator, meaning it is designed to handle only AI-specific tasks. This can make it more power-efficient than the integrated graphics processing units (GPUs) generally used for AI workloads.
Meta is working with Taiwan Semiconductor Manufacturing Co (TSMC, 台積電) to produce the chip, this person said.
The test deployment began after Meta finished its first “tape-out” of the chip, a significant marker of success in silicon development work that involves sending an initial design through a chip factory, the other source said.
A typical tape-out costs tens of millions of US dollars and takes roughly three to six months to complete, with no guarantee the test will succeed. A failure would require Meta to diagnose the problem and repeat the tape-out step.
Meta and TSMC declined to comment.
Meta executives have said they want to start using their own chips by next year for training, or the compute-intensive process of feeding the AI system reams of data to “teach” it how to perform.
As with the inference chip, the goal for the training chip is to start with recommendation systems and later use it for generative AI products like chatbot Meta AI, the executives said.
“We’re working on how would we do training for recommender systems and then eventually how do we think about training and inference for gen AI,” Meta chief product officer Chris Cox said at the Morgan Stanley technology, media and telecom conference last week.
Cox described Meta’s chip development efforts as “kind of a walk, crawl, run situation” so far, but said executives considered the first-generation inference chip for recommendations to be a “big success.”
Meta previously pulled the plug on an in-house custom inference chip after it flopped in a small-scale test deployment similar to the one it is doing now for the training chip, instead reversing course and placing orders for billions of dollars worth of Nvidia GPUs in 2022.
The social media company has remained one of Nvidia’s biggest customers since then, amassing an arsenal of GPUs to train its models, including for recommendations and ads systems and its Llama foundation model series. The units also perform inference for the more than 3 billion people who use its apps each day.
The value of those GPUs has been thrown into question this year, as AI researchers increasingly express doubts about how much more progress can be made by continuing to “scale up” large language models by adding ever more data and computing power.
Those doubts were reinforced with the late-January launch of new low-cost models from Chinese start-up DeepSeek (深度求索), which optimize computational efficiency by relying more heavily on inference than most incumbent models.
In a DeepSeek-induced global rout in AI stocks, Nvidia shares lost as much as a fifth of their value at one point. They subsequently regained most of that ground, with investors wagering the company’s chips will remain the industry standard for training and inference, although they have dropped again on broader trade concerns.
‘DECENT RESULTS’: The company said it is confident thanks to an improving world economy and uptakes in new wireless and AI technologies, despite US uncertainty Pegatron Corp (和碩) yesterday said it plans to build a new server manufacturing factory in the US this year to address US President Donald Trump’s new tariff policy. That would be the second server production base for Pegatron in addition to the existing facilities in Taoyuan, the iPhone assembler said. Servers are one of the new businesses Pegatron has explored in recent years to develop a more balanced product lineup. “We aim to provide our services from a location in the vicinity of our customers,” Pegatron president and chief executive officer Gary Cheng (鄭光治) told an online earnings conference yesterday. “We
LEAK SOURCE? There would be concern over the possibility of tech leaks if TSMC were to form a joint venture to operate Intel’s factories, an analyst said Taiwan Semiconductor Manufacturing Co (TSMC, 台積電) yesterday stayed mum after a report said that the chipmaker has pitched chip designers Nvidia Corp, Advanced Micro Devices Inc and Broadcom Inc about taking a stake in a joint venture to operate Intel Corp’s factories. Industry sources told the Central News Agency (CNA) that the possibility of TSMC proposing to operate Intel’s wafer fabs is low, as the Taiwanese chipmaker has always focused on its core business. There is also concern over possible technology leaks if TSMC were to form a joint venture to operate Intel’s factories, Concord Securities Co (康和證券) analyst Kerry Huang (黃志祺)
It was late morning and steam was rising from water tanks atop the colorful, but opaque-windowed, “soapland” sex parlors in a historic Tokyo red-light district. Walking through the narrow streets, camera in hand, was Beniko — a former sex worker who is trying to capture the spirit of the area once known as Yoshiwara through photography. “People often talk about this neighborhood having a ‘bad history,’” said Beniko, who goes by her nickname. “But the truth is that through the years people have lived here, made a life here, sometimes struggled to survive. I want to share that reality.” In its mid-17th to
‘MAKE OR BREAK’: Nvidia shares remain down more than 9 percent, but investors are hoping CEO Jensen Huang’s speech can stave off fears that the sales boom is peaking Shares in Nvidia Corp’s Taiwanese suppliers mostly closed higher yesterday on hopes that the US artificial intelligence (AI) chip designer would showcase next-generation technologies at its annual AI conference slated to open later in the day. The GPU Technology Conference (GTC) in California is to feature developers, engineers, researchers, inventors and information technology professionals, and would focus on AI, computer graphics, data science, machine learning and autonomous machines. The event comes at a make-or-break moment for the firm, as it heads into the next few quarters, with Nvidia CEO Jensen Huang’s (黃仁勳) keynote speech today seen as having the ability to