The Facebook parent said that it is working on a new AI-optimized data center design and the second phase of its 16,000 GPU supercomputer for AI research. Credit: Sdecoret / Getty Images Facebook parent company Meta has revealed plans for the development of its own custom chip for running artifical intelligence models, and a new data center architecture for AI workloads. “We are executing on an ambitious plan to build the next generation of Meta’s AI infrastructure and today, we’re sharing some details on our progress. This includes our first custom silicon chip for running AI models, a new AI-optimized data center design and the second phase of our 16,000 GPU supercomputer for AI research,” Santosh Janardhan, head of infrastructure at Meta, wrote in a blog post Thursday. Meta’s custom chip for running AI models, called Meta Training and Inference Accelerator (MTIA), is designed to provide greater compute power and efficiency than CPUs on the market today, according to Janardhan. MTIA is customized for internal workloads such as content understanding, feeds, generative AI, and ad ranking, the company said, adding that the first version of the chip was designed in 2020. Meta’s announcement of the strides it is making to produce its own custom chips for running AI models comes at a time when other large technology companies — driven by the proliferation of large language models and generative AI —are either working on or have already launched their own chips for AI workloads Earlier this month, news reports claimed that Microsoft was working with chip-maker AMD to develop its own chip for running AI workloads. AWS has also released its own chip for running AI workloads. On its part, Meta also said Thursday that its new data center design will be optimized to train AI models, a process that enables them to better their performance as they ingest more data.. “This new data center will be an AI-optimized design, supporting liquid-cooled AI hardware and a high-performance AI network connecting thousands of AI chips together for data center-scale AI training clusters,” Janardhan wrote, adding that the new data center systems will be faster and more cost-effective to build than earlier facilities. In addition to the new data center design, the company said that it was working on developing AI supercomputers that will support training of next-generation AI models, power augmented reality tools, and support real-time translation technology. ENDS Related content news Nvidia to power India’s AI factories with tens of thousands of AI chips India’s cloud providers and server manufacturers plan to boost Nvidia GPU deployment nearly tenfold by the year’s end compared to 18 months ago. By Prasanth Aby Thomas Oct 24, 2024 5 mins GPUs Artificial Intelligence Data Center news New middleware doubles GPU computational efficiency for AI workloads in trials, says Fujitsu The company says the computing broker is aimed at solving the GPU shortage for compute-intensive workloads by improving resource allocation and memory management across AI platforms and applications. By Elizabeth Montalbano Oct 22, 2024 4 mins GPUs Artificial Intelligence news analysis Nvidia: Latest news and insights Here’s what you need to know about the AI and processor giant’s latest product and company news. By Dan Muse Oct 15, 2024 6 mins Artificial Intelligence news US targets advanced AI and cloud firms with new reporting proposal This, along with other AI regulations, sparks worries for enterprises about escalating compliance costs and curbing innovation. By Prasanth Aby Thomas Sep 10, 2024 1 min Regulation Artificial Intelligence Cloud Computing PODCASTS VIDEOS RESOURCES EVENTS NEWSLETTERS Newsletter Promo Module Test Description for newsletter promo module. Please enter a valid email address Subscribe