
Meta, on Wednesday, unveiled its next-generation Meta Coaching and Inference Accelerator (MTIA), its household of custom-made chipsets for artificial intelligence (AI) workloads. The improve to its AI chipset comes nearly a yr after the corporate launched the primary AI chips. These Inference Accelerators will energy the tech large’s current and future merchandise, companies, and the AI that lies inside its social media platforms. Specifically, Meta highlighted that the capabilities of the chipset shall be used to serve its rating and suggestion fashions.
Making the announcement by way of its blog post, Meta mentioned, “The subsequent technology of Meta’s large-scale infrastructure is being constructed with AI in thoughts, together with supporting new generative AI (GenAI) services and products, suggestion methods, and superior AI analysis. It is an funding we count on will develop within the years forward because the compute necessities to assist AI fashions enhance alongside the fashions’ sophistication.”
The brand new AI chip provides important enhancements in each energy technology and effectivity as a result of enhancements in its structure, as per Meta. The subsequent technology of MTIA doubles the compute and reminiscence bandwidth in comparison with its predecessor. It may additionally serve Meta’s suggestion fashions that it makes use of to personalise content material for its customers on its social media platforms.
On the {hardware} of the chipset, Meta mentioned that the system has a rack-based design that holds as much as 72 accelerators the place three chassis comprise 12 boards and every of them homes two accelerators. The processor clocks at 1.35GHz which is far sooner than its predecessor at 800MHz. It may additionally run at the next output of 90W. The material between the accelerators and the host has additionally been upgraded to PCIe Gen5.
The software program stack is the place the corporate has made main enhancements. The chipset is designed to be totally built-in with PyTorch 2.0 and associated options. “The decrease degree compiler for MTIA takes the outputs from the frontend and produces extremely environment friendly and device-specific code,” the corporate defined.
The outcomes to date present that this MTIA chip can deal with each the low complexity (LC) and excessive complexity (HC) rating and suggestion fashions which can be parts of Meta’s merchandise. Throughout these fashions, there is usually a ~10x-100x distinction in mannequin measurement and the quantity of compute per enter pattern. As a result of we management the entire stack, we will obtain better effectivity in comparison with commercially obtainable GPUs. Realizing these good points is an ongoing effort and we proceed to enhance efficiency per watt as we construct up and deploy MTIA chips in our methods.
With the rise of AI, many tech corporations at the moment are specializing in manufacturing customised AI chipsets that may cater to their specific wants. These processors supply large compute energy over servers which permits them to carry merchandise resembling generalist AI chatbots and AI instruments for particular duties.