Intel launches ‘Gaudi 3 accelerator’; A GenAI strategy for enterprise

Intel Gaudi 3 accelerator
Rep.Image | Courtesy: Intel
By Arya M Nair, Official Reporter
  • Follow author on

Intel, an American multinational corporation and technology company, has introduced the Intel Gaudi 3 accelerator to bring performance, openness and choice to enterprise generative AI (GenAI), at the Intel Vision 2024 customer and partner conference.

The company also unveiled a suite of new open scalable systems, next-gen products and strategic collaborations to accelerate GenAI adoption. With only 10 percent of enterprises successfully moving GenAI projects into production last year, Intel’s latest offerings address the challenges businesses face in scaling AI initiatives.

Enterprises are looking to scale GenAI from pilot to production. To do so, they need readily available solutions, built on performant and cost- and energy-efficient processors like the Intel Gaudi 3 AI accelerator, that also address complexity, fragmentation, data security and compliance requirements.

Pat Gelsinger_Intel launches Gaudi 3 accelerator
Pat Gelsinger
CEO – Intel

“Innovation is advancing at an unprecedented pace, all enabled by silicon, and every company is quickly becoming an AI company. Intel is bringing AI everywhere across the enterprise, from the PC to the data center to the edge. Our latest Gaudi, Xeon and Core Ultra platforms are delivering a cohesive set of flexible solutions tailored to meet the changing needs of our customers and partners and capitalize on the immense opportunities ahead.”

The Intel Gaudi 3 AI accelerator will power AI systems with up to tens of thousands of accelerators connected through the common standard of Ethernet. Intel Gaudi 3 promises 4x more AI computing for BF16 and a 1.5x increase in memory bandwidth over its predecessor. The accelerator will deliver a significant leap in AI training and inference for global enterprises looking to deploy GenAI at scale.

In comparison to Nvidia H100, Intel Gaudi 3 is projected to deliver 50 percent faster time-to-train on average3 across Llama2 models with 7B and 13B parameters, and GPT-3 175B parameter model. Additionally, Intel Gaudi 3 accelerator inference throughput is projected to outperform the H100 by 50 percent on average1 and 40 percent for inference power-efficiency averaged2 across Llama 7B and 70B parameters, and Falcon 180B parameter models.

Intel Gaudi 3 provides open, community-based software and industry-standard Ethernet networking. And it allows enterprises to scale flexibly from a single node to clusters, super-clusters and mega-clusters with thousands of nodes, supporting inference, fine-tuning and training at the largest scale.

Intel Gaudi 3 will be available to OEMs, including Dell Technologies, Hewlett Packard Enterprise, Lenovo and Supermicro, in the second quarter of 2024.

Don’t Miss | OpenAI rolls out ‘Voice Engine’ to recreate human voices

YOU MAY LIKE