Intel Unveils Gaudi 3 AI Accelerator, Claims Superior Performance Over NVIDIA H100
At the 2024 Intel Vision Conference held in Phoenix, Arizona, Intel unveiled its latest artificial intelligence chip, the Gaudi 3 AI accelerator, aiming to challenge NVIDIA's dominant position in the AI chip market. Gaudi 3 is designed to provide exceptional performance, openness, and flexibility for enterprise-grade generative AI, helping businesses overcome challenges in expanding AI applications.
According to official data from Intel, Gaudi 3 offers an average 50% higher inference performance compared to NVIDIA's H100 GPU, while improving power efficiency by 40%, all at a fraction of the cost. In terms of training performance, Intel predicts that Gaudi 3 will be 1.7 times faster than H100.
The Gaudi 3 chip adopts advanced 5nm manufacturing process, a significant improvement over the previous generation Gaudi 2's 7nm process. This advancement endows it with advanced features and transistors, greatly enhancing AI computing power, network bandwidth, and memory bandwidth.
Intel CEO Pat Gelsinger stated, "Innovation is evolving at an unprecedented pace, and it is all supported by silicon technology. Today, every company is rapidly transforming into an AI company. Intel is bringing the power of AI to every corner of the enterprise, from personal computers to data centers, and to edge computing."
Gaudi 3 will be supplied to original equipment manufacturers (OEMs) including Dell Technologies, Hewlett Packard Enterprise, Lenovo, and Supermicro, expanding the product choices for enterprises in the AI data center market. Additionally, Intel announced collaborations with numerous new customers and partners such as Bharti Airtel, Bosch, CtrlS, IBM, IFF, Landing AI, Ola, NAVER, NielsenIQ, Roboflow, and Seekr to promote the application and development of the Gaudi accelerator.
In addition to the Gaudi 3 accelerator, Intel has also updated its next-generation products and services, covering various areas of enterprise AI. The company has introduced a new data center, cloud, and edge processor brand - Intel Xeon 6, which is expected to deliver exceptional efficiency and enhanced AI performance.
Although NVIDIA still holds approximately 80% of the market share in the AI chip market, Intel provides a strong competitive choice for enterprises seeking alternatives to NVIDIA's proprietary solutions with its open approach and Ethernet network that complies with industry standards. Intel also emphasizes the importance of its software stack, promoting its end-to-end support and open-source nature, and collaborating with partners such as Hugging Face, PyTorch, DeepSpeed, and Mosaic to simplify the software porting process and accelerate the deployment of Gaudi 3 systems.
Sachin Katti, Senior Vice President of Intel's Network Group, said, "We are closely collaborating with the software ecosystem, building an open reference software, and allowing you to freely assemble solutions instead of being forced to purchase complete solution building blocks."
As cloud service providers and enterprises continue to build infrastructure for deploying AI software, the data center AI market is expected to continue growing. This indicates that although NVIDIA remains the major manufacturer of most AI chips, there is still enough room for competitors to develop. The high cost of running generative AI and purchasing Nvidia GPUs has led many enterprises to seek additional suppliers to reduce costs.
Intel's move into the AI chip market comes at a critical time for the company's transformation. It is striving to reshape itself as a leader in advanced chip technology and actively building chip manufacturing capabilities in the United States and overseas. Despite the company's stock performance lagging behind competitors NVIDIA and AMD, Intel believes that its open approach and competitive pricing strategy will attract enterprises that are unwilling to rely solely on a single supplier to meet high-value demands such as AI hardware.