Huang Renxun brings re:Invent to life on paper

2023-11-30

The influence of Jensen Huang is everywhere, and so is NVIDIA. The company seems to dominate the spotlight in all major events, whether it's Google Cloud Next, AWS re:Invent, or Microsoft Ignite, it always brings the party atmosphere to a climax.

In a recent re:Invent keynote, AWS and NVIDIA announced a strategic initiative to provide a new category of supercomputing infrastructure, software, and services for generative AI.

The two companies also decided to deploy NVIDIA's long-awaited GH200 chip, originally planned for release in 2024. The installation of the NVIDIA GH200 chip will take place in AWS's cloud infrastructure, emphasizing the global availability of this advanced hardware for AWS customers.

NVIDIA x AWS

The event also showcased its efforts to build the world's fastest GPU-driven AI supercomputer, which will be a huge leap forward in reshaping the industry and driving technological advancements at unprecedented speeds. This innovation is named "Project Cieba" and aims to house 16,384 NVIDIA GH200 superchips, delivering an astonishing 65 ExaFLOPS of AI computing power to drive NVIDIA's next wave of generative AI innovation.

In addition, AWS is also collaborating with NVIDIA to introduce three new Amazon EC2 instances, including the P5e instance for large-scale generative AI and high-performance computing (HPC) workloads, as well as the G6 and G6e instances for a wide range of applications such as AI refinement, inference, and graphics processing.

"NVIDIA and AWS are collaborating across the entire compute stack, from AI infrastructure, acceleration libraries, and foundational models to generative AI services," said Jensen Huang, CEO of NVIDIA.

NVIDIA x Hyperscale Cloud Providers

Clearly, NVIDIA is thriving in collaborative environments, and this became even more evident at AWS re:Invent as it also shared partnerships with competitors Google Cloud, Microsoft Azure, and Oracle.

At this event, NVIDIA and Microsoft announced their collaboration to launch the AI Foundry service on Microsoft Azure, aimed at facilitating the development of custom generative AI applications for enterprises and startups. This service integrates NVIDIA's AI technology and DGX Cloud AI supercomputing with Azure's infrastructure, providing a comprehensive solution for creating and deploying custom AI models.

The partnership also emphasizes the use of NVIDIA's AI foundational models and tools for custom model development, making these advancements easily accessible through Azure's cloud platform and marketplace. This collaboration marks an important step in promoting the development and deployment of advanced AI applications across various industries.

"Many of Google's products are built and served on NVIDIA GPUs, and many of our customers are seeking NVIDIA-accelerated computing to efficiently develop LLMs, thus advancing generative AI," shared Thomas Kurian, CEO of Google Cloud, at this year's Next event.

At Google Cloud Next, NVIDIA collaborates with Google to advance AI computing, software, and services while enhancing AI supercomputing capabilities.

The two companies are working together to optimize Google's PaxML for NVIDIA GPUs, facilitating the development of large-scale language models, and have integrated serverless Spark with NVIDIA GPUs to accelerate data processing. Additionally, Google Cloud announced the adoption of NVIDIA H100 GPUs in its A3 VMs and Vertex AI platform, gaining access to resources such as the NVIDIA DGX GH200 AI supercomputer.

Recently, Oracle also announced a longstanding partnership with NVIDIA to accelerate enterprise AI adoption, helping customers address business challenges.

In a recent interview with AIM, Oracle stated that it is fully equipped in terms of infrastructure, with NVIDIA choosing OCI as the first hyperscale cloud provider to offer NVIDIA DGX Cloud. Chris Chelliah of Oracle said, "When NVIDIA thinks about cloud computing and data, they think of Oracle." He mentioned that the company utilizes MySQL HeatWave data for real-time anomaly detection on NVIDIA clusters to serve its customers.

NVIDIA is Everywhere

NVIDIA's diverse partnerships with leading cloud providers and hyperscale cloud service providers provide it with a unique positioning in various aspects of the AI field. While everyone is busy building their internal silicon capabilities to handle AI workloads, the nature of these partnerships seems to be rapidly evolving.

From an innovation and AI advancement perspective, Google Cloud appears to be NVIDIA's favorite, while Microsoft Azure is a key partner in enterprise coverage and application development due to its strong enterprise focus and extensive customer base.

Oracle stands out from others in data management and AI-driven solutions, particularly with its emphasis on real-time data processing capabilities. On the other hand, AWS plays a crucial role in AI solutions centered around security, addressing the growing concerns about AI security and reliability.

Overall, these partnerships provide NVIDIA with a multifaceted platform to expand its AI capabilities and market coverage. Each partnership aligns with NVIDIA's strategic focus areas, whether it's AI innovation, enterprise applications, data management, or ensuring the security of AI solutions. In short, everyone loves NVIDIA.