Connect with us


AWS and Nvidia to partner on new AI supercomputing infrastructure



Amazon Web Services (AWS) and Nvidia on Tuesday announced new initiatives in their strategic collaboration that will focus on adding supercomputing capabilities to the companies’ artificial intelligence (AI) infrastructure.

The announcement came at the AWS re:Invent conference and features several notable projects. One major new initiative is known as Project Ceiba, a supercomputer that will be integrated with several AWS services. It will give Nvidia access to a comprehensive set of AWS capabilities, including its Virtual Private Cloud encrypted networking and high-performance block storage. 

Project Ceiba will be used for research and development aimed at advancing AI for large language models (LLMs), graphics – including images, videos and 3D generation – in addition to simulations, digital biology, robotics, self-driving cars, Earth-2 climate prediction and more.


AWS and Nvidia announced several new AI initiatives as part of their strategic collaboration. (AWS: Photo by Jaap Arriens/NurPhoto via Getty Images | Nvidia: Photo by Ahmet Serdar Eser/Anadolu via Getty Images / Getty Images)

AWS and Nvidia will also partner in powering Nvidia DGX Cloud, an AI supercomputing service that gives enterprises access to multi-node supercomputing to train complex LLMs and generative AI models. It will be integrated with Nvidia AI Enterprise software and provide customers with direct access to Nvidia’s AI experts.

Ticker Security Last Change Change %
AMZN AMAZON.COM INC. 147.03 -0.70 -0.47%
NVDA NVIDIA CORP. 478.21 -4.21 -0.87%

Amazon will become the first cloud provider to offer Nvidia’s GH200 Grace Hopper Superchips with multi-node NVLink technology with its Elastic Cloud Compute (EC2) platform. The Nvidia Superchips will allow Amazon EC2 to provide up to 20 terabytes of memory to power terabyte-scale workloads.


Jensen Huang Nvidia Taipei Computex

Jensen Huang, co-founder and chief executive officer of Nvidia Corp., speaks during the Taipei Computex expo in Taipei, Taiwan, on Monday, May 29, 2023. (Photographer: I-Hwa Cheng/Bloomberg via Getty Images / Getty Images)

Nvidia will also integrate its NeMo Retriever microservice into AWS to help users boost their development of generative AI tools like chatbots and summarization tools that leverage accelerated semantic retrieval. 

Nvidia BioNeMo – which is available on Amazon SageMaker and will be incorporated on AWS on Nvidia DGX Cloud – helps pharmaceutical companies speed up the drug discovery process by simplifying and accelerating the training of AI models using their own data.

“Generative AI is transforming cloud workloads and putting accelerated computing at the foundation of diverse content generation,” said Jensen Huang, founder and CEO of Nvidia. “Driven by a common mission to deliver cost-effective, state-of-the-art generative AI to every customer, Nvidia and AWS are collaborating across the entire computing stack, spanning AI infrastructure, acceleration libraries, foundation models, and generative AI services.”


Adam Selipsky AWS

Amazon Web Services (AWS) CEO Adam Selipsky delivers a keynote address during AWS re:Invent 2022, a conference hosted by Amazon Web Services, at The Venetian Las Vegas on November 29, 2022, in Las Vegas, Nevada. (Noah Berger/Getty Images for Amazon Web Services / Getty Images)

“AWS and Nvidia have collaborated for more than 13 years, beginning with the world’s first GPU cloud instance,” said Amazon Web Services CEO Adam Selipsky. “We continue to innovate with Nvidia to make AWS the best place to run GPUs, combining next-gen Nvidia Grace Hopper Superchips with AWS’s EFA powerful networking, EC2 UltraClusters’ hyper-scale clustering, and Nitro’s advanced virtualization capabilities.”

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *