The companies are extending their AI partnership, and one key initiative is a supercomputer that will be integrated with AWS services and used by Nvidia’s own R&D teams. Credit: Shutterstock Amazon Web Services and Nvidia have announced an expansion of their alliance that includes plans to add supercomputing capabilities to AWS’s artificial intelligence (AI) infrastructure. The companies announced the news at the AWS re:Invent conference in Las Vegas. The biggest of the initiatives is Project Ceiba, a supercomputer that will be hosted by AWS for Nvidia’s own research and development teams. It will feature 16,384 Nvidia GH200 Superchips and be capable of processing 65 exaflops of AI, the companies said. The Project Ceiba supercomputer will be integrated with a number of AWS services, including Amazon Virtual Private Cloud (VPC) encrypted networking and Amazon Elastic Block Store high-performance block storage. Nvidia plans to use the supercomputer for research and development to advance AI for LLMs, graphics and simulation, digital biology, robotics, self-driving cars, Earth-2 climate prediction and more. New Amazon EC2 G6e instances featuring Nvidia L40S GPUs and G6 instances powered by L4 GPUs are also in the works, AWS announced. L4 GPUs are scaled back from the Hopper H100 but offer much more power efficiency. These new instances are aimed at startups, enterprises, and researchers looking to experiment with AI. Nvidia also shared plans to integrate its NeMo Retriever microservice into AWS to help users with their development of generative AI tools like chatbots. NeMo Retriever is a generative AI microservice that enables enterprises to connect custom LLMs to enterprise data, so the company can generate proper AI responses based on their own data. “Generative AI is transforming cloud workloads and putting accelerated computing at the foundation of diverse content generation,” said Jensen Huang, founder and CEO of Nvidia, in a statement. “Driven by a common mission to deliver cost-effective, state-of-the-art generative AI to every customer, Nvidia and AWS are collaborating across the entire computing stack, spanning AI infrastructure, acceleration libraries, foundation models, and generative AI services.” In other news, AWS will be the first cloud provider to bring Nvidia’s GH200 Grace Hopper Superchips to the cloud. The Nvidia GH200 NVL32 multi-node platform connects 32 Grace Hopper Superchips by Nvidia’s NVLink and NVSwitch interconnects. The platform will be available on Amazon Elastic Compute Cloud (EC2) instances connected with Amazon’s networking, virtualization (AWS Nitro System), and hyper-scale clustering (Amazon EC2 UltraClusters). AWS will host Nvidia’s DGX Cloud cluster of GPUs for AI. DGX Cloud on AWS will accelerate training of generative AI and LLMs that can reach beyond 1 trillion parameters. Related content news High-bandwidth memory nearly sold out until 2026 While it might be tempting to blame Nvidia for the shortage of HBM, it’s not alone in driving high-performance computing and demand for the memory HPC requires. By Andy Patrizio May 13, 2024 3 mins CPUs and Processors High-Performance Computing Data Center news CHIPS Act to fund $285 million for semiconductor digital twins Plans call for building an institute to develop digital twins for semiconductor manufacturing and share resources among chip developers. By Andy Patrizio May 10, 2024 3 mins CPUs and Processors Data Center news HPE launches storage system for HPC and AI clusters The HPE Cray Storage Systems C500 is tuned to avoid I/O bottlenecks and offers a lower entry price than Cray systems designed for top supercomputers. By Andy Patrizio May 07, 2024 3 mins Supercomputers Enterprise Storage Data Center news Lenovo ships all-AMD AI systems New systems are designed to support generative AI and on-prem Azure. By Andy Patrizio Apr 30, 2024 3 mins CPUs and Processors Data Center PODCASTS VIDEOS RESOURCES EVENTS NEWSLETTERS Newsletter Promo Module Test Description for newsletter promo module. Please enter a valid email address Subscribe