AWS claims its new Amazon EC2 M7g and R7g instances provide 25% better performance vs. the past generation of instances. Credit: CIS Amazon Web Services (AWS) has announced availability of its new Amazon EC2 M7g and R7g instances, the latest generation of instances for memory-intensive applications and running Amazons custom Arm processor, known as Graviton3. This is the second offering of Graviton3-based instances from AWS. It previously announced specific instances for compute-intensive workloads last May. Both the M7g and the R7g instances deliver up to 25% higher performance than equivalent sixth-generation instances. Part of the performance bump comes from the adoption of DDR5 memory, which offers up to 50% higher memory bandwidth than DDR4. But there’s also considerable performance gain from the new Graviton3 chip. Amazon claims that compared to instances run on Graviton2, the new M7g and R7g instances offer up to 25% higher compute performance, nearly twice the floating point performance, twice the cryptographic performance, and up to three times faster machine-learning inference. The M7g instances are for general purpose workloads such as application servers, microservices, and mid-sized data stores. M7g instances scale from one virtual CPU with 4GiB of memory and 12.5Gbps of network bandwidth to 64 vCPUs with 256GiB of memory and 30Gbps of network bandwidth. (A GiB is a gibibyte, a different method of measuring storage. The term 1GB implies 1GB of storage, but it actually represents 0.93GB. To avoid confusion and promote accuracy, 1GiB represents 0.93GB, but the term gibibyte hasn’t caught on.) The R7g instances are tuned for memory-intensive workloads such as in-memory databases and caches, and real-time big-data analytics. R7g instances scale from 1 vCPU and 8GB of memory with 12.5Gbps of network bandwidth to 64 vCPUs with 512GB of memory and 30 Gbps of network bandwidth. New AWS AI partnership AWS has also announced an expanded partnership with startup Hugging Face to make more of its AI tools available to AWS customers. These include Hugging Face’s language-generation tool for building generative AI applications to perform tasks like text summarization, answering questions, code generation, image creation, and writing essays and articles. The models will run on AWS’s purpose-built ML accelerators for the training (AWS Trainium) and inference (AWS Inferentia) of large language and vision models.The benefits of the models include faster training and scaling low-latency, high-throughput inference. Amazon claims Trainium instances offer 50% lower cost-to-train vs. comparable GPU-based instances. Hugging Face models on AWS can be used three ways: through SageMaker JumpStart, AWS’s tool for building and deploying machine-language models; the Hugging Face AWS Deep Learning Containers (DLCs); or tutorials to deploy customer models to AWS Trainium or AWS Inferentia. Related content news High-bandwidth memory nearly sold out until 2026 While it might be tempting to blame Nvidia for the shortage of HBM, it’s not alone in driving high-performance computing and demand for the memory HPC requires. By Andy Patrizio May 13, 2024 3 mins CPUs and Processors High-Performance Computing Data Center news CHIPS Act to fund $285 million for semiconductor digital twins Plans call for building an institute to develop digital twins for semiconductor manufacturing and share resources among chip developers. By Andy Patrizio May 10, 2024 3 mins CPUs and Processors Data Center news HPE launches storage system for HPC and AI clusters The HPE Cray Storage Systems C500 is tuned to avoid I/O bottlenecks and offers a lower entry price than Cray systems designed for top supercomputers. By Andy Patrizio May 07, 2024 3 mins Supercomputers Enterprise Storage Data Center news Lenovo ships all-AMD AI systems New systems are designed to support generative AI and on-prem Azure. By Andy Patrizio Apr 30, 2024 3 mins CPUs and Processors Data Center PODCASTS VIDEOS RESOURCES EVENTS NEWSLETTERS Newsletter Promo Module Test Description for newsletter promo module. Please enter a valid email address Subscribe