AWS Announces Three New Amazon EC2 Instances Powered by AWS-Designed Chips

December 1, 2021

LAS VEGAS, Dec. 1, 2021 — Tuesday, at AWS re:Invent, Amazon Web Services, Inc. (AWS), an Amazon.com, Inc. company, announced three new Amazon Elastic Compute Cloud (Amazon EC2) instances powered by AWS-designed chips that help customers significantly improve the performance, cost, and energy efficiency of their workloads running on Amazon EC2. New C7g instances powered by next-generation AWS Graviton3 processors deliver up to up to 25% better performance than current generation C6g instances powered by AWS Graviton2 processors. New Trn1 instances powered by AWS Trainium chips provide the best price performance and the fastest time to train most machine learning models in Amazon EC2. New storage-optimized Im4gn/Is4gen/I4i instances based on AWS-designed AWS Nitro SSDs (solid-state drives) offer the best storage performance for I/O-intensive workloads running on Amazon EC2. Together, these instances herald the arrival of new Amazon EC2 instances based on AWS-designed chips that help customers power their most business-critical applications.

“With our investments in AWS-designed chips, customers have realized huge price performance benefits for some of today’s most business-critical workloads. These customers have asked us to continue pushing the envelope with each new EC2 instance generation,” said David Brown, Vice President, Amazon EC2 at AWS. “AWS’s continued innovation means customers are now getting brand new, game changing instances to run their most important workloads with significantly better price performance than anywhere else.”

C7g instances powered by new AWS Graviton3 processors deliver up to 25% better performance compared to current generation C6g instances powered by AWS Graviton2 processors

Customers like DirecTV, Discovery, Epic Games, Formula 1, Honeycomb.io, Intuit, Lyft, MercardoLibre, NextRoll, Nielsen, SmugMug, Snap, Splunk, and Sprinklr have seen significant performance gains and reduced costs from running AWS Graviton2-based instances in production since they launched in 2020. The Graviton2 instance portfolio offers 12 different instances that include general purpose, compute optimized, memory optimized, storage optimized, burstable, and accelerated computing instances, so customers have the deepest and broadest choice of cost-effective and power-efficient compute in the cloud. As customers bring more compute intensive workloads like high performance computing (HPC), gaming, and machine learning inference to the cloud, and as their compute, storage, memory, and networking demands grow, they are looking for even better price performance and energy efficiency to run these demanding workloads.

C7g instances, powered by next generation AWS Graviton3 processors, provide up to 25% better performance for compute-intensive workloads compared to current generation C6g instances powered by Graviton2 processors. AWS Graviton3 processors also deliver up to 2x higher floating point performance for scientific, machine learning, and media encoding workloads, up to 2x faster performance for cryptographic workloads, and up to 3x better performance for machine learning workloads compared to previous generation AWS Graviton2 processors. AWS Graviton3 processors are also more energy efficient, using up to 60% less energy for same performance than comparable EC2 instances. C7g instances are the first in the cloud to feature the latest DDR5 memory, which provides 50% higher memory bandwidth versus AWS Graviton2-based instances to improve the performance of memory-intensive applications like scientific computing. C7g instances also deliver 20% higher networking bandwidth compared to AWS Graviton2-based instances. C7g instances support Elastic Fabric Adapter (EFA), which allows applications to communicate directly with network interface cards, providing lower and more consistent latency, to enhance the performance of applications that require parallel processing at scale like HPC and video encoding. C7g instances are available today in preview. To learn more about C7g instances, visit aws.amazon.com/ec2/instance-types/c7g.

Trn1 instances powered by AWS Trainium chips provide the best price performance and the fastest time to train most machine learning models in Amazon EC2

More and more customers are building, training, and deploying machine learning models to power applications that have the potential to reinvent their businesses and customer experiences. However, to ensure improved accuracy, these machine learning models must consume ever-growing amounts of training data, which causes them to become increasingly expensive to train. This dilemma can have the effect of limiting the number of machine learning models that customers are able to deploy. AWS provides the broadest and deepest choice of compute offerings for machine learning, including the EC2 P4d instances featuring NVIDIA A100 Tensor Core GPUs and EC2 DL1 instances featuring Gaudi accelerators from Habana Labs. But even with the fastest accelerated instances available today, it can still be prohibitively expensive and time consuming to train ever-larger machine learning models.

Trn1 instances powered by AWS Trainium chips offer the best price performance and the fastest machine learning model training in Amazon EC2, providing up to 40% lower cost to train deep learning models compared to the latest P4d instances. Trn1 instances offer 800 Gbps EFA networking bandwidth (2x higher than the latest EC2 GPU-based instances) and integrate with Amazon FSx for Lustre high performance storage—enabling customers to launch Trn1 instances with EC2 UltraClusters capability. With EC2 UltraClusters, developers can scale machine learning training to 10,000+ Trainium accelerators interconnected with petabit-scale networking, giving customers on-demand access to supercomputing-class performance to cut training time from months to days for even the largest and most complex models. Trn1 instances are available today in preview. To learn more about Trn1 instances, visit aws.amazon.com/ec2/instance-types/trn1.

Im4gn/Is4gen/I4i instances featuring new AWS Nitro SSDs deliver the best storage performance for I/O intensive-workloads

Today, customers use I3/I3en storage-optimized instances for applications that require direct access to data sets on local storage like scale-out transactional and relational databases (e.g. MySQL and PostgreSQL), NoSQL databases (e.g. Cassandra, MongoDB, Redis, etc.), big data (e.g. Hadoop), and data analytics workloads (e.g. Spark, Hive, Presto, etc.). I3/I3en instances offer Non-Volatile Memory Express (NVMe) SSD-backed instance storage optimized for low latency, high I/O performance, and throughput at a low cost. Customers appreciate the fast transaction times I3/I3en instances provide, but as they evolve their workloads to process even more complex transactions on larger data sets, they need even higher compute performance and faster access to data, without higher costs.

Im4gn/Is4gen/I4i instances are architected to maximize the storage performance of I/O-intensive workloads. Im4gn/Is4gen/I4i instances offer up to 30 TB of NVMe storage from AWS-designed AWS Nitro SSDs, delivering up to 60% lower I/O latency and 75% lower latency variability compared to previous generation I3 instances to maximize application performance. AWS Nitro SSDs are tightly integrated with the AWS Nitro System via optimizations in the storage stack, hypervisor, and hardware. Because AWS is managing both the hardware and firmware of the AWS Nitro SSDs, customers benefit from improved functionality because SSD updates are delivered more quickly compared to using commercial SSDs. Im4gn instances (available today) feature AWS Graviton2 processors and provide up to 40% better price performance and up to 44% lower cost per TB of storage compared to I3 instances. Is4gen instances (available today) also use AWS Graviton2 processors and provide up to 15% lower cost per TB of storage and up to 48% better compute performance compared to I3en instances. To get started with Im4gn/Is4gen instances, visit aws.amazon.com/ec2/instance-types/i4g. I4i instances (available soon) feature 3rd generation Intel Scalable processors (Ice Lake), delivering up to 55% better compute performance than current generation I3 instances. To learn more about I4i instances, visit aws.amazon.com/ec2/instance-types/i4i.

SAP HANA is a world’s leading in-memory database that serves as the foundation of the SAP Business Technology Platform. “Over the past decade, SAP HANA has helped customers manage their most mission critical transactional and analytics workloads,” said Irfan Khan, President of HANA Database & Analytics at SAP. “AWS investments and innovations on ARM-based AWS Graviton processors and SAP HANA Cloud are a great match with potential to deliver step-wise operation and performance improvement benefits to our enterprise customers, and to SAP’s cloud analytics and data management solutions powered by SAP HANA Cloud.”

Twitter is what’s happening and what people are talking about right now. “Twitter is working on a multi-year project to leverage the AWS Graviton-based EC2 instances to deliver Twitter timelines. As part of our ongoing engineering to drive further efficiencies, we tested the new Graviton3-based C7g instances,” said Nick Tornow, Head of Platform at Twitter. “Across a number of benchmarks that we’ve found to be representative of the performance of Twitter workloads, we found Graviton3-based C7g instances deliver 20%-80% higher performance versus Graviton2-based C6g instances, while also reducing tail latencies by as much as 35%. We are excited to utilize Graviton3-based instances in the future to realize significant price performance benefits.”

Formula 1 (F1) racing began in 1950 and is the world’s most prestigious motor racing competition, as well as the world’s most popular annual sporting series. “We had already seen that Graviton2-based C6gn instances provided us the best price performance for some of our CFD workloads. We have now found Graviton3 C7g instances to be 40% faster than the Graviton2 C6gn instances for those same simulations,” said Pat Symonds, CTO at Formula 1 Management. “We’re excited that EFA will be standard on this instance type, and given this much improved price performance, we expect Graviton3-based instances to become the optimal choice to run all of our CFD workloads.”

Founded in 1991, Epic Games is the creator of Fortnite, Unreal, Gears of War, Shadow Complex, and the Infinity Blade series of games. Epic’s Unreal Engine technology brings high-fidelity, interactive experiences to PC, console, mobile, AR, VR, and the Web. “As we look to the future and building increasingly immersive and compelling experiences for players, we are excited to use AWS Graviton3-based EC2 instances,” said Mark Imbriaco, Senior Director of Engineering at Epic Games. “Our testing has shown they are suitable for even the most demanding, latency-sensitive workloads while providing significant price performance benefits and expanding what is possible within Fortnite and any Unreal Engine created experience.”

Honeycomb develops an observability platform that enables engineering teams to visualize, analyze, and improve cloud application quality and performance. “We’re excited to have tested our high-throughput telemetry ingestion workload against early preview instances of AWS Graviton3 and have seen a 35% performance increase for our workload over Graviton2,” said Liz Fong-Jones, Principal Developer Advocate at honeycomb.io. “We were able to run 30% fewer instances of C7g than C6g serving the same workload, and with 30% reduced latency. We are looking forward to adopting AWS Graviton3-powered C7g instances in production once they are generally available.”

Anthropic builds reliable, interpretable, and steerable AI systems that will have many opportunities to create value commercially and for public benefit. “Our research interests span multiple areas including natural language, human feedback, scaling laws, reinforcement learning, code generation, and interpretability. A major key to our success is access to modern infrastructure that allows us to spin up very large fleets of high-performance deep learning accelerators,” said Tom Brown, Co-founder at Anthropic. “We are looking forward to using Trn1 instances powered by AWS Trainium, as their unprecedented ability to scale to tens of thousands of nodes and higher network bandwidth will enable us to iterate faster while keeping our costs under control.”

Splunk is a leading data platform provider and is designed to investigate, monitor, analyze, and act on data at any scale. “We run C/C++ based workloads for indexing and searching event data. Our workload is CPU bound and benefits from high capacity and low latency SSD storage,” said Brad Murphy, Vice President, Cloud Platform & Infrastructure at Splunk. “When evaluating the new Im4gn/Is4gen instances powered by AWS Graviton2, we observed an up to 50% decrease in search runtime compared to I3/I3en instances, which we currently use. This makes Im4gn and Is4gen instances a great choice for running our storage-intensive workloads with significant price performance improvement and lower TCO.”

Sprinklr helps the world’s biggest companies make their customers happier across 30+ digital channels—using the most advanced, sophisticated AI engine built for the enterprise to create insight-driven strategies and better customer experiences. “We benchmarked our Java-based search workloads on Amazon EC2 Im4gn/Is4gen instances powered by AWS Graviton2 processors. Smaller Is4gen instances offer similar performance compared to larger I3en instances, presenting an opportunity to meaningfully reduce the TCO,” said Abhay Bansal, Vice President of Engineering at Sprinklr. “We also saw a significant 50% reduction in latency for queries when moving our workloads from I3 to Im4gn instances, indicating a significant 40% price performance benefit. Moving to AWS Graviton2-based instances was easy, taking two weeks to complete benchmarking. We are very happy with our experience and look forward to running these workloads in production on Im4gn and Is4gen instances.”

Redis Enterprise powers mission critical apps and services for over 8,000 organizations globally by enabling software teams to create a high-performance data layer for the real-time world. “We’re thrilled to see the Amazon EC2 I4i instances using the new low latency AWS Nitro SSDs that deliver better transaction speed than the previous generation instances,” said Yiftach Shoolman, Co-Founder and CTO at Redis. “We expect the faster storage performance and higher networking and processor speeds of the I4i instances will deliver significant improvements at an even more attractive total cost of ownership for our customers who use Redis-on-Flash on I4i instances.”

About Amazon Web Services

For over 15 years, Amazon Web Services has been the world’s most comprehensive and broadly adopted cloud offering. AWS has been continually expanding its services to support virtually any cloud workload, and it now has more than 200 fully featured services for compute, storage, databases, networking, analytics, machine learning and artificial intelligence (AI), Internet of Things (IoT), mobile, security, hybrid, virtual and augmented reality (VR and AR), media, and application development, deployment, and management from 81 Availability Zones within 25 geographic regions, with announced plans for 27 more Availability Zones and nine more AWS Regions in Australia, Canada, India, Indonesia, Israel, New Zealand, Spain, Switzerland, and the United Arab Emirates. Millions of customers—including the fastest-growing startups, largest enterprises, and leading government agencies—trust AWS to power their infrastructure, become more agile, and lower costs. To learn more about AWS, visit aws.amazon.com.

About Amazon

Amazon is guided by four principles: customer obsession rather than competitor focus, passion for invention, commitment to operational excellence, and long-term thinking. Amazon strives to be Earth’s Most Customer-Centric Company, Earth’s Best Employer, and Earth’s Safest Place to Work. Customer reviews, 1-Click shopping, personalized recommendations, Prime, Fulfillment by Amazon, AWS, Kindle Direct Publishing, Kindle, Career Choice, Fire tablets, Fire TV, Amazon Echo, Alexa, Just Walk Out technology, Amazon Studios, and The Climate Pledge are some of the things pioneered by Amazon. For more information, visit amazon.com/about.


Source: AWS

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

MLPerf Inference 4.0 Results Showcase GenAI; Nvidia Still Dominates

March 28, 2024

There were no startling surprises in the latest MLPerf Inference benchmark (4.0) results released yesterday. Two new workloads — Llama 2 and Stable Diffusion XL — were added to the benchmark suite as MLPerf continues Read more…

Q&A with Nvidia’s Chief of DGX Systems on the DGX-GB200 Rack-scale System

March 27, 2024

Pictures of Nvidia's new flagship mega-server, the DGX GB200, on the GTC show floor got favorable reactions on social media for the sheer amount of computing power it brings to artificial intelligence.  Nvidia's DGX Read more…

Call for Participation in Workshop on Potential NSF CISE Quantum Initiative

March 26, 2024

Editor’s Note: Next month there will be a workshop to discuss what a quantum initiative led by NSF’s Computer, Information Science and Engineering (CISE) directorate could entail. The details are posted below in a Ca Read more…

Waseda U. Researchers Reports New Quantum Algorithm for Speeding Optimization

March 25, 2024

Optimization problems cover a wide range of applications and are often cited as good candidates for quantum computing. However, the execution time for constrained combinatorial optimization applications on quantum device Read more…

NVLink: Faster Interconnects and Switches to Help Relieve Data Bottlenecks

March 25, 2024

Nvidia’s new Blackwell architecture may have stolen the show this week at the GPU Technology Conference in San Jose, California. But an emerging bottleneck at the network layer threatens to make bigger and brawnier pro Read more…

Who is David Blackwell?

March 22, 2024

During GTC24, co-founder and president of NVIDIA Jensen Huang unveiled the Blackwell GPU. This GPU itself is heavily optimized for AI work, boasting 192GB of HBM3E memory as well as the the ability to train 1 trillion pa Read more…

MLPerf Inference 4.0 Results Showcase GenAI; Nvidia Still Dominates

March 28, 2024

There were no startling surprises in the latest MLPerf Inference benchmark (4.0) results released yesterday. Two new workloads — Llama 2 and Stable Diffusion Read more…

Q&A with Nvidia’s Chief of DGX Systems on the DGX-GB200 Rack-scale System

March 27, 2024

Pictures of Nvidia's new flagship mega-server, the DGX GB200, on the GTC show floor got favorable reactions on social media for the sheer amount of computing po Read more…

NVLink: Faster Interconnects and Switches to Help Relieve Data Bottlenecks

March 25, 2024

Nvidia’s new Blackwell architecture may have stolen the show this week at the GPU Technology Conference in San Jose, California. But an emerging bottleneck at Read more…

Who is David Blackwell?

March 22, 2024

During GTC24, co-founder and president of NVIDIA Jensen Huang unveiled the Blackwell GPU. This GPU itself is heavily optimized for AI work, boasting 192GB of HB Read more…

Nvidia Looks to Accelerate GenAI Adoption with NIM

March 19, 2024

Today at the GPU Technology Conference, Nvidia launched a new offering aimed at helping customers quickly deploy their generative AI applications in a secure, s Read more…

The Generative AI Future Is Now, Nvidia’s Huang Says

March 19, 2024

We are in the early days of a transformative shift in how business gets done thanks to the advent of generative AI, according to Nvidia CEO and cofounder Jensen Read more…

Nvidia’s New Blackwell GPU Can Train AI Models with Trillions of Parameters

March 18, 2024

Nvidia's latest and fastest GPU, codenamed Blackwell, is here and will underpin the company's AI plans this year. The chip offers performance improvements from Read more…

Nvidia Showcases Quantum Cloud, Expanding Quantum Portfolio at GTC24

March 18, 2024

Nvidia’s barrage of quantum news at GTC24 this week includes new products, signature collaborations, and a new Nvidia Quantum Cloud for quantum developers. Wh Read more…

Alibaba Shuts Down its Quantum Computing Effort

November 30, 2023

In case you missed it, China’s e-commerce giant Alibaba has shut down its quantum computing research effort. It’s not entirely clear what drove the change. Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Shutterstock 1285747942

AMD’s Horsepower-packed MI300X GPU Beats Nvidia’s Upcoming H200

December 7, 2023

AMD and Nvidia are locked in an AI performance battle – much like the gaming GPU performance clash the companies have waged for decades. AMD has claimed it Read more…

DoD Takes a Long View of Quantum Computing

December 19, 2023

Given the large sums tied to expensive weapon systems – think $100-million-plus per F-35 fighter – it’s easy to forget the U.S. Department of Defense is a Read more…

Synopsys Eats Ansys: Does HPC Get Indigestion?

February 8, 2024

Recently, it was announced that Synopsys is buying HPC tool developer Ansys. Started in Pittsburgh, Pa., in 1970 as Swanson Analysis Systems, Inc. (SASI) by John Swanson (and eventually renamed), Ansys serves the CAE (Computer Aided Engineering)/multiphysics engineering simulation market. Read more…

Choosing the Right GPU for LLM Inference and Training

December 11, 2023

Accelerating the training and inference processes of deep learning models is crucial for unleashing their true potential and NVIDIA GPUs have emerged as a game- Read more…

Intel’s Server and PC Chip Development Will Blur After 2025

January 15, 2024

Intel's dealing with much more than chip rivals breathing down its neck; it is simultaneously integrating a bevy of new technologies such as chiplets, artificia Read more…

Baidu Exits Quantum, Closely Following Alibaba’s Earlier Move

January 5, 2024

Reuters reported this week that Baidu, China’s giant e-commerce and services provider, is exiting the quantum computing development arena. Reuters reported � Read more…

Leading Solution Providers

Contributors

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Shutterstock 1179408610

Google Addresses the Mysteries of Its Hypercomputer 

December 28, 2023

When Google launched its Hypercomputer earlier this month (December 2023), the first reaction was, "Say what?" It turns out that the Hypercomputer is Google's t Read more…

AMD MI3000A

How AMD May Get Across the CUDA Moat

October 5, 2023

When discussing GenAI, the term "GPU" almost always enters the conversation and the topic often moves toward performance and access. Interestingly, the word "GPU" is assumed to mean "Nvidia" products. (As an aside, the popular Nvidia hardware used in GenAI are not technically... Read more…

Shutterstock 1606064203

Meta’s Zuckerberg Puts Its AI Future in the Hands of 600,000 GPUs

January 25, 2024

In under two minutes, Meta's CEO, Mark Zuckerberg, laid out the company's AI plans, which included a plan to build an artificial intelligence system with the eq Read more…

Google Introduces ‘Hypercomputer’ to Its AI Infrastructure

December 11, 2023

Google ran out of monikers to describe its new AI system released on December 7. Supercomputer perhaps wasn't an apt description, so it settled on Hypercomputer Read more…

China Is All In on a RISC-V Future

January 8, 2024

The state of RISC-V in China was discussed in a recent report released by the Jamestown Foundation, a Washington, D.C.-based think tank. The report, entitled "E Read more…

Intel Won’t Have a Xeon Max Chip with New Emerald Rapids CPU

December 14, 2023

As expected, Intel officially announced its 5th generation Xeon server chips codenamed Emerald Rapids at an event in New York City, where the focus was really o Read more…

IBM Quantum Summit: Two New QPUs, Upgraded Qiskit, 10-year Roadmap and More

December 4, 2023

IBM kicks off its annual Quantum Summit today and will announce a broad range of advances including its much-anticipated 1121-qubit Condor QPU, a smaller 133-qu Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire