:Imagine a system that can explore multiple approaches to complex problems, drawing on its understanding of vast amounts of data, from scientific datasets to source code to business documents, and reasoning through the possibilities in real time,” Brown stated in a blog post. “This lightning-fast reasoning isn’t waiting on the horizon. It’s happening today in our customers’ AI production environments. The scale of the AI systems that our customers are building today—across drug discovery, enterprise search, software development, and more—is truly remarkable. And there’s much more ahead.”
Brown said P6e-GB200 UltraServers have been deployed in third-generation EC2 UltraClusters, which creates a single fabric for AWS’s largest data centers. Third-generation UltraClusters cut power consumption by up to 40 percent and reduce cabling requirements by more than 80 percent, improving efficiency and reducing potential points of failure, according to Brown.
AWS uses Elastic Fabric Adapter (EFA) with its Scalable Reliable Datagram protocol, which routes traffic across multiple network paths, built to maintain smooth operation even during congestion or failures. Brown said P6e-GB200 and P6-B200 instances with EFAv4 show up to 18 percent faster collective communications in distributed training compared to P5en instances that use EFAv3.
While P6-B200 instances are air-cooled, P6e-GB200 UltraServers use liquid cooling, enabling higher compute density in large NVLink domain architectures, according to AWS. P6e-GB200 servers provide configurable liquid-to-chip cooling in both new and existing data centers, “so we can support both liquid-cooled accelerators and air-cooled network and storage infrastructure in the same facility.”
P6e-GB200 UltraServers will also be available through NVIDIA DGX Cloud. DGX Cloud is a unified AI platform with NVIDIA’s AI software stack.




