OpenAI has signed a landmark $38 billion deal with Amazon Web Services (AWS) in a seven-year partnership that grants the AI company immediate access to hundreds of thousands of NVIDIA GPUs for training and deploying advanced AI models. This agreement marks OpenAI’s first major computing partnership outside Microsoft Azure and signals a strategic shift toward multi-cloud infrastructure as the company scales ChatGPT and develops next-generation AI capabilities.
Table of Contents
The partnership, announced November 2, 2025, will see OpenAI utilize Amazon EC2 UltraServers equipped with state-of-the-art NVIDIA GB200 and GB300 GPUs, with deployment expected by the end of 2026 and expansion continuing through 2027. This massive infrastructure investment underscores the escalating compute demands of frontier AI development and positions AWS as a critical backbone for OpenAI’s ambitious roadmap.
What the AWS-OpenAI Partnership Includes
The $38 billion agreement provides OpenAI with unprecedented access to AWS’s cloud infrastructure, specifically designed for large-scale AI workloads. The partnership centers on Amazon EC2 UltraServers, a sophisticated architectural platform that clusters hundreds of thousands of NVIDIA GPUs on the same network to enable low-latency performance across interconnected systems.
OpenAI will leverage both NVIDIA GB200 and GB300 GPUs organized through EC2 UltraServers, creating computing clusters optimized for maximum AI processing efficiency. These clusters support diverse workloads ranging from serving inference requests for ChatGPT to training trillion-parameter models, with the flexibility to adapt as OpenAI’s needs evolve.
AWS has committed to providing capacity comprising hundreds of thousands of GPUs alongside the ability to scale to tens of millions of CPUs, particularly for rapidly expanding agentic AI workloads that require distributed computing power. All contracted capacity is targeted for deployment before the end of 2026, with provisions for further expansion into 2027 and beyond.
The AWS-OpenAI partnership is a $38 billion, seven-year agreement announced in November 2025 that gives OpenAI immediate access to AWS cloud infrastructure, including hundreds of thousands of NVIDIA GB200/GB300 GPUs via Amazon EC2 UltraServers, to support ChatGPT inference and next-generation AI model training.
Why OpenAI Chose AWS Infrastructure
OpenAI’s decision to partner with AWS reflects the company’s growing compute requirements and strategic move toward infrastructure diversification. According to OpenAI CEO Sam Altman, “Scaling frontier AI requires massive, reliable compute,” and AWS’s world-class infrastructure provides the scale, security, and performance necessary for this next era of AI development.
AWS brings exceptional experience running large-scale AI infrastructure securely and reliably, with computing clusters exceeding 500,000 chips already in operation. This proven track record at scale was a determining factor for OpenAI as the company seeks to push models to new heights of intelligence.
The partnership also came at an opportune moment following OpenAI’s recent corporate restructuring, which granted the company greater flexibility in cloud provider relationships. Microsoft, OpenAI’s largest investor with a 27% stake in the newly formed OpenAI Group, waived its “first right of refusal” for exclusive compute provision, effectively opening the door for multi-cloud partnerships.
AWS CEO Matt Garman emphasized that “AWS’s best-in-class infrastructure will serve as a backbone for their AI ambitions,” highlighting the breadth and immediate availability of optimized compute that demonstrates AWS’s unique positioning to support vast AI workloads.
Amazon EC2 UltraServers Explained
Amazon EC2 UltraServers represent AWS’s cutting-edge infrastructure designed specifically for demanding AI workloads. These servers connect multiple EC2 instances using dedicated, high-bandwidth, and low-latency accelerator interconnects, creating unified computing clusters that function as single powerful systems.
The P6e-GB200 UltraServers, which OpenAI will utilize, are accelerated by NVIDIA GB200 NVL72 technology and can access up to 72 NVIDIA Blackwell GPUs within one NVLink domain. This configuration delivers 360 petaflops of FP8 compute performance (without sparsity) and 13.4 TB of total high-bandwidth memory (HBM3e).
Each NVIDIA Grace Blackwell Superchip in these systems connects two high-performance NVIDIA Blackwell tensor core GPUs with an NVIDIA Grace CPU using the NVLink-C2C interconnect, providing 10 petaflops of FP8 compute and up to 372 GB HBM3e memory per superchip. The superchip architecture colocates GPU and CPU within one compute module, significantly increasing bandwidth between processors compared to previous generation instances.
These UltraServers are deployed in third-generation EC2 UltraClusters and represent AWS’s first large-scale liquid-cooled systems, offering up to 28.8 Tbps networking via Elastic Fabric Adapter v4 (EFAv4). This infrastructure is integrated with managed services like Amazon SageMaker HyperPod, Amazon EKS, and NVIDIA DGX Cloud for streamlined AI development workflows.
Amazon EC2 UltraServers are high-performance computing platforms that connect multiple EC2 instances via dedicated accelerator interconnects, supporting up to 72 NVIDIA Blackwell GPUs with 360 petaflops of compute and 13.4 TB of memory for training and deploying advanced AI models.
OpenAI’s Multi-Cloud Strategy
This AWS partnership represents a significant evolution in OpenAI’s infrastructure strategy, marking a departure from the company’s historical primary reliance on Microsoft Azure. While Microsoft remains a key investor and technology partner, OpenAI is now pursuing a diversified, multi-cloud approach to meet escalating compute demands.
The multi-cloud strategy provides several advantages for OpenAI, including access to diverse GPU technologies, reduced dependency on any single provider, and the flexibility to optimize workloads across different infrastructure platforms. By some estimates, OpenAI has signed approximately $1 trillion worth of infrastructure deals in 2025, including a $300 billion Oracle agreement and the $500 billion Stargate project with Oracle and SoftBank.
OpenAI’s models already became available on Amazon Bedrock earlier in 2025, making the company one of the most popular publicly available model providers on the platform with thousands of customers using OpenAI models for agentic workflows, coding, scientific analysis, and mathematical problem-solving. This existing integration provided a foundation for the expanded compute partnership announced in November.
The diversification strategy also builds a more resilient compute pipeline for OpenAI, crucial for continuous innovation in a rapidly evolving field where infrastructure availability can constrain development timelines. AWS’s leadership in cloud infrastructure combined with OpenAI’s pioneering advancements in generative AI creates a powerful ecosystem that will help millions of users continue deriving value from ChatGPT and future OpenAI products.
Impact on ChatGPT and AI Development
The AWS infrastructure will directly support ChatGPT’s operational needs and enable OpenAI to scale its most popular AI assistant to meet growing demand. AWS CEO Matt Garman confirmed that the infrastructure will provide “the compute they need, from serving inference for ChatGPT to training next-generation models to scaling agentic AI workloads”.
Inference workloads the process of running trained models to generate responses for user queries require massive computational resources as ChatGPT serves millions of daily users. The low-latency performance enabled by EC2 UltraServers’ clustered architecture ensures ChatGPT can deliver fast, responsive interactions even during peak usage periods.
Beyond serving existing products, the AWS partnership positions OpenAI to train increasingly sophisticated models that push the boundaries of artificial intelligence. Training frontier models with trillions of parameters demands enormous computing clusters operating in coordination, precisely the capability AWS UltraClusters provide.
The infrastructure specifically supports agentic AI workloads autonomous AI systems that can plan, execute tasks, and interact with external tools with minimal human intervention. OpenAI’s ability to scale to tens of millions of CPUs through AWS infrastructure will be critical for deploying these next-generation capabilities broadly.
The AWS partnership provides OpenAI with the computational infrastructure to run ChatGPT inference at scale, train trillion-parameter frontier models, and deploy agentic AI workloads, with hundreds of thousands of NVIDIA GPUs delivering the performance needed for real-time responses and advanced model development.
AWS vs Azure for OpenAI Workloads
| Factor | AWS (New Partnership) | Microsoft Azure (Existing) | 
|---|---|---|
| GPU Access | Hundreds of thousands of NVIDIA GB200/GB300 GPUs via EC2 UltraServers | Existing NVIDIA GPU infrastructure, primary compute provider since 2019 | 
| Compute Scale | Up to tens of millions of CPUs, clusters exceeding 500K chips | Large-scale infrastructure with Azure OpenAI Service integration | 
| Deployment Timeline | Immediate access, full deployment by end of 2026 | Already operational with established workloads | 
| Infrastructure Type | Amazon EC2 UltraServers with liquid cooling, 360 PFLOPS per cluster | Azure ML optimized compute with global data centers | 
| Strategic Relationship | $38B seven-year partnership, first major AWS-OpenAI compute deal | Longtime partner, 27% investor in OpenAI Group, integrated Azure OpenAI Service | 
| Integration | Models available on Amazon Bedrock for enterprise customers | Deep integration with Microsoft 365, Azure AI Foundry | 
The comparison reveals OpenAI’s strategic balancing act between leveraging Microsoft’s deep integration and investment while accessing AWS’s massive scale and infrastructure expertise. This multi-cloud approach provides OpenAI with greater negotiating power, technical flexibility, and reduced risk of infrastructure bottlenecks constraining AI development.
Financial and Market Implications
The $38 billion commitment represents one of the largest cloud infrastructure deals in history and underscores the extraordinary capital requirements of frontier AI development. This agreement will grow over its seven-year term as OpenAI’s compute needs continue expanding with each successive model generation.
For AWS, the partnership strengthens its position in the competitive AI infrastructure market and demonstrates the company’s ability to support the most demanding workloads at unprecedented scale. AWS CEO Matt Garman characterized the deal as “a powerful reminder of why organizations trust AWS when they need serious scale, security, and performance for their most demanding workloads”.
The announcement followed AWS’s strong third quarter 2025 financial results, which showed $33 billion in total sales, up 20% year over year, with operating income reaching $11.4 billion. The OpenAI partnership positions AWS to maintain growth momentum as AI infrastructure becomes an increasingly significant revenue driver for cloud providers.
For Microsoft, the partnership represents both a challenge and confirmation of OpenAI’s value. While Microsoft retains its 27% ownership stake and Azure relationship, the AWS deal demonstrates OpenAI’s intent to diversify infrastructure dependencies rather than remain exclusively tied to a single cloud provider.
Industry analysts view the partnership as validation of the multi-cloud trend in AI development, where companies hedge infrastructure risks and optimize performance by leveraging multiple providers’ strengths. The deal also highlights the extreme capital intensity of AI leadership, with infrastructure investments reaching unprecedented levels as companies compete to build the most capable models.
Timeline and Deployment Details
Immediate (November 2025): OpenAI begins utilizing AWS compute resources immediately following the partnership announcement. Initial workloads include ChatGPT inference and select training operations on available GPU capacity.
Q1-Q4 2026: AWS progressively deploys contracted infrastructure, including the buildout of EC2 UltraServers clusters with hundreds of thousands of NVIDIA GB200 and GB300 GPUs. OpenAI gradually migrates additional workloads and training operations to AWS infrastructure during this period.
End of 2026: Target date for full deployment of all committed capacity under the initial $38 billion agreement. At this milestone, OpenAI will have access to the complete infrastructure footprint specified in the partnership terms.
2027 and Beyond: Provision for further expansion beyond the initial capacity commitment as OpenAI’s compute requirements continue growing with new model development and increased user demand. The seven-year partnership provides framework for ongoing scaling through 2031.
The phased deployment approach balances OpenAI’s immediate needs with the practical realities of building massive-scale infrastructure. AWS’s experience deploying AI-optimized data centers positions the company to meet aggressive timelines while maintaining reliability and performance standards.
Technical Architecture and Performance
AWS designed OpenAI’s infrastructure deployment with a sophisticated architecture optimized for maximum AI processing efficiency. The core innovation lies in clustering NVIDIA GPUs via Amazon EC2 UltraServers on unified networks, enabling low-latency performance across interconnected systems.
This clustering approach allows OpenAI to treat distributed GPU resources as cohesive computing environments, essential for training models that exceed the memory and processing capacity of individual servers. The NVLink interconnect technology within each UltraServer provides high-bandwidth communication between GPUs, minimizing the performance penalties typically associated with distributed computing.
The inclusion of both NVIDIA GB200 and GB300 GPUs provides OpenAI with cutting-edge tensor processing capabilities optimized for AI workloads. These Blackwell-generation processors deliver substantial improvements in FP8 compute performance and memory bandwidth compared to previous generations, directly translating to faster training times and lower inference latency.
AWS’s third generation EC2 UltraClusters integrate these UltraServers with ultra high speed networking (up to 28.8 Tbps via EFAv4) and liquid cooling systems that enable sustained peak performance. The liquid cooling represents AWS’s first large scale deployment of this technology and proves essential for managing the thermal output of densely packed GPU clusters.
The infrastructure’s flexibility constitutes another critical advantage, with clusters designed to support varied workloads from real-time inference serving to multi-week training runs. OpenAI can dynamically allocate resources based on operational priorities, maximizing utilization and cost efficiency across the massive infrastructure footprint.
Security and Compliance Considerations
AWS brings extensive security expertise to the partnership, with infrastructure designed to meet the rigorous requirements of enterprise and government customers. The AWS Nitro System, which powers P6e-GB200 UltraServers, provides hardware level isolation and security for compute instances.
OpenAI’s workloads will benefit from AWS’s comprehensive compliance certifications, including HIPAA, GDPR, ISO standards, and SOC 2, ensuring the infrastructure meets regulatory requirements across different jurisdictions. These certifications prove particularly important as OpenAI expands ChatGPT and other services into regulated industries like healthcare and finance.
Data encryption in transit and at rest, combined with AWS’s identity and access management systems, protects sensitive training data and model weights throughout the development lifecycle. AWS’s global infrastructure also enables data residency compliance, allowing OpenAI to maintain data within specific geographic regions when required by local regulations.
The partnership agreement includes provisions for dedicated, AI-optimized infrastructure guaranteeing OpenAI exclusive access to specific resources. This isolation ensures OpenAI’s workloads aren’t impacted by other AWS customers and provides additional security boundaries around proprietary model development.
Industry Response and Competitive Landscape
The AWS-OpenAI partnership has sent ripples through the cloud computing and AI industries, with competitors reassessing their positions in the escalating infrastructure arms race. Microsoft, despite its continued investment in OpenAI, now faces increased competition for AI infrastructure workloads from its primary cloud rival.
Google Cloud and other providers are expected to pursue similar mega-deals with AI companies seeking diversified infrastructure options. The partnership validates the strategic importance of AI workloads for cloud providers and signals that infrastructure capacity may become a competitive differentiator in the AI race.
For OpenAI’s competitors including Anthropic, Google DeepMind, and Meta AI, the AWS partnership demonstrates the capital requirements and infrastructure partnerships necessary to compete at the frontier of AI development. Companies without access to similar compute resources may struggle to train competitive models as parameter counts and training dataset sizes continue expanding.
The NVIDIA GPU market also receives validation from the partnership, with OpenAI specifically choosing NVIDIA’s latest Blackwell architecture for its infrastructure buildout. This preference reinforces NVIDIA’s dominance in AI accelerators despite increasing competition from AMD, Intel, and custom silicon providers.
Future Implications for AI Development
This partnership establishes a precedent for AI infrastructure at unprecedented scale and suggests future frontier models will require even more substantial compute resources. The seven-year commitment horizon indicates OpenAI anticipates sustained growth in computational requirements through at least 2031.
The shift toward agentic AI workloads specifically mentioned in the partnership signals OpenAI’s development priorities beyond ChatGPT. These autonomous agents represent the next evolution of AI capabilities, requiring different infrastructure characteristics with emphasis on CPU resources for task execution alongside GPU power for reasoning.
OpenAI’s multi-cloud strategy may become standard practice for major AI companies, with diversification viewed as essential for ensuring uninterrupted access to compute resources. This trend could reshape cloud provider dynamics, encouraging competition and innovation in AI-optimized infrastructure.
The partnership also highlights the interdependence of AI advancement and infrastructure innovation. AWS’s development of specialized systems like EC2 UltraServers directly enables AI companies to pursue more ambitious models, while AI workload demands drive infrastructure providers to push technical boundaries.
As AI models continue growing in capability and application, partnerships of this scale and structure will likely become more common, potentially consolidating AI development among organizations with access to massive compute resources. This dynamic raises important questions about AI accessibility and whether smaller organizations can compete effectively in frontier model development.
FAQ Section (From PAA/Related Searches)
What does the AWS OpenAI partnership mean for ChatGPT users?
The partnership ensures ChatGPT has access to massive computing infrastructure for handling growing user demand, reducing latency, and improving response quality. AWS’s reliable, scalable infrastructure supports the real-time inference workloads required to serve millions of ChatGPT users simultaneously, with the low-latency performance of EC2 UltraServers enabling faster response times. Users can expect continued service reliability and potential performance improvements as OpenAI leverages this expanded infrastructure.
How does AWS compute infrastructure compare to Microsoft Azure for AI?
AWS and Azure both offer world-class AI infrastructure, but with different strengths. AWS provides massive scale with proven experience running clusters exceeding 500,000 chips and specialized offerings like EC2 UltraServers optimized for distributed GPU workloads. Azure excels in integration with Microsoft’s ecosystem, including Office 365 and Azure OpenAI Service, with strong AutoML capabilities and user-friendly interfaces. OpenAI’s multi-cloud approach leverages both providers’ advantages for optimal performance and flexibility.
What are agentic AI workloads mentioned in the partnership?
Agentic AI workloads refer to autonomous AI systems that can plan, execute tasks, interact with external tools, and make decisions with minimal human intervention. Unlike traditional AI that responds to specific prompts, agentic AI can pursue complex goals across multiple steps, requiring different infrastructure characteristics with emphasis on CPU resources for task execution alongside GPU power for reasoning. The AWS partnership specifically provisions tens of millions of CPUs to support OpenAI’s scaling of these next-generation capabilities.
Will OpenAI still use Microsoft Azure after the AWS partnership?
Yes, OpenAI will continue using Microsoft Azure alongside AWS as part of a multi-cloud infrastructure strategy. Microsoft remains OpenAI’s largest investor with 27% ownership in OpenAI Group and maintains a strategic partnership extending beyond compute infrastructure. The AWS deal represents diversification rather than replacement, with OpenAI leveraging multiple cloud providers to meet escalating compute demands, reduce dependency risks, and optimize workload performance.
What NVIDIA GPUs will OpenAI use on AWS?
OpenAI will use NVIDIA’s latest Blackwell-generation GPUs, specifically GB200 and GB300 models, via Amazon EC2 UltraServers. These GPUs feature Grace Blackwell Superchips that deliver 10 petaflops of FP8 compute per chip and up to 372 GB HBM3e memory. AWS will cluster up to 72 GPUs within single NVLink domains, providing 360 petaflops of compute per cluster for training trillion-parameter models and serving high-volume inference workloads.
How does this partnership affect other AI companies?
The AWS-OpenAI partnership sets new benchmarks for AI infrastructure scale and signals the extreme capital requirements for frontier model development. Competitors like Anthropic, Google DeepMind, and Meta AI face pressure to secure similar compute resources to remain competitive. The partnership may accelerate industry consolidation around organizations with access to massive infrastructure, while also driving cloud providers to compete aggressively for AI workload customers with specialized offerings and competitive pricing.
Featured Snippet Boxes
What is the AWS OpenAI partnership?
The AWS OpenAI partnership is a $38 billion, seven-year strategic agreement announced in November 2025 that provides OpenAI with immediate access to AWS cloud infrastructure, including hundreds of thousands of NVIDIA GB200/GB300 GPUs via Amazon EC2 UltraServers, to support ChatGPT operations and next-generation AI model training.
How much is the OpenAI AWS deal worth?
The OpenAI AWS deal is worth $38 billion over seven years, with all contracted compute capacity targeted for deployment by the end of 2026 and provisions for further expansion through 2027 and beyond.
Why did OpenAI partner with AWS?
OpenAI partnered with AWS to access massive, reliable compute infrastructure for scaling frontier AI, diversify beyond Microsoft Azure dependence, and leverage AWS’s proven experience running large-scale AI clusters exceeding 500,000 chips with superior security and performance.
What are Amazon EC2 UltraServers?
Amazon EC2 UltraServers are high-performance computing platforms that connect multiple EC2 instances via dedicated accelerator interconnects, supporting up to 72 NVIDIA Blackwell GPUs with 360 petaflops of FP8 compute and 13.4 TB of high-bandwidth memory for AI training and inference.
When will OpenAI start using AWS infrastructure?
OpenAI began using AWS infrastructure immediately following the November 2025 partnership announcement, with progressive deployment continuing through 2026 and full capacity expected by the end of 2026, with expansion into 2027 and beyond.
Does this partnership affect OpenAI’s relationship with Microsoft?
Microsoft remains a key investor with a 27% stake in OpenAI Group and continues as an infrastructure partner via Azure, but waived exclusive compute rights, allowing OpenAI to pursue a multi-cloud strategy with AWS for diversified infrastructure access.
Source: OpenAI
