OpenAI and Nvidia are reportedly in talks to create a $100 billion global network of AI data centers, with a new chip leasing model at the core of their partnership.
Instead of selling expensive AI chips directly, Nvidia may lease its high-performance GPUs to OpenAI, giving the company access to computing power without massive upfront costs. This move could revolutionise the way AI companies build infrastructure and scale their services.
The data centers—called “AI factories”—would be spread across multiple regions worldwide, ensuring faster AI processing, reduced latency, and better global access.
Industry experts believe this leasing model will lower barriers for AI adoption while creating a steady revenue stream for Nvidia and helping OpenAI expand more aggressively in the race against rivals like Anthropic, Google DeepMind, and Meta.
Why It Matters
- For OpenAI: Faster scaling without billions in upfront hardware purchases.
- For Nvidia: Recurring revenue instead of one-time sales, securing long-term dominance in the AI chip market.
- For Businesses: Affordable access to cutting-edge AI computing.
Summary Table
| Key Detail | Information |
|---|---|
| Partnership | OpenAI & Nvidia |
| Investment Size | $100 Billion |
| Business Model | Chip leasing (GPUs rented, not sold) |
| Infrastructure | Global AI data centers (“AI factories”) |
| Benefits to OpenAI | Reduced upfront costs, faster scaling |
| Benefits to Nvidia | Recurring revenue, stronger market dominance |
| Industry Impact | Lower entry barriers, global AI adoption boost |
| Rivals | Anthropic, Google DeepMind, Meta |








