All revisions

Revision #1

System

about 3 hours ago

OpenAI's $18 Billion Chip Gambit Hits a Wall — and It Reveals the Fragile Economics of the AI Boom

OpenAI's plan to build its own custom AI chips — a bet central to the company's long-term financial viability — has run into a financing impasse that raises uncomfortable questions about the sustainability of the broader AI infrastructure boom.

The company's $18 billion chip partnership with Broadcom, code-named Project Nexus, has stalled because Broadcom will only finance the first phase of the buildout if Microsoft agrees to purchase roughly 40% of the custom chips [1][2]. Microsoft has not signed a firm purchase agreement, and OpenAI's head of compute, Sachin Katti, has said that Microsoft's involvement would make the deal "financially unattractive" and the proposed commercial structure "likely unworkable" [2][3].

The result is a catch-22 that neither side appears willing to resolve: the chip designer wants a guaranteed buyer, the buyer's participation would undermine the economics for the company that actually needs the chips, and the company that needs the chips cannot finance the project alone.

The Deal That Won't Close

Project Nexus centers on a custom inference chip called "Jalapeno," designed specifically to power ChatGPT and OpenAI's other AI products [1][4]. The first-generation chip, also referred to as "Titan," is being developed with Broadcom's ASIC design services and manufactured on TSMC's N3 (3-nanometer) process, with mass production targeted for late 2026 [5][6]. A second-generation chip, Titan 2, is planned for TSMC's more advanced A16 (1.6nm) process node, with deployment targeted for 2027 [5].

The $18 billion covers an initial tranche of a much larger buildout requiring approximately 1.3 gigawatts of data center capacity [1][2]. Under the proposed financing structure, Microsoft would buy the chips, install them in its Azure data centers, and rent the computing capacity back to OpenAI [2]. This arrangement would give Broadcom the financial security of a creditworthy buyer, but it would leave OpenAI paying Microsoft for access to chips OpenAI itself designed — an arrangement that Katti's team views as defeating the purpose of building custom silicon in the first place [3].

The disagreement also extends to physical infrastructure. OpenAI envisions specialized data centers optimized for its custom silicon, while Microsoft prefers standardized, versatile data center designs that can serve multiple customers and workloads [3]. This architectural divergence has made Microsoft reluctant to commit to chips that may not integrate cleanly with its broader infrastructure strategy.

Why OpenAI Needs Its Own Chips

The financial imperative behind Project Nexus becomes clear when examining OpenAI's spending trajectory. OpenAI President Greg Brockman testified under oath during the Musk-Altman trial that the company expects to spend $50 billion on computing power in 2026 alone [7]. Against projected 2026 revenue of approximately $29.4 billion, compute costs alone would consume nearly double the company's income [8][7].

OpenAI Revenue vs. Projected Compute Costs

OpenAI's own internal forecasts project cumulative losses of $44 billion over the 2023–2028 period, before an anticipated profit of $14 billion in 2029 [9]. The company's annualized revenue run rate reached $12.7 billion in Q1 2026, up from $3.7 billion in 2023, but infrastructure costs have grown faster than revenue [8].

Custom chips offer a potential path to closing this gap. Nvidia's Blackwell B200 delivers inference at approximately $0.02 per million tokens — about 4.5x cheaper than the H100 at $0.09 per million tokens [10]. But even Nvidia's latest hardware carries significant markup. A custom ASIC (application-specific integrated circuit) — a chip designed for a single type of workload rather than general-purpose computing — can deliver further cost reductions by eliminating unnecessary transistors and optimizing specifically for inference on OpenAI's model architectures [11].

The question is whether those savings justify an $18 billion upfront investment that would take years to pay back — and whether OpenAI can actually execute on chip design at the pace required to stay ahead of Nvidia's accelerating roadmap.

Microsoft's Strategic Retreat

The financing snag cannot be understood in isolation from the broader restructuring of the Microsoft-OpenAI relationship. In April 2026, the two companies announced a revamped partnership agreement that fundamentally altered their commercial arrangement [12][13]:

  • Microsoft's exclusive cloud hosting rights for OpenAI products were eliminated. OpenAI can now serve all of its products through Amazon, Google, or any other cloud provider [12].
  • Revenue share payments from OpenAI to Microsoft will continue through 2030 but are now subject to a total cap [12].
  • Microsoft's license to OpenAI's intellectual property extends through 2032 but is no longer exclusive [13].
  • The AGI clause — which previously allowed Microsoft to determine its response if OpenAI achieved artificial general intelligence — was dropped entirely [14].

From Microsoft's perspective, committing $7 billion to purchase chips for a partner that is actively diversifying away from Azure is a difficult proposition. Microsoft is also developing its own custom AI chip, Maia 200, which the company claims delivers 3x the FP4 performance of Amazon's Trainium3 and 30% better performance per dollar than its existing fleet [15]. Buying OpenAI's chips would mean investing in a competing silicon program while already funding its own.

The Competitive Landscape

OpenAI is entering the custom chip race years behind competitors who have already iterated through multiple hardware generations.

Custom AI Chip Program Investments
Source: Industry reports and company disclosures
Data as of May 1, 2026CSV

Google's TPU (Tensor Processing Unit) program is the most mature custom silicon effort in the industry, with its seventh generation (Ironwood) now deployed at scale [15][16]. Google has invested an estimated $25 billion or more cumulatively over more than a decade, and its TPUs are available to external customers through Google Cloud [16].

Amazon's Trainium/Inferentia program, built on its 2015 acquisition of Israeli chip startup Annapurna Labs, has evolved into an estimated $10 billion-plus annual run-rate business within AWS [16][17]. Trainium 3 is now sampling, and Amazon has committed $225 billion to its AI chip strategy [17].

Meta's MTIA (Meta Training and Inference Accelerator) takes a different approach: it is purpose-built for internal workloads like ad ranking and content recommendation, with no external commercialization [15]. Meta has three distinct chip generations either shipping or sampling in 2026 [15].

OpenAI's custom silicon team, by contrast, currently has approximately 40 people [5]. The team is led by Richard Ho, who previously spent nine years at Google leading the TPU program and later served as SVP at photonic computing company Lightmatter [18][19]. Ho brings deep expertise, but building a competitive chip organization from 40 engineers — while Google's TPU team numbers in the thousands — represents a structural disadvantage.

If Project Nexus stalls or is scaled back, OpenAI risks losing the hardware talent it has recruited. The AI chip job market remains intensely competitive, and engineers who joined to build production silicon are unlikely to stay for an indefinite research project with uncertain funding [18].

The Case for Microsoft's Caution

Microsoft's hesitation has a rational basis that extends beyond the bilateral relationship with OpenAI. No company outside the established hyperscalers — Google, Amazon, Meta, and Microsoft itself — has successfully brought a custom AI chip from design to production deployment at scale. The track record of custom silicon projects across the industry is littered with delays and cost overruns [20].

Nvidia's roadmap continues to accelerate. The company's Vera Rubin architecture, planned for 2026-2027, targets 50 petaflops of FP4 performance with 288GB of HBM4 memory [15]. Each Nvidia generation has delivered 4-15x inference improvements over its predecessor [10]. For Microsoft, the question is whether OpenAI's first-generation custom chip — which will have no production track record when it arrives — can outperform Nvidia hardware that benefits from decades of software ecosystem investment and continuous iteration.

Microsoft's CFO has also attributed $25 billion of the company's record capital expenditure budget to rising memory chip prices [20]. Committing additional billions to an unproven chip program while managing existing infrastructure costs represents a concentrated risk.

Furthermore, Microsoft now holds its own chip program in Maia 200, which offers the company direct control over architecture, timeline, and cost — without the governance complexities of a joint venture with an increasingly independent partner [15].

Alternative Funding Paths

If Microsoft does not commit, OpenAI has several potential alternative financing routes, each with significant strings attached.

SoftBank has already invested over $60 billion in OpenAI, including a $30 billion commitment as part of the company's record $110 billion funding round in February 2026 [21][22]. SoftBank took on a $40 billion loan to cover its OpenAI commitment, but has recently scaled back a separate $10 billion margin loan backed by OpenAI shares to $6 billion after lender pushback [23][24]. SoftBank's appetite for additional OpenAI exposure may be reaching its limits.

The Stargate Project, a $500 billion joint venture with SoftBank, Oracle, and Abu Dhabi-based MGX, provides an off-balance-sheet vehicle for infrastructure spending [25][26]. Stargate has already surpassed 10GW of committed AI infrastructure capacity [26]. However, Stargate is designed for data center buildout, not chip manufacturing — and its governance structure means OpenAI does not have unilateral control over spending decisions.

Saudi Arabia's Public Investment Fund (PIF) has been in discussions to participate in OpenAI funding rounds [27]. Saudi Arabia has designated 2026 as its national Year of Artificial Intelligence, and the PIF has channeled capital into AI infrastructure including a $2.7 billion data center contract in Riyadh [28][29]. However, Saudi capital comes with geopolitical considerations — including US government scrutiny over AI chip export controls and data sovereignty requirements — that could complicate OpenAI's regulatory position.

A standalone bond issuance is theoretically possible given OpenAI's $852 billion valuation [8], but the company's projected $14 billion loss in 2026 [9] and negative free cash flow profile would likely command significant yield premiums.

What This Means for the AI Capex Cycle

The financing impasse at OpenAI arrives as the broader AI infrastructure boom reaches a critical inflection point. The five largest US cloud and AI infrastructure providers have collectively committed to spending between $660 billion and $720 billion on capital expenditure in 2026, nearly doubling 2025 levels [20][30].

Hyperscaler AI Capital Expenditures

But the gap between spending growth and revenue growth is widening. In 2025, revenues for Alphabet, Amazon, Meta, and Microsoft grew an average of 16.5%, while capex growth averaged 60%. The projected 2026 figures are more extreme: revenue growth averaging 15.5% against capex growth averaging 80% [20]. Free cash flow is trending downward across the hyperscaler cohort [20].

Nvidia has reported $33 billion in customer receivables, reflecting delayed payment settlements — a signal that even Nvidia's customers are straining to keep up with the pace of spending [4]. If OpenAI — a company valued at $852 billion, backed by SoftBank's $60 billion and Microsoft's $13 billion, and anchored by a $500 billion infrastructure joint venture — cannot close an $18 billion chip commitment, it raises a question the market has largely avoided: at what point does AI capex exceed the industry's ability to finance it?

The answer matters beyond OpenAI. Custom ASIC programs from Google, Amazon, Microsoft, and Meta are growing at an estimated 44.6% compound annual growth rate, targeting the inference workloads that now represent two-thirds of all AI compute [15]. Analysts project Nvidia's inference market share could fall from over 90% to 20-30% by 2028 as these custom programs scale [15]. But that projection assumes the capital to fund these programs continues to flow — an assumption the OpenAI-Broadcom impasse has called into question.

The Path Forward

OpenAI faces three realistic options: restructure the financing terms to make Broadcom comfortable without Microsoft's guaranteed purchase agreement, find an alternative anchor buyer to replace Microsoft's role, or reduce the initial scope of the chip rollout [4].

None of these options is cost-free. Restructuring the deal likely means accepting worse economics. Finding an alternative buyer means ceding strategic leverage. Scaling back the program means delaying the cost savings that custom silicon was supposed to deliver — and potentially losing the hardware engineers who signed on to build chips, not wait for financing to close.

The deeper issue is structural. OpenAI has positioned itself as a company that needs to spend $50 billion a year on compute [7], but it generates less than $30 billion in revenue [8] and relies on a network of investors, joint ventures, and partnership agreements to bridge the gap. Each of those relationships carries its own incentives and constraints. When the interests of the chip designer, the chip manufacturer, the cloud partner, and the financial backers all need to align for a single deal to close, the probability of impasse increases with every additional stakeholder.

Project Nexus may yet find its financing. But the fact that it has stalled — at this scale, involving these counterparties, at this moment in the AI investment cycle — is itself a data point that the market will need to price.

Sources (30)

  1. [1]
    OpenAI and Broadcom in discussions over financing of $18bn custom chip projectdatacenterdynamics.com

    Broadcom has told OpenAI it will only finance the first phase if Microsoft agrees to purchase roughly 40% of the custom chips.

  2. [2]
    OpenAI's AI Chip Deal With Broadcom Hits $18 Billion Financing Snagwinbuzzer.com

    The deal requires 1.3 gigawatts of data center capacity. Microsoft would buy chips, install them, and rent capacity back to OpenAI.

  3. [3]
    OpenAI's AI Chip Deal With Broadcom Hits $18 Billion Financing Snagtheinformation.com

    Sachin Katti, OpenAI's head of compute, said Microsoft involvement would make the deal financially unattractive and likely unworkable.

  4. [4]
    OpenAI's $18 Billion Chip Financing Trouble Rattles AI Boom Narrativefinance.yahoo.com

    The financing gap represents an early tranche of a multi-year accelerator covering 10 gigawatts of OpenAI-designed chips. Nvidia reported $33 billion in customer receivables.

  5. [5]
    OpenAI Reportedly to Deploy Custom AI Chip on TSMC N3 by End-2026trendforce.com

    OpenAI's custom chip team has approximately 40 people. First-gen chip on TSMC N3 process, second-gen Titan 2 planned for A16 (1.6nm) process.

  6. [6]
    What's Behind OpenAI's First Custom Chip Design With TSMC?aimagazine.com

    OpenAI's chip development leverages Broadcom's ASIC design services with TSMC manufacturing, targeting mass production in second half of 2026.

  7. [7]
    OpenAI Projects $50 Billion in Compute Spending for 2026opentools.ai

    Greg Brockman revealed under oath that OpenAI expects to spend $50 billion on computing power in 2026, representing roughly 1.25 million B200 GPUs.

  8. [8]
    OpenAI revenue, valuation & fundingsacra.com

    OpenAI's post-money valuation went from $157B in October 2024 to $852B in March 2026. Annualized revenue run rate reached $12.7 billion in Q1 2026.

  9. [9]
    OpenAI's own forecast predicts $14 billion loss in 2026finance.yahoo.com

    OpenAI expects to lose $44 billion cumulatively from 2023 through 2028 before turning a $14 billion profit in 2029.

  10. [10]
    Benchmarking LLM Inference on NVIDIA B200, H200, H100medium.com

    NVIDIA B200 delivers inference at approximately $0.02 per million tokens, roughly 4.5x cheaper than H100 at $0.09 per million tokens.

  11. [11]
    Inside the Custom AI Chip Race: Google, AWS, Microsoft, Meta, OpenAIhashrateindex.com

    Comprehensive comparison of hyperscaler custom AI chip programs including Google TPU, Amazon Trainium, Microsoft Maia, and Meta MTIA.

  12. [12]
    OpenAI shakes up partnership with Microsoft, capping revenue share paymentscnbc.com

    OpenAI can now serve all products across any cloud provider. Revenue share payments to Microsoft subject to a total cap through 2030.

  13. [13]
    Microsoft and OpenAI end exclusivity agreementtomshardware.com

    Microsoft's license to OpenAI's IP extends through 2032 but is no longer exclusive. OpenAI can now partner with Amazon, Google, and others.

  14. [14]
    Microsoft and OpenAI Drop AGI Clause from Agreementaitoolly.com

    The AGI clause that allowed Microsoft to determine its response if OpenAI achieved AGI was dropped from the restructured agreement.

  15. [15]
    The Custom AI Chip Race in 2026: Meta, Google, Amazon, and Microsoft vs. Nvidianerdleveltech.com

    Custom ASICs grow at 44.6% CAGR. Microsoft Maia 200 claims 3x FP4 performance of Trainium3 and 30% better price-performance than existing fleet.

  16. [16]
    Google, Amazon and others now make their own custom AI chipscnbc.com

    Google's TPU program is the most mature custom silicon effort. Amazon's Trainium has 30-40% better price performance than rival hardware on AWS.

  17. [17]
    Amazon Trainium 3: The $225 Billion Chip Bet Threatening NVIDIA's Dominanceoplexa.com

    Amazon has committed $225 billion to its AI chip strategy. Trainium evolved from quiet experiment to $10B+ annual run-rate business.

  18. [18]
    OpenAI appoints former Google TPU lead as head of hardwaredatacenterdynamics.com

    Richard Ho spent nine years at Google leading the TPU program. Previously SVP at photonic computing company Lightmatter.

  19. [19]
    Lightmatter Names Top Google TPU Engineer Richard Ho as VP of Hardware Engineeringbusinesswire.com

    Richard Ho co-invented methods for using machine learning to design chip architectures at Google. Earlier co-founded 0-In Design Automation.

  20. [20]
    Big Tech's $700 billion AI spending spree has no clear end in sightfortune.com

    Hyperscalers have committed $660-720 billion in 2026 capex. Revenue growth of 15.5% vs. capex growth of 80% raises sustainability concerns.

  21. [21]
    SoftBank Is in Talks to Invest $30 Billion More in OpenAIbloomberg.com

    SoftBank's $30 billion commitment brings its total OpenAI investment to over $60 billion as part of OpenAI's $110 billion funding round.

  22. [22]
    OpenAI raises $110B in one of the largest private funding rounds in historytechcrunch.com

    OpenAI's record $110 billion raise included $50B from Amazon, $30B from Nvidia, and $30B from SoftBank against a $730 billion valuation.

  23. [23]
    SoftBank cuts OpenAI-backed margin loan target from $10B to $6Bamericanbazaaronline.com

    SoftBank scaled back its OpenAI-backed margin loan target after lender pushback over AI valuations and debt exposure.

  24. [24]
    Why SoftBank's new $40B loan points to a 2026 OpenAI IPOtechcrunch.com

    SoftBank took on a $40 billion loan to cover its $30 billion OpenAI commitment.

  25. [25]
    Announcing The Stargate Projectopenai.com

    The Stargate Project is a $500 billion AI infrastructure joint venture with SoftBank, Oracle, and MGX, planned over four years.

  26. [26]
    OpenAI, Oracle, and SoftBank expand Stargate with five new AI data center sitesopenai.com

    OpenAI has surpassed 10GW of committed AI infrastructure capacity, with more than 3GW added in the last 90 days.

  27. [27]
    OpenAI in talks with Saudi's PIF, UAE's MGX for $40 billion funding roundmenabytes.com

    Saudi Arabia's PIF and UAE's MGX among key investors being courted for OpenAI's massive funding round led by SoftBank.

  28. [28]
    Saudi Arabia's PIF Recalibrates 2026 Strategy, Channelling Investment into AIthesauditimes.net

    PIF identified AI as receiving the highest concentration of capital in 2026. Saudi Arabia designated 2026 as the Year of Artificial Intelligence.

  29. [29]
    PIF-backed Humain secures up to $1.2bn to expand AI infrastructurearabnews.com

    Saudi Arabia's Humain received $1.2 billion financing framework for 250 megawatts of AI data center capacity.

  30. [30]
    AI Capex 2026: The $690B Infrastructure Sprintfuturumgroup.com

    The five largest US cloud providers committed $660-690 billion in 2026 capex, nearly doubling 2025 levels.