Google's TPU Spurs AI Cost Revolution, Challenging OpenAI's Nvidia Dependence

Author's Avatar
May 13, 2025
Article's Main Image

As the AI landscape intensifies, companies are focusing on cost-efficiency when integrating AI technologies. Google (GOOGL, Financial) recently introduced its Tensor Processing Unit (TPU), significantly lowering AI compute costs to just 20% of what OpenAI spends relying on Nvidia's (NVDA) GPUs. Currently, a single Nvidia H100 chip costs about $3,000, with market prices soaring up to $35,000, placing a heavy financial burden on enterprises. In contrast, Google’s aggressive pricing strategy for TPU is rapidly attracting business users.

Beyond hardware, API pricing plays a crucial role in platform selection. Google's Gemini 2.5 Pro API is considerably cheaper than OpenAI's o3 model, allowing small to medium enterprises to implement generative AI more affordably. Google promotes an open AI ecosystem with its platforms, contrasting with OpenAI's integrated approach with Microsoft services like Azure and Office 365.

From a technical perspective, Google's Gemini 2.5 Pro excels in handling extensive data, while OpenAI's o3 is noted for logical reasoning. In application, Google leverages its Cloud platform to integrate AI features for businesses, while OpenAI taps into global markets via ChatGPT and Microsoft 365 Copilot, boasting 800 million active users.

Disclosures

I/We may personally own shares in some of the companies mentioned above. However, those positions are not material to either the company or to my/our portfolios.