Two chips cover training and inference workloads, with Google claiming speed and cost gains over prior generation.
Briefing
Google first deployed TPUs internally for its own workloads before offering them externally on Google Cloud. The internal deployment reduced Google's own Nvidia GPU consumption but did not materially dent Nvidia's hyperscaler revenue because AWS and Azure continued large-scale GPU purchases. The parallel strategy then mirrors the current dual-offering approach.
AWS Trainium and Inferentia chips launched as Nvidia alternatives on AWS. Adoption remained limited relative to Nvidia GPU instances because developer tooling and software ecosystems favored CUDA. Google's TPU faces the same ecosystem lock-in problem, as most AI training frameworks are CUDA-optimized, constraining switching speed.

Victory Giant's 50% Hong Kong IPO debut, driven by its status as a Nvidia PCB supplier, shows how concentrated investor exposure to the Nvidia supply chain has become. A credible TPU alternative gaining hyperscaler share introduces a valuation risk for that supply chain premium that the IPO pricing did not reflect.

Tim Cook's departure and John Ternus's appointment as Apple CEO raised immediate questions about Apple's AI silicon acceleration. A hardware-focused CEO pursuing faster on-device AI inference directly competes with both Nvidia GPU deployments and Google TPU cloud inference for agentic AI workloads.
7 hours ago