Google Places Order for Million Chips, Aiming to Bypass NVIDIA and AMD

kyojuro czwartek, 30 października 2025

Anthropic and Google Cloud have unveiled a major partnership centered around TPU technology, committing to provide up to one million TPU chips alongside comprehensive cloud services throughout the contract's duration. This initiative aims to deliver "more than a gigawatt of computing power" by the next year. Google considers this commitment its largest individual deal involving custom AI chips, marking a significant industry milestone outside of the conventional NVIDIA and AMD technology transactions. Anthropic, which has exclusively relied on Google Cloud since 2023, will now expand its AI capabilities within the same technological framework, ensuring rapid capacity enhancement while minimizing migration expenses.

Unlike general-purpose GPUs, TPUs are specifically designed as ASICs for machine learning tasks. Anthropic continues to rely on TPUs, citing "price/performance and efficiency" as key advantages while also highlighting its thorough experience in using TPUs for training and deploying models. The strategic plan to utilize "up to one million TPUs" combined with Google Cloud's complementary services suggests that Anthropic's research and production will benefit from highly integrated AI infrastructure. In the broader industry landscape, NVIDIA's CEO, Jensen Huang, has recognized Google TPU alongside other specialized chips like Amazon Trainium as central to market competition. Anthropic's commitment to TPUs solidifies its position as Google's largest external TPU customer, significantly impacting AI application trends on the frontlines.

Maintaining the established TPU architecture allows a seamless continuation of toolchains, compilers, and server frameworks, reducing the demand for cross-platform optimization. Utilizing this expanded computing power will facilitate enhanced model training and online services, touted by the company as "equipping our R&D team with a leading-edge AI optimization infrastructure for years ahead." The projection of channeling more than a gigawatt of computing power "as early as next year" highlights a phased rather than one-time delivery, synchronized with Google Cloud's data center growth and supply-demand dynamics. Contrastingly, OpenAI and Oracle continue to pursue "multi-GW" level agreements with NVIDIA, indicating the emergence of dual pathways within the industry to either scale GPU clusters or harness ASICs for targeted scenarios.

This venture importantly symbolizes diversification on the supply side of AI technology. With AI models expanding rapidly in size and inference throughput, leading AI companies have begun integrating ASICs alongside general-purpose GPUs to achieve a more predictable cost efficiency and energy curve. This shift aligns with the "GPU vs. ASIC" debate emphasized by Jensen Huang: while GPUs boast mature ecosystems and versatility, ASICs offer tailored efficiency and performance. Despite speculative interpretations surrounding Anthropic's strategic pivot from NVIDIA, the official stance stands: Anthropic emerges as the largest TPU customer, providing Google Cloud's self-developed chip ecosystem crucial operational scale.

Collectively, Anthropic's collaboration with Google Cloud regarding a million TPU commitment exemplifies the advancing commercialization of non-GPU computing in large model training and service contexts. In the immediate term, it equips Anthropic with substantial new resources and reduces intra-stack migration costs; in the longer perspective, it could drive TPU ecosystem advances in compiling, scheduling, and maintenance tools. Overarching these developments is the shifting market landscape shaped by the concurrent advancement of both GPU and ASIC technologies. As this debate evolves, the enhancement of supply chains and the optimization of cost structures stand to become pivotal factors driving AI industry evolution.

Powiązane wiadomości

© 2025 - TopCPU.net