
Google Cloud chief reveals the long game: a decade of silicon and the energy battle behind the AI boom | Fortune
While the world scrambles to adapt to the explosive demand for generative AI, Google Cloud CEO Thomas Kurian says his company isnāt reacting to a trend, but rather executing a strategy set in motion 10 years ago. In a recent panel for Fortune Brainstorm AI , Kurian detailed how Google anticipated the two biggest bottlenecks facing the industry today: the need for specialized silicon, and the looming scarcity of power. According to Kurian , Googleās preparation began well before the current hype cycle. āWeāve worked on TPUs since 2014 ... a long time before AI was fashionable,ā Kurian said, referring to Googleās custom Tensor Processing Units. The decision to invest early was driven by a fundamental belief that chip architecture could be radically redesigned to accelerate machine learning. The energy premonition Perhaps more critical than the silicon itself was Googleās foresight regarding the physical constraints of computing. While much of the industry focused on speed, Google was calculating the electrical cost of that speed. āWe also knew that the most problematic thing that was going to happen was going to be energy because energy and data centers were going to become a bottleneck alongside chips,ā Kurian said. This prediction influenced the design of their infrastructure. Kurian said Google designed its machines āto be super efficient in delivering the maximum number of flops per unit of energy.ā This efficiency is now a critical competitive advantage as AI adoption surges, placing unprecedented strain on global power grids. Kurian said the energy challenge is more complex than simply finding more power, noting that not all energy sources are compatible with the specific demands of AI training. āIf youāre running a cluster for training ... the spike that you have with that computation draws so much energy that you canāt handle that from some forms of energy production,ā he said. To combat this, Google is pursuing a three-pronged strategy: diversifying energy sources, utilizing AI to manage thermodynamic exchanges within data centers, and developing fundamental technologies to create new forms of energy. In a moment of recursive innovation, Kurian said āthe control systems that monitor the thermodynamics in our data centers are all governed by our AI platform.ā The āzero sumā fallacy Despite Googleās decade-long investment in its own silicon, Kurian pushed back against the narrative that the rise of custom chips threatens industry giants like Nvidia . He argues that the press often frames the chip market as a āzero sum game,ā a view he considers incorrect. āFor those of us who have been working on AI infrastructure, thereās many different kinds of chips and systems that are optimized for many different kinds of models,ā Kurian said. He characterized the relationship with Nvidia as a partnership rather than a rivalry, noting that Google optimizes its Gemini models for Nvidia GPUs and recently collaborated to allow Gemini to run on Nvidia clusters while protecting Googleās intellectual property. āAs the market grows,ā he said, āweāre creating opportunity for everybody.ā The full stack advantage Kurian attributed Google...
Preview: ~500 words
Continue reading at Fortune
Read Full Article