Google's ultra-private CEO Sundar Pichai is showing me around Googleplex, its California headquarters. A walkway runs along the length of it, passing by a giant dinosaur skeleton, a beach volleyball pitch and dozens of Googlers lunching under the hazy November sun.
But it's a laboratory, hidden away at the back of the campus behind some trees, that he is most excited for me to see.
This is where the invention that Google believes is its secret weapon is being developed.
Known as a Tensor Processing Unit (or TPU), it looks like an unassuming little chip but, says Mr Pichai, it will one day power every AI query that goes through Google. This makes it potentially one of the most important objects in the world economy right now.
"AI is the most profound technology humanity [has ever worked] on," he insists. "It has potential for extraordinary benefits - we will have to work through societal disruptions."
But the confusing question lingering over the AI hype is whether it is a bubble at risk of bursting - as, if so, it may well be a spectacular burst akin to the dotcom crash at the start of the century, with consequences for us all.
Bloomberg via Getty ImagesA walkway runs along the length of Googleplex, passing by a giant dinosaur skeleton, a beach volleyball pitch and staff lunching in the winter sun
The Bank of England has already warned of a "sudden correction" in global financial markets, saying "market valuations appear stretched" for tech AI firms. Meanwhile. OpenAI boss Sam Altman has speculated that "there are many parts of AI that I think are kind of bubbly right now".
Asked whether Google would be immune from a potential bubble burst, Mr Pichai said it could weather that potential storm - but for all his starry-eyed excitement around the possibilities of AI, he also issued a warning: "I think no company is going to be immune, including us."
So why, then, is Google investing more than $90bn a year in the AI build-out, a three-fold increase in just four years, at the very moment these suggestions are being discussed?
The big AI surge - and the big risk
The AI surge - of which Google is just one part - is, in cash terms, the biggest market boom the world has seen.
Its numbers are extraordinary - there is $15 trillion of market value at Google and four other tech giants whose headquarters are all within a short drive of one another.
Chipmaker turned AI systems pioneer Nvidia in Santa Clara is now worth more than $5 trillion. A 10-minute drive south, in Cupertino, is Apple HQ, hovering around $4 trillion; while 15 minutes west is $1.9 trillion Meta (previously Facebook). And in the centre of San Francisco, OpenAI was recently valued at $500bn.

Google's parent firm Alphabet, headquartered in Mountain View, is worth about $3.3 trillion, and has almost doubled in value since April, (which every Googler on campus will no doubt be feeling through the value of their stock options)
The purely financial consequences of this trend are significant enough.
The value of the shares in these companies (and a few others outside Silicon Valley, such as Microsoft in Seattle) have helped cushion the US economy from the impact of trade wars, and kept retirement plans and investments buoyant - and not just in the US.
Yet it comes with a big risk. That is, the incredible dependence of US stock market growth on the performance of a handful of tech giants. The Magnificent 7 - Alphabet, Amazon, Apple, Meta, Microsoft, Nvidia, and Tesla - collectively comprise one third of the valuation of America's entire S&P 500.
And that market value is now more highly concentrated in a few firms than it was during the dotcom bubble in 1999, according to the IMF.
Mr Pichai points out that every decade or so come these "inflection points": the personal computer, then the internet in the late 1990s, followed by mobile and cloud. "Now it's clearly the era of Artificial Intelligence."
But as for the big question - is it a bubble?
Mr Pichai argues there are two ways of thinking about it. First, there is "palpably exciting" progress of services that people and companies are using.
But he concedes: "It's also true when we go through these investment cycles, there are moments we overshoot collectively as an industry…
"So I think it's both rational and there are elements of irrationality through a moment like this."

Sundar Pichai: "We have this phrase at Google, which is 'uncomfortably exciting'."
Now, a distinction is emerging in the markets between those businesses that rely on often borrowed money and complicated deals to access the chips that power their AI, and the biggest tech companies, such as Google, Microsoft and Amazon, which can fund investment in chips and data from their own pockets.
Which brings us to Google's own silicon chips, or their prized TPUs.
'Restricted': inside the silicon chip lab
The lab, where they are tested, is the size of a five-a-side football pitch with a mesh of multi-coloured wires and deep blue blinking lights. Signs all around read: "restricted".
What's striking is the sheer noise - this is down to the cooling systems, which are needed to help control the temperature of the chips, which can get incredibly hot when crunching trillions of calculations.

Google's TPU cluster is developing the chip that could soon power all its AI searches
The TPUs are designed to help power AI machines. And they work differently from other types of chips.
The CPU (central processing unit) is the primary component of a computer - essentially its brain - that performs most of the processing and control functions, while GPUs (graphics processing units) perform more specialised processing, executing many parallel tasks at once - this can include AI.
However Asics (application-specific integrated circuits), are chips custom-built for a specific purpose, for example, a specific AI algorithm. And the TPU is a specialist Google-designed type of Asic.

There are several versions of TPUs: the Ironwood is the latest. The TPUs are part of Sundar Pichai's overall strategy of owning the entire scientific supply chain - from the silicon to the data, plus the AI models and everything in between
A core aspect of the AI boom has been the mad dash to amass lots of top-performing chips and put them into data centres (or the physical facilities that store, process and run large amounts of data and software).
Nvidia's boss Jensen Huang once coined the term "AI factories" to describe the massive data centres full of pods and racks of super chips, connected to huge energy and cooling systems.
(Tech bosses such as Mark Zuckerberg have referred to some being the size of Manhattan. The Google TPU lab is somewhat more modest, testing out the technology for deployment elsewhere.)
Stories abound of tech bros begging chip makers for hundreds of thousands of these highly engineered pieces of silicon. Take the recent dinner at Nobu in Palo Alto, where Elon Musk and Larry Ellison, the founder and head of Oracle, tried to woo Nvidia's Jensen Huang, to sell them more of them.
As Mr Ellison put it: "I would describe the dinner as me and Elon begging Jensen for GPUs. Please take our money - no, no take more. You're not taking enough. We need you to take more, please!"
It is precisely the race to access the power of as many as possible of these high performance chips, and to scale them up into massive data centres, that is driving an AI boom - and there's a perception that the only way to win is to keep spending.
The chips race - and the OpenAI storm
The terrace of the Rosewood Sand Hill hotel, a sprawling 16-acre estate near the Santa Cruz mountains that serves crab rolls and $35 signature vodka martinis, is where the big Silicon Valley deal-making gets done. It's close to Stanford University and Meta's HQ, as well as the headquarters of major venture capital firms.
There are whispered rumours about who will be next to announce customised AI chips - Asics - to compete with Google and Nvidia.
Just before I visited, something of a storm was brewing about the investment plans of OpenAI, which Elon Musk co-founded.
The firm, which started as a not-for-profit but has since established a commercial structure, has been the focus of a web of cross-investments involving buying up chips and other computer hardware needed for AI processing.
Few in the industry doubt OpenAI's phenomenal user growth - in particular the popularity of its chatbot, ChatGPT. It has ambitions to design its own custom AI chips, but some have speculated about whether it might need government support to achieve this.
Getty ImagesSam Altman: 'What we do think might make sense is governments building (and owning) their own AI infrastructure'
In a podcast episode that aired last month, an OpenAI investor questioned how the company's spending commitments tallied with its revenues, to which co-founder Sam Altman shot back, challenging the revenue figures quoted, and adding: "If you want to sell your shares, I'll find you a buyer. Enough."
He has since shared a lengthy post on X, explaining, among other things, that OpenAI is looking at commitments of about $1.4 trillion over the next eight years and why he believes now is the time to invest in scaling up their technology.
"I do not think the government should be writing insurance policies for AI companies," he said.
But he also said: "What we do think might make sense is governments building (and owning) their own AI infrastructure."
Getty ImagesElon Musk and Larry Ellison are said to have begged Nvidia's Jensen Huang to sell them more top-performing chips during a dinner in Nobu
Elsewhere, there have been notable very recent falls in share prices of AI infrastructure companies - Coreweave, a start-up that supplies OpenAI, saw its shares lose 26% of their value earlier this month.
Plus, there have been some reactions in markets for perceived credit risk among other firms. And while most of these tech share prices have generally climbed higher over the course of 2025, there has been a mild dip more generally in the past few days.
ChatGPT versus Gemini 3.0
None of this has dampened the excitement over AI's potential within the industry. Google's consumer AI model, Gemini 3.0, launched to great fanfare earlier this week — this will pitch Google in a direct battle with OpenAI and its still-dominant ChatGPT for the market share.
What we don't yet know is whether it marks an end to the days of chatbots going rogue and recommending glue as a pizza ingredient. So, is the end result of all this fantastic investment is that information is less reliable, I asked Mr Pichai.
"I think if you only construct systems standalone and you only rely on that, [that] would be true," he told me. "Which is why I think we have to make the information ecosystem has to be much richer than just having AI technology being the sole product in it."
But I put it to him that truth matters. His response: "truth matters".
You must be logged in to post a comment.