Categories
Asides

Why Didn’t Google Build ChatGPT?

When OpenAI released ChatGPT, I believe the company that was most shocked was Google.

They had DeepMind. They had Demis Hassabis. By all accounts, Google had some of the best researchers in the world. So why couldn’t they build ChatGPT—or even release it?

Google also had more data than anyone else.
So why did that not help? Perhaps it was because they had too much big data—so much of it optimized for search and advertising that it became a liability in the new paradigm of language generation. Data that had once been a strategic asset was now too noisy, too structurally biased to be ideal for training modern AI.

Having a large amount of data is no longer the condition for innovation. Instead, what matters now is a small amount of critical data, and a team with a clear objective for the model’s output. That’s what makes today’s AI work.

That’s exactly what OpenAI demonstrated. In its early days, they didn’t have access to massive GPU clusters. Their partnership with Microsoft only came later, after GPT-3. They launched something that moved the world—with minimal resources, and a lot of design and training ingenuity. It wasn’t about quantity of data, but quality. Not about how much compute you had, but how you structured your model. That was the disruptive innovation.

And what did Big Tech do in response? They began buying up GPUs. To preempt competition. They secured more computing power than they could even use, just to prevent others from accessing it.

It was a logical move to block future disruptions before they could even begin. In language generation AI especially, platforms like Twitter and Facebook—where raw, unfiltered human expression is abundant—hold the most valuable data. These are spaces full of emotion, contradiction, and cultural nuance. Unlike LinkedIn, which reflects structured, formalized communication, these platforms capture what it means to be human.

That’s why the data war began. Twitter’s privatization wasn’t just a media shakeup. Although never explicitly stated, Twitter’s non-public data has reportedly been used in xAI’s LLM training. The acquisition likely aimed to keep that “emotional big data” away from competitors. Cutting off the API and changing domains was a visible consequence of that decision.

And just as Silicon Valley was closing in—hoarding data and GPUs—DeepSeek emerged from an entirely unexpected place.

A player from China, operating under constraints, choosing architectures that didn’t rely on cutting-edge chips, yet still managing to compete in performance. That was disruptive innovation in its purest form.

What Google had, OpenAI didn’t. What OpenAI had, Google didn’t. That difference now seems to signal the future shape of our digital world.

Leave a ReplyCancel reply

This site uses Akismet to reduce spam. Learn how your comment data is processed.

Discover more from iAm Hiro Shinohara

Subscribe now to keep reading and get access to the full archive.

Continue reading

Exit mobile version