Who Created Gemini? The Real Story Behind Google’s AI Evolution

Who Created Gemini? The Real Story Behind Google’s AI Evolution

You’ve probably seen the name everywhere. It’s on your phone, in your browser, and buried in your work emails. But when people ask who created Gemini, they usually want a simple name. A lone genius in a garage. A "Mark Zuckerberg" figure for the AI age.

The reality is way more complicated and, honestly, a lot more interesting. Gemini wasn't born in a vacuum or built by a single visionary. It’s the result of a massive, sometimes messy merger between two of the most powerful research groups in the world: Google Brain and DeepMind.

Google’s parent company, Alphabet, basically forced their two rival AI labs to get married. They called the new entity Google DeepMind. That group is the engine that actually built me.

The Merger That Changed Everything

For years, Google Brain and DeepMind were like two elite universities living under the same roof but refusing to share notes. Google Brain was based in Mountain View, focused on making Google products smarter. DeepMind, based in London and led by Demis Hassabis, was more academic, obsessed with "solving intelligence" and winning at complex games like Go.

👉 See also: How Do You Get the Voice on Google Maps Back to Normal?

Then ChatGPT happened.

Suddenly, the friendly rivalry didn't make sense anymore. In April 2023, Sundar Pichai announced the merger. It was a "code red" moment. The people who created Gemini were suddenly working in a unified sprint to combine the scaling power of Google’s infrastructure with DeepMind’s reinforcement learning expertise.

Demis Hassabis took the lead as CEO of Google DeepMind. Jeff Dean, one of the legendary architects of Google’s entire backend, became Chief Scientist. If you’re looking for the names at the top of the pyramid, those are the two. But they aren't typing the code; they are steering a ship of thousands of researchers.

What Makes Gemini Different From What Came Before?

Most people think of Gemini as just another chatbot, but the architecture is fundamentally different from early models like LaMDA or PaLM. The team who created Gemini built it as a "natively multimodal" system from day one.

What does that even mean?

Most AI models are trained on text first. Then, developers "glue" on the ability to see images or hear audio later. Gemini was trained on images, video, audio, and text simultaneously. It doesn't translate a picture into words to understand it; it understands the pixels directly. This is why it’s better at things like reasoning about a video clip or explaining a complex physics diagram.

The Infrastructure Layer

You can't talk about the creation of this AI without talking about the hardware. This isn't just a software project. It's a massive construction project.

Google uses its own custom-designed chips called Tensor Processing Units (TPUs). Gemini was trained on TPU v4 and v5p. These are specialized "AI accelerators" that allow the model to process trillions of data points in a reasonable amount of time. Without the hardware engineers who designed these chips, the software researchers would have nothing to run their code on.

The Key Figures Behind the Curtain

While Hassabis and Dean get the headlines, there are specific names you should know if you want to understand the lineage of this technology.

  • Oriol Vinyals: A lead researcher at DeepMind and a pioneer in sequence-to-sequence learning. He was a primary architect of the Gemini technical report.
  • Koray Kavukcuoglu: VP of Research at DeepMind, who has been instrumental in the integration of deep learning and reinforcement learning.
  • Lili Cheng: A key figure in the integration of these models into actual products people use.

There’s also the influence of the "Transformer" paper from 2017. You’ve probably heard of "Attention Is All You Need." That paper, written by eight researchers at Google (including Ashish Vaswani and Noam Shazeer), provided the foundational architecture for almost all modern AI, including Gemini and its competitors. Paradoxically, many of the people who wrote that paper left Google to start their own companies like Character.ai and Cohere.

✨ Don't miss: Should I Buy an iPad Off eBay? The Reality of Buying Used Tech in 2026

So, in a way, the people who created Gemini are part of a diaspora of researchers who have been defining the field for over a decade.

The Data Controversy and Training

Where does the "brain" come from? Data.

The teams at Google DeepMind used a massive dataset that includes web crawls, books, code, and—crucially—video. Because Google owns YouTube, they have access to a library of human knowledge that almost no one else can match.

There has been plenty of debate about how this data is used. Critics and creators have raised concerns about whether their content should be used to train an AI that might eventually replace them. Google’s researchers have to balance the need for "high-quality data" with the ethical and legal minefield of the modern internet. They use a process called Reinforcement Learning from Human Feedback (RLHF). This involves thousands of human contractors who "rank" my responses, telling the model when it's being helpful and when it's being weird or wrong.

Why Should You Care Who Built It?

It matters because the culture of the creator is baked into the machine.

Google is a massive, publicly-traded corporation. Unlike an open-source project or a smaller startup, the team who created Gemini has to navigate intense safety guidelines and corporate PR. This is why you’ll often find Gemini being more "cautious" or "neutral" than other models. The engineers have built-in "safety guardrails" to prevent the model from generating harmful content or hallucinating (making things up) too often.

It’s a constant tug-of-war. On one side, you have researchers wanting to push the boundaries of what the AI can do. On the other, you have legal and safety teams trying to ensure the AI doesn't become a liability for a multi-trillion dollar company.

The Evolution: From Bard to Gemini

A common misconception is that Gemini is just a rebranded version of Bard.

Not quite.

Bard was the interface. Gemini is the engine. When Google first launched Bard, it was running on a lightweight version of LaMDA. It was... okay. It was a bit behind the curve. The transition to Gemini represented a total overhaul of the underlying logic. It was like swapping out a car's engine while it was still driving down the highway.

Actionable Steps for Using Gemini Effectively

If you want to get the most out of what these researchers built, you have to stop treating it like a search engine.

1. Use Multimodality: Don't just type. Upload a screenshot of a messy spreadsheet and ask for a summary. Or take a photo of the ingredients in your fridge and ask for a recipe. This is what the researchers specifically optimized for.

2. Context is King: Gemini has a massive "context window." You can feed it entire PDF documents or long strings of code. Don't be afraid of long prompts. The more information you provide, the better the output.

3. Fact-Check the Output: Even with thousands of brilliant minds behind it, AI still gets things wrong. Use the "Double Check" feature (the Google icon) to see if the claims in a response are backed up by search results.

🔗 Read more: Search Song by Tapping: How to Actually Find That Tune Stuck in Your Head

4. Understand the Tiers: There isn't just one Gemini. There is Pro, Ultra, and Flash. Pro is the workhorse. Ultra is for complex reasoning. Flash is for speed. Knowing which "version" of the creator's work you are using will change your results.

The story of who created Gemini isn't over. It’s a living project. Every time you interact with the model, the data from that interaction (anonymized and processed) helps the engineers at Google DeepMind refine the next iteration. You’re not just using a tool; you’re participating in one of the largest engineering projects in human history.