Join us at Radio City Music Hall, July 15th! >>

Google: The AI Company

Fall 2025, Episode 1

ACQ2 Episode

October 5, 2025
October 5, 2025

The Complete History & Strategy of Google

Google faces the greatest innovator's dilemma in history. They invented the Transformer — the breakthrough technology powering every modern AI system from ChatGPT to Claude (and, of course, Gemini). They employed nearly all the top AI talent: Ilya Sutskever, Geoff Hinton, Demis Hassabis, Dario Amodei — more or less everyone who leads modern AI worked at Google circa 2014. They built the best dedicated AI infrastructure (TPUs!) and deployed AI at massive scale years before anyone else. And yet... the launch of ChatGPT in November 2022 caught them completely flat-footed. How on earth did the greatest business in history wind up playing catch-up to a nonprofit-turned-startup?

Today we tell the complete story of Google's 20+ year AI journey: from their first tiny language model in 2001 through the creation Google Brain, the birth of the transformer, the talent exodus to OpenAI (sparked by Elon Musk's fury over Google’s DeepMind acquisition), and their current all-hands-on-deck response with Gemini. And oh yeah — a little business called Waymo that went from crazy moonshot idea to doing more rides than Lyft in San Francisco, potentially building another Google-sized business within Google. This is the story of how the world's greatest business faces its greatest test: can they disrupt themselves without losing their $140B annual profit-generating machine in Search?

Sponsors:

Many thanks to our fantastic Fall ‘25 Season partners:

Acquired’s 10th Anniversary Celebration!

Links:

Carve Outs:

Sponsors:

More Acquired: 

Join the Slack
Get Email Updates
Become a Limited PartnerJoin the Slack

Get New Episodes:

Thank you! You're now subscribed to our email list, and will get new episodes when they drop.

Oops! Something went wrong while submitting the form

Transcript: (disclaimer: may contain unintentionally confusing, inaccurate and/or amusing transcription errors)

Ben: I went and looked at a studio, well a little office that I was going to turn into a studio nearby, but it was not good at all. It had drop ceilings so I could hear the guy in the office next to me, you would be able to hear him talking on episodes.

David: Third co-host.

Ben: Third co-host.

David: Is it Howard?

Ben: No, it was a lawyer. It seemed to be talking through some horrible problem that I didn’t want to listen to, but I could hear every word.

David: Does he want millions of people listening to his conversation?

Ben: All right, let’s do a podcast.

David: Let’s do a podcast.

Ben: Welcome to the fall 2025 season of Acquired, the podcast about great companies and the stories and playbooks behind them. I’m Ben Gilbert.

David: I’m David Rosenthal.

Ben: And we are your hosts. Here’s a dilemma. Imagine you have a profitable business. You make giant margins on every single unit you sell. And the market you compete in is also giant, one of the largest in the world, you might say. But then on top of that, lucky for you, you also are a monopoly in that giant market with 90% share and a lot of lock-in.

David: And when you say monopoly, monopoly as defined by the US government.

Ben: That is correct. But then imagine this. In your research lab, your brilliant scientists come up with an invention. This particular invention, when combined with a whole bunch of your old inventions by all your other brilliant scientists, turns out to create the product that is much better for most purposes than your current product.

You launch the new product based on this new invention, especially because out of pure benevolence, your scientists had published research papers about how awesome the new invention is and lots of the inventions before also. Now there are new startup competitors quickly commercializing that invention. Of course David, you change your whole product to be based on a new thing, right?

David: This sounds like a movie.

Ben: Yes. But here is the problem. You haven’t figured out how to make this new incredible product anywhere near as profitable as your old giant cash printing business. Maybe you shouldn’t launch that new product.

David, this sounds like quite the dilemma to me. Of course, listeners, this is Google today and in perhaps the most classic textbook case of The Innovator’s Dilemma ever. The entire AI revolution that we are in right now is predicated by the invention of the Transformer out of the Google Brain team in 2017.

Think OpenAI and ChatGPT, Anthropic, NVIDIA hitting all time highs. All the craziness right now depends on that one research paper published by Google in 2017.

And consider this. Not only did Google have the densest concentration of AI talent in the world 10 years ago that led to this breakthrough, but today they have just about the best collection of assets that you could possibly ask for.

They’ve got a top tier AI model with Gemini. They don’t rely on some public cloud to host their model. They have their own in Google Cloud that now does $50 billion in revenue. That is real scale. They’re a chip company with their Tensor Processing Units or TPUs, which is the only real scale deployment of AI chips in the world besides NVIDIA GPUs. Maybe AMD, maybe, but these are definitely the top two.

David: Somebody put it to me in research that if you don’t have a foundational frontier model or you don’t have an AI chip, you might just be a commodity in the AI market. And Google is the only company that has both.

Ben: Google still has a crazy bench of talent, and despite ChatGPT becoming the Kleenex of the era, Google does still own the text box, the single one that is the front door to the Internet for the vast majority of people, anytime anyone has intent to do anything online.

But the question remains. What should Google do strategically? Should they risk it all and lean into their birthright to win in artificial intelligence? Or will protecting their gobs of prophets from search hamstring them as the AI wave passes them by? But perhaps first we must answer the question, how did Google get here, David Rosenthal. Listeners today, we tell the story of Google the AI company.

David: Woo.

Ben: You like that David? Was that good?

David: I love. Did you hire a Hollywood scriptwriting consultant without telling me?

Ben: I wrote that 100% myself with no AI, thank you very much.

David: No AI.

Ben: Well listeners, if you want to know every time an episode drops, vote on future episode topics, or get access to corrections from past episodes, check out our email list. That’s acquired.fm/email.

Come talk about this episode with the entire Acquired community in Slack after you listen. That’s acquired.fm/slack.

David: Speaking of the Acquired community, we have an anniversary celebration coming up.

Ben: We do.

David: Ten years of the show. We’re going to do an open Zoom call with everyone to celebrate, like how we used to do our LP calls back in the day with LPs. We are going to do that on October 20th, 2025 at 4:00 PM Pacific time. Check out the show notes for more details.

Ben: If you want more Acquired, check out our interview show, ACQ2. Our last interview was super fun. We sat down with Tobi Lütke, the founder and CEO of Shopify about how AI has changed his life and where he thinks it will go from here. So search ACQ2 and any podcast player. Before we dive in, we want to briefly thank our presenting partner, J.P. Morgan Payments.

David: Just like how we say every company has a story, every company’s story is powered by payments, and J.P. Morgan payments is a part of so many of their journeys from seed to IPO and beyond.

Ben: So with that, this show is not investment advice. David and I may have investments in the companies we discuss, and this show is for informational and entertainment purposes only. David, Google, the AI company.

David: So Ben, as you were alluding to in that fantastic intro—really, you’re up in the game again—if we rewind 10 years ago from today before the Transformer paper comes out, all of the following people, as we’ve talked about before, were Google employees. Ilya Sutskever, founding chief scientist of OpenAI, who along with Geoff Hinton and Alex Krizhevsky had done the seminal AI work on AlexNet and just published that a few years before.

All three of them were Google employees, as was Dario Amodei, the founder of Anthropic; Andrej Karpathy, chief scientist at Tesla until recently; Andrew Ng, Sebastian Thrun, Noam Shazeer; all the DeepMind folks, Demis Hassabis, Shane Legg, Mustafa Suleyman—Mustafa, now in addition to in the past having been a founder of DeepMind, runs AI at Microsoft—basically every single person of note in AI worked at Google with the one exception of Yann Le Cun who worked at Facebook.

Ben: It’s pretty difficult to trace a big AI lab now back and not find Google in its origin story.

David: The analogy here is it’s almost as if at the dawn of the computer era itself, a single company like (say) IBM had hired every single person who knows how to code. It’d be like if anybody else wants to write a computer program, oh sorry. You can’t do that. Anybody who knows how to program works at IBM.

This is how it was with AI and Google in the mid 2010s. But learning how to program a computer wasn’t so hard that people out there couldn’t learn how to do it. Learning how to be an AI researcher, significantly more difficult.

Ben: It was the stuff of very specific PhD programs with a very limited set of advisors, and a lot of infighting in the field of where the direction of the field was going, what was legitimate versus what was crazy heretical religious stuff.

David: Then, yes, the question is how do we get to this point? Well, it goes back to the start of the company. Larry Page always thought of Google as an artificial intelligence company. In fact, Larry Page’s dad was a computer science professor and had done his PhD at the University of Michigan in machine learning and artificial intelligence, which was not a popular field in computer science back then.

Ben: In fact, a lot of people thought specializing in AI was a waste of time because so many of the big theories from 30 years prior to that had been disproven at that point, or at least people thought they were disproven. It was frankly contrarian for Larry’s dad to spend his life and career and research work in AI.

David: And that rubbed off on Larry. If you squint the PageRank algorithm that Google was founded upon, is a statistical method. You could classify it as part of AI within computer science. Larry, of course, was always dreaming much, much bigger.

There’s the quote that we’ve said before on this show, in the year 2000, two years after Google’s founding, when Larry says, “Artificial intelligence would be the ultimate version of Google. If we had the ultimate search engine, it would understand everything on the web, it would understand exactly what you wanted, and it would give you the right thing.

That’s obviously artificial intelligence. We’re nowhere near doing that now. However, we can get incrementally closer, and that is basically what we work on here.” It’s always been an AI company.

Ben: And that was in 2000.

David: Well, one day, in either late 2000 or early 2001—the timelines are a bit hazy here—Google engineer named George Herrick is talking over lunch with Ben Gomes, famous Google engineer who I think would go on to lead search, and a relatively new engineering hire named Noam Shazeer.

Now, George was one of Google’s first 10 employees, incredible engineer. Just like Larry Page’s dad, he had a PhD in machine learning from the University of Michigan. Even when George went there, it was still a relatively rare contrarian subfield within computer science.

The three of them are having lunch, and George says offhandedly to the group that he has a theory from his time as a PhD student, that compressing data is actually technically equivalent to understanding it. The thought process is if you can take a given piece of information and make it smaller, store it away, and then later reinstantiate it in its original form, the only way that you could possibly do that is if whatever force is acting on the data actually understands what it means.

You’re losing information going down to something smaller, and then recreating the original thing. It’s like you’re a kid in school. You learn something in school, you read a long textbook, you store the information in your memory, then you take a test to see if you really understood the material. And if you can recreate the concepts, then you really understand it.

Ben: Which foreshadows big LLMs today are like compressing the entire world’s knowledge into some number of terabytes, that’s just like the smashed down little vector set—little at least compared to all the information in the world—but it’s that idea. You can store all the world’s information in an AI model in something that is incomprehensible and hard to understand. But then if you uncompress it, you can bring knowledge back to its original form.

David: And these models demonstrate understanding, right?

Ben: Eh, do they? That’s the question. They certainly mimic understanding.

David: So this conversation is happening—this is 25 years ago—and Noam the new hire, the young buck, stops in his tracks and he’s like, wow. If that’s true, that’s really profound.

Ben: Is this in one of Google’s micro kitchens?

David: This is in one of Google’s micro kitchens. They’re having lunch.

Ben: Where did you find this, by the way? A 25-year-old…

David: This is In the Plex. This is a small little passage in Stephen Levy’s great book that’s been a source for all of our Google episodes, In the Plex. This is a small little throwaway passage in here about this because this book came out before ChatGPT and AI and all that.

So Noam latches onto George, and keeps vibing over this idea. Over the next couple of months, the two of them decide, in the most Google-y fashion possible, that they are just going to stop working on everything else, and they’re going to go work on this idea on language models and compressing data and can they generate machine understanding with data. And if they can do that, that that would be good for Google.

I think this coincides with that period in 2001 when Larry Page fired all the managers in the engineering organization. Everybody was just doing whatever they wanted to do.

Ben: Funny.

David: There’s this great quote from George in the book. “A large number of people thought it was a really bad thing for Noam and I to spend our talents on, but Sanjay Ghemawat,” Sanjay, of course, being Jeff Dean’s famous prolific coding partner, “thought it was cool.”

George would posit the following argument to any doubters that they came across. “Sanjay thinks it’s a good idea, and no one in the world is as smart as Sanjay, so why should Noam and I accept your view that it’s a bad idea?”

Ben: It’s like if you beat the best team in football, are you the new best team in football no matter what?

David: Yeah. All of this ends up taking Noam and George deep down the rabbit hole of probabilistic models for natural language. Meaning, for any given sequence of words that appears on the Internet, what is the probability for another specific sequence of words to follow? This should sound pretty familiar for anybody who knows about LLMs work today.

Ben: Oh, like a next word predictor.

David: Or a next token predictor if you generalized it. The first thing that they do with this work is they create the ‘did you mean’ spelling correction in Google search.

Ben: Oh, that came out of this?

David: That came out of this. Noam created this. This is huge for Google because obviously it’s a bad user experience when you mistype a query and then need to type another one, but it’s a tax to Google’s infrastructure because every time these mistyped queries are going, Google’s infrastructure goes and serves the results to that query that are useless and immediately overwritten with the new one.

Ben: And it’s a really tightly scoped problem where you can see like, oh wow, 80% of the time that someone types in ‘god groomer,’ oh they actually mean dog groomer and they retype it. If it’s really high confidence, then you actually just correct it without even asking them, and then ask them if they want to opt out instead of opting in. It’s a great feature and it’s a great first use case for this in a very narrowly scoped domain.

David: Totally. So they get this way and they keep working on it, Noam and George, and they end up creating a fairly “large” (for the time) language model that they call affectionately PHIL—the Probabilistic Hierarchical Inferential Learner.

Ben: These AI researchers love creating their backronyms.

David: They love their word puns. Fast forward to 2003, and Susan Wojcicki and Jeff Dean are getting ready to launch AdSense. They need a way to understand the content of these third-party webpages, the publishers, in order to run the Google Ad corpus against them. Well PHIL is the tool that they use to do it.

Ben: I had no idea that language models were involved in this.

David: Jeff Dean borrows PHIL and famously uses it to code up his implementation of AdSense in a week (because he’s Jeff Dean) and boom, AdSense. This is billions of dollars of new revenue to Google overnight because it’s the same corpus of better AdWords that are search ads, that they’re now serving on third-party pages. They just massively expanded the inventory for the ads that they already have in the system. Thanks to PHIL.

Ben: Thanks to PHIL.

All right. This is a moment where we got to stop and just give some Jeff Dean facts. Jeff Dean is going to be the through line of this episode of wait, how did Google pull that off? How did Jeff Dean just go home and over the weekend rewrite some entire giant distributed system and figure out all of Google’s problems?

Back when Chuck Norris facts were big, Jeff Dean facts became a thing internally at Google. I just want to give you some of my favorites. The speed of light in a vacuum used to be about 35 miles per hour. Then Jeff Dean spent a weekend optimizing physics.

David: So good.

Ben: Jeff Dean’s PIN is the last four digits of pie.

David: Only Googlers would come up with these.

Ben: To Jeff Dean, NP means no problemo.

David: Oh yeah. I’ve seen that one before. I think that one’s my favorite. Oh man. So, so good. Also a wonderful human being who we spoke to in research and was very, very helpful. Thank you, Jeff.

So, language models definitely work, definitely going to drive a lot of value for Google, and they also fit pretty beautifully into Google’s mission to organize the world’s information, make it universally accessible and useful. If you can understand the world’s information, compress it, and then recreate it. that fits the mission I think. I think that checks the box.

Ben: Absolutely.

David: PHIL gets so big that apparently by the mid-2000s, PHIL is using 15% of Google’s entire data center infrastructure. I assume a lot of that is AdSense ad serving, but also ‘did you mean’ and all the other stuff that they start using it for within Google.

Ben: So early natural language systems, computationally expensive.

David: Yes. So now mid-2000s, fast forward to 2007, which is a very, very big year for the purposes of our story. Google had just recently launched the Google Translate product. This is the era of all the great, great products coming out of Google that we’ve talked about—Maps, Gmail, Docs and all the wonderful things that Chrome and Android are going to come later.

Ben: They had a 10-year run where they basically launched everything you know of at Google except for search, Truly in a 10 year run. Then there were about 10 years after that from 2013-on where they basically didn’t launch any new products that you’ve heard about until we get to Gemini, which is this fascinating thing. But this 2003–2013 era was just so rich with hit after hit after hit.

David: Magical. One of those products was Google Translate, not the same level of user base or perhaps impact on the world as Gmail or Maps or whatnot, but still a magical, magical product. The chief architect for Google Translate was another incredible machine learning PhD named Franz Och.

Franz had a background in natural language processing and machine learning, and that was his PhD. He was German, he got his PhD in Germany. At the time, DARPA…

Ben: The Defense Advanced Research Projects Agency division of the government.

David: …had one of their famous challenges going for machine translation. Google and Franz of course enters this, and Franz builds an even larger language model that blows away the competition in this year’s version of the DARPA challenge. This is either 2006 or 2007, gets an astronomically high blue score. For the time it’s called the bilingual evaluation understudy is the algorithmic benchmark for judging the quality of translations. At the time, higher than anything else possible.

Jeff Dean hears about this and the work that Franz and the Translate team have done. It’s like, this is great. This is amazing. When are you guys going to ship this in production?

Ben: Oh, I heard this story.

David: Jeff and Noam talk about this on the DoorDash podcast. That episode is so, so good. Franz is like, no, no, no, no. Jeff, you don’t understand. This is research. This isn’t for the product. We can’t ship this model that we built. This is a n gram language model—gram’s like a number of words in a cluster—and we’ve trained it on a corpus of two trillion words from the Google search index.

This thing is so large it takes it 12 hours to translate a sentence. The way the DARPA challenge worked in this case was you got a set of sentences on Monday and then you had to submit your machine translation of those set of sentences by Friday.

Ben: Plenty of time for the servers to run.

David: They were like, okay, so we have whatever number of hours it is from Monday to Friday. Let’s use as much compute as we can to translate these couple of sentences.

Ben: Hey, learn the rules of the game and use them to your advantage.

David: Exactly. Jeff Dean being the engineering equivalent of Chuck Norris, he’s like, hmm, let me see your code. Jeff goes and parachutes in and works with the Translate team for a few months. He re-architects the algorithm to run on the words and the sentences in parallel instead of sequentially. Because when you’re translating a set of sentences or a set of words in a sentence, you don’t necessarily need to do it in order. You can break up the problem into different pieces, work on it independently. You can parallelize it.

Ben: And you won’t get a perfect translation. But imagine you just translate every single word, you can at least go translate those all at the same time in parallel, reassemble the sentence, and mostly understand what the initial meaning was.

David: Yup. And as Jeff knows very well, because he and Sanjay basically built it with Urs Hölzle, Google’s infrastructure is extremely parallelizable. Distributed, You can break up workloads into little chunks, send them all over the various data centers that Google has, reassemble the projects, return that to the user.

Ben: They are the single best company in the world at parallelizing workloads across CPUs, across multiple data centers.

David: We’re still talking CPUs here. Jeff’s work with the team gets that average sentence translation time down from 12 hours to 100 milliseconds. Then they ship it in Google Translate and it’s amazing.

Ben: This sounds like a Jeff Dean fact. Well, it used to take 12 hours, and then Jeff Dean took a few months with it. Now it’s 100 milliseconds.

David: So this is the first “large” language model used in production in a product at Google. They see how well this works. Maybe we could use this for other things like predicting search queries as you type. That might be interesting.

Of course, the crown jewel of Google’s business that also might be an interesting application for this, the ad quality score for AdWords is literally the predicted click-through rate on a given set of ad copy. You can see how an LLM that is really good at ingesting information, understanding it, and predicting things based on that might be really useful for calculating ad quality for Google.

Ben: Which is the direct translation to Google’s bottom line.

David: Indeed. Obviously, all of that is great on the language model front. I said 2007 was a big year. Also in 2007 begins the momentous intersection of several computer science professors on the Google campus.

In April of 2007, Larry Page hires Sebastian Thrun from Stanford to come to Google and work first part-time and then full-time on machine learning applications. Sebastian was the head of SAIL (the Stanford Artificial Intelligence Laboratory), legendary AI laboratory that was big in the first wave of AI back in the 60s–70s when Larry’s dad was active in the field. Then actually shut down for a while and then had been restarted and re-energized here in the early 2000s. Sebastian was the leader, the head of SAIL.

Ben: Funny story about Sebastian, the way that he actually comes to Google. Sebastian was kind enough to speak with us to prep for this episode. I didn’t realize it was basically an acquihire. He and some (I think it was) grad students were in the process of starting a company, had term sheets from Benchmark and Sequoia. Larry came over and said, what if we just acquire your company before it’s even started in the form of signing bonuses?

David: Probably a very good decision on their part. SAIL, this group within the CS Department at Stanford not only had some of the most incredible, most accomplished professors and PhD AI researchers in the world. They also had this stream of Stanford undergrads that would come through and work there as researchers while they were working on their CS degrees or symbolic system degrees or whatever it was that they were doing at Stanford undergrads.

One of those people was Chris Cox, who’s the Chief Product Officer at Meta.

Chris: No way.

David: Yeah, that was how he got his start in all of this in AI. Obviously, Facebook and Meta are going to come back into the story here in a little bit. You really can’t make this up.

Another undergrad who passed through SAIL while Sebastian was there was a young freshman and sophomore who would later drop out of Stanford to start a company that went through Y Combinator’s very first batch in summer 2005.

Ben: I’m on the edge of my seat. Who is this?

David: Any guesses?

Ben: Dropbox, Reddit. I’m trying to think who else was in the first batch.

David: Oh, no, but way more on the nose for this episode. The company was a failed local mobile social network.

Ben: Oh, Sam Altman. Looped.

David: Sam Altman.

Ben: That’s amazing. He was at SAIL at the same time?

David: He was at SAIL, yup, as an undergrad researcher.

Ben: Wow.

David: Wild, right? We told you that It’s a very small set of people that are all doing all of this.

Ben: Man, I miss those days. Sam presenting at the WWDC with Steve Jobs on stage, with the double popped collar. Different time in tech.

David: The double popped collar. That was amazing. That was a vibe. That was a moment. Oh man. All right, so April 2007 Sebastian comes over from SAIL into Google. One of the first things he does over the next set of months is a project called Ground Truth for Google Maps.

Ben: Which is essentially Google Maps.

David: It is essentially Google Maps. Before Ground Truth, Google Maps existed as a product, but they had to get all the mapping data from a company called Tele Atlas.

Ben: I think there were two. They were a duopoly. Navtech was the other one.

David: Yeah, Navtech and Tele Atlas.

Ben: But it was this crappy source of truth map data that everyone used, and you really couldn’t do any better than anyone else because you all just used the same data.

David: It was not that good and it cost a lot of money. Tele Atlas and Navtech were multi-billion dollar companies. I think maybe one or both of them were public at some point, then got acquired. But a lot of money, a lot of revenue.

Ben: Yup. Sebastian’s first thing was Street View. He already had the experience of orchestrating this fleet of all these cars to drive around and take pictures.

David: Then coming into Google, Ground Truth is this moonshot-type project to recreate all the Tele Atlas data.

Ben: Mostly from their own photographs of streets, from Street View. And they incorporated some other data. There was census data they used, and I think it was 40-something data sources to bring it all together. But ground Truth was this very ambitious effort to create new maps from whole cloth.

David: Just like all of the AI and AI-enabled projects within Google that we’re talking about here, works very, very well, very quickly. Huge win.

Ben: Well, especially when you hire a thousand people in India to help you sift through all the discrepancies and the data and actually hand draw all the maps.

David: We are not yet in an era of a whole lot of AI automation. On the back of this win with Ground Truth, Sebastian starts lobbying to Larry and Sergey. Hey, we should do this a lot. We should bring in AI professors, academics. I know all these people, into Google part-time.

They don’t have to be full-time employees. Let them keep their posts in academia, but come here and work with us on projects for our products. They’ll love it. They get to see their work used by millions and millions of people. We’ll pay them, they’ll make a lot of money, they’ll get Google stock, and they get to stay professors at their academic institutions.

Ben: Win, win, win.

David: Win, win, win. As you would expect, Larry and Serge are like, yeah, yeah, yeah, that’s a good idea. Let’s do that. More of that. So in December of 2007, Sebastian brings in a relatively little known machine learning professor from the University of Toronto named Geoff Hinton to the Google campus to come and give a tech talk. Not yet hiring him, but come give a tech talk to all the folks at Google, and talk about some of the new work, Geoff, that you and your PhD and postdoc students there at the University of Toronto are doing on blazing new paths with neural networks.

Ben: And Geoff Hinton, for anybody who doesn’t know the name, now very much known as the godfather of neural networks and really the godfather of the whole direction that AI went in.

David: Modern AI.

Ben: He was a fringe academic at this point in history. Neural networks were not a respected subtree of AI.

David: No, totally not.

Ben: And part of the reason is there had been a lot of hype 30–40 years before around neural networks that just didn’t pan out. It was effectively (everyone thought) disproven and certainly backwater.

David: Ben, do you remember from our NVIDIA episodes my favorite piece of trivia about Geoff Hinton?

Ben: Oh yes, that his great-grandfather was George Boole?

David: Yup. He is the great-great-grandson of George and Mary Boole who invented Boolean algebra and Boolean logic.

Ben: Which is hilarious now that I know more about this because that’s the basic building block of symbolic logic of defined deterministic computer science logic. The hilarious thing about neural nets is it’s not. It’s not symbolic AI, it’s not I feed you the specific instructions and you follow a big if-then tree. It is non-deterministic. It is the opposite of that field.

David: Which actually just underscores (again) how heretical this branch of machine learning and computer science was. Ben, as you were saying earlier, neural networks not a new idea and had all of this great promise in theory, but in practice just took too much computation to do multiple layers.

You could really only have a single or maybe small single-digit number of layers in a computer neural network up until this time. But Geoff and his former postdoc, guy named Yann LeCun, start evangelizing within the community. Hey, if we can find a way to have multi-layered, deep layered neural networks, something we call deep learning, we could actually realize the promise here.

Ben: It’s not that the idea is bad. It’s that the implementation, which would take a ton of compute to actually do all the math, to do all the multiplication required to propagate through layer after layer after layer of neural networks to detect and understand and store patterns, if we could actually do that, a big multi-layered neural network would be very valuable and possibly could work.

David: Here we are now in 2007, mid-2000s, Moore’s law has increased enough that you could actually start to try to test some of these theories. So Geoff comes and he gives this talk at Google. It’s on YouTube, you can go watch it, We’ll link to it in the show notes.

Ben: Really?

David: This is incredible. This is an artifact of history sitting there on YouTube, and people at Google, Sebastian, Jeff Dean, all the other folks we are talking about, get very, very, very excited because they’ve already been doing stuff like this with Translate and the language models that they’re working with, that’s not using deep neural networks that Geoff’s working on.

Here’s this whole new architectural approach, that if they could get it to work, would enable these models that they’re building to work way better, recognize more sophisticated patterns, understand the data better. Very, very promising.

Ben: Again, all in theory at this point.

David: Sebastian Thrun brings Geoff Hinton into the Google fold after this tech talk, (I think) first as a consultant over the next couple of years, then this is a basic later, Geoff Hinton technically becomes an intern at Google. That’s how they get around the part-time, full-time policies here.

Ben: He was a summer intern somewhere around 2011–2012. Mind you, at this point he’s 60 years old.

David: In the next couple of years after 2007 here, Sebastian’s concept of bringing these computer science, machine learning academics into Google as contractors or part-time interns—basically letting them keep their academic posts and work on big projects for Google’s products internally—goes so well that by late 2009, Sebastian, Larry, and Sergey decided, hey, we should just start a whole new division within Google. And it becomes Google X, the Moonshot Factory.

The first project within Google X, Sebastian leads himself…

Ben: Oh, David, don’t say it, don’t say it.

David: I won’t say the name of it. We will come back to it later. But for our purposes for now, the second project would be critically important not only for our story, but—

Ben: To the whole world.

David: Everything in AI changing the entire world, and that second project is called Google Brain. But before we tell the Google Brain story, now is a great time to thank our friends at J.P. Morgan payments.

Ben: Today we are going to talk about one of the core components of J.P. Morgan payments, their treasury solutions. Now, treasury is something that most listeners probably do not spend a lot of time thinking about, but it’s fundamental to every company.

David: Treasury used to be just a back office function, but now great companies are using it as a strategic lever. With J.P. Morgan Payments treasury solutions, you can view and manage all your cash positions in real time and all of your financial activities across 120 currencies in 200 countries.

Ben: And the other thing that they acknowledge really in their whole strategy is that every business has its own quirk, so it’s not a cookie cutter approach. They work with you to figure out what matters most for you and your business and then help you gain clarity, control, and confidence.

David: So whether you need advanced automation or just want to cut down on manual processes and approvals, their real-time treasury solutions are designed to keep things running smoothly. Whether your treasury is in the millions or billions, or perhaps like the company we’re talking about this episode in the hundreds of billions of dollars.

Ben: And they have some great strategic offerings like pay by bank, which lets customers pay you directly from their bank account. It’s simple, secure, tokenized, and you get faster access to funds and enhance data to optimize revenue and reduce fees. This lets you send and receive real-time payments instantly just with a single API connection to J.P. Morgan.

David: And because J.P. Morgan’s platform is global, that one integration lets you access 45 countries and counting and lets you scale basically infinitely as you expand. As we’ve said before, J.P. Morgan Payments moves $10 trillion a day. So scale is not an issue for your business.

Ben: Not at all. If you’re wondering how to actually manage all that global cash, J.P. Morgan again has you covered with their liquidity and account solutions that make sure you have the right amount of cash and the right currencies in the right places for what you need.

David: So whether you’re expanding into new markets or just want more control over your funds, J.P. Morgan Payments is the partner you want to optimize liquidity, streamline operations, and transform your treasury. To learn more about how J.P. Morgan can help you and your company, just go to jpmorgan.com/acquired and tell them that Ben and David sent you.

Ben: All right, David. Google Brain.

David: When Sebastian left Stanford full-time and joined Google full-time, of course somebody else had to take over SAIL. The person who did is another computer science professor, brilliant guy named Andrew Ng.

Ben: This is all the hits.

David: All the hits. This is all the AI hits on this episode. What does Sebastian do? He recruits Andrew to come part-time. Start spending a day a week on the Google campus. This coincides right with the start of X and Sebastian formalizing this division.

One day in the 2010-201 timeframe, Andrew’s spending his day a week on the Google campus and he bumps into who else? Jeff Dean. Jeff Dean is telling Andrew about what he and Franz have done with language models and what Geoff Hinton is doing in deep learning.

Of course, Andrew knows all this and Andrew’s talking about what he and SAIL are doing at Stanford, and they decide, you know? The time might finally be right to try and take a real big swing on this within Google and build a massive, really large, deep learning model in the vein of what Geoff Hinton has been talking about on highly paralyzable Google infrastructure.

Ben: And when you say the time might be right, Google had tried twice before and neither project really worked. They tried this thing called Brains on Borg. Borg is an internal system that they used to run all of their infrastructure. They tried the Cortex project, and neither of these really worked. So there’s a little bit of scar tissue in the research group at Google of, are large-scale neural networks actually going to work for us on Google infrastructure?

David: So the two of them, Andrew Ng and Jeff Dean, pull in Greg Corrado, who is a neuroscience PhD and amazing researcher who was already working at Google. And in 2011, the three of them launch the second official project within X, appropriately enough called Google Brain. The three of them get to work building a really, really big deep neural network model.

Ben: And if they’re going to do this, they need a system to run it on. Google is all about taking this frontier research and then doing the architectural and engineering system to make it actually run.

David: So Jeff Dean is working on this system on the infrastructure, and he decides to name the infrastructure Distbelief, which of course is a pun, both on the distributed nature of the system and also on (of course) the word disbelief because…

Ben: No one thought it was going to work.

David: …most people in the field thought this was not going to work, and most people in Google thought this was not going to work.

Ben: And here’s a little bit on why. It’s a little technical, but follow me for a second. All the research from that period of time pointed to the idea that you needed to be synchronous. So all the compute needed to be really dense, happening on a single machine, with really high parallelism, like what GPUs do, that you really would want it all happening in one place. It’s really easy to go look up and see, hey, what are the computed values for everything else in the system before I take my next move?

What Jeff Dean wrote with Distbelief was the opposite. It was distributed across a whole bunch of CPU cores and potentially all over a data center or maybe even in different data centers.

In theory, this is really bad because it means you would need to be constantly waiting around on any given machine for the other machines to sync their updated parameters before you could proceed. But instead, the system actually worked asynchronously without bothering to go and get the latest parameters from other cores. You were updating parameters on stale data. You would think that wouldn’t work. The crazy thing is it did.

Okay, so you’ve got Distbelief. What do they do with it now? They want to do some research. So they try out, can we do cool neural network stuff? And what they do in a paper that they submitted in 2011, right at the end of the year, is—and I’ll just give you the name of the paper first—Building High-Level Features Using Large-Scale Unsupervised Learning, but everyone just calls it the cat paper.

David: The cat paper.

Ben: You talk to anyone at Google, you talk to anyone in AI, they’re like, oh yeah, the cat paper.

What they did was they trained a large nine-layer neural network to recognize cats from unlabeled frames of YouTube videos using 16,000 CPU cores on a thousand different machines.

Listeners, just to underscore how seminal this is, we actually talked with Sundar in prep for the episode, and he cited seeing the cat paper come across his desk as one of the key moments that sticks in his brain in Google’s story.

David: A little later on they would do a TGIF where they would present the results of the cat paper. You talk to people at Google, they’re like, that TGIF. Oh my God. That’s what it all changed.

Ben: It proved that large neural networks could actually learn meaningful patterns without supervision and without labeled data. Not only that, it could run on a distributed system that Google built to actually make it work on their infrastructure. That is a huge unlock of the whole thing. Google’s got this big infrastructure asset, can we take this theoretical computer science idea that the researchers have come up with, and use Distbelief to actually run it on our system?

David: That is the amazing technical achievement here that is almost secondary to the business impact of the cat paper. I think it’s not that much of a leap to say that the cat paper led to probably hundreds of billions of dollars of revenue generated by Google, Facebook, and ByteDance over the next decade.

Ben: Definitely. Pattern recognizers in data.

David: So YouTube had a big problem at this time, which was that people would upload these videos and there are tons of videos being uploaded to YouTube, but people are really bad at describing what is in the videos that they’re uploaded.

YouTube is trying to become more of a destination site, trying to get people to watch more videos, trying to build a feed, increase dwell time, et cetera. The problem is the recommender is trying to figure out what to feed, and it’s only just working off titles and descriptions that people were writing about their own videos.

Ben: And whether you’re searching for a video or they’re trying to figure out what video to recommend next, they need to know what the video’s about.

David: Yup. So the cat paper proves that you can use this technology, a deep neural network running on Distbelief to go inside of the videos in the YouTube library and understand what they were about, and use that data to then figure out what videos to serve to people.

Ben: If you can answer the question cat or not a cat, you can answer a whole lot more questions too.

David: Here’s a quote from Jeff Dean about this. “We built a system that enabled us to train pretty large neural nets through both model and data parallelism. We had a system for unsupervised learning on 10 million randomly selected YouTube frames,” as you were saying, Ben. “It would build up unsupervised representations based on trying to reconstruct the frame from the high-level representations. We got that working and training on 2000 computers using 16,000 cores.

After a little while, that model was actually able to build a representation at the highest neural net level where one neuron would get excited by images of cats. It had never been told what a cat was, but it had seen enough examples of them in the training data of head-on facial views of cats that that neuron would then turn on for cats and not much else.”

Ben: It’s so crazy. This is the craziest thing about unlabeled data, unsupervised learning that a system can learn what a cat is without ever being explicitly told what a cat is. And that there’s a cat neuron.

David: Then there’s an iPhone neuron, a San Francisco Giants neuron, and all the things that YouTube recommends.

Ben: Not to mention porn filtering, explicit content filtering.

David: Not to mention copyright identification and enabling revenue share with copyright holders. This leads to everything in YouTube. Basically puts YouTube on the path to today becoming the single biggest property on the Internet and the single biggest media company in the planet.

This kicks off a 10-year period from 2012 when this happens, until ChatGPT on November 30th 2022, when AI is already shaping the human existence for all of us and driving hundreds of billions of dollars of revenue.

It’s just in the YouTube feed, then Facebook borrows it, they hire Yann LeCun and they start Facebook AI research, then they bring it into Instagram, then TikTok and ByteDance take it, then it goes back to Facebook and YouTube with reels and shorts. This is the primary way that humans on the planet spend their leisure time for the next 10 years.

Ben: This is my favorite, David Rosenthal, is everyone talks about 2022-onward as the AI era. I love this point from you that actually for anyone that could make good use of a recommender system and a classifier system basically in a company with a social feed, the AI era started in 2012.

David: Yes, the AI era started in 2012, and part of it was the cat paper. The other part of it was what Jensen at NVIDIA always calls the big bang moment for AI, which was AlexNet.

We talked about Geoff Hinton. Back at the University of Toronto, he’s got two grad students who he’s working with in this era—Alex Krizhevsky and Ilya Sutskever.

Ben: Of course.

David: Future co-founder and chief scientist of OpenAI. The three of them are working with Geoff’s deep neural network ideas and algorithms to create an entry for the famous ImageNet competition in computer science.

Ben: This is Fei-Fei Li’s thing from Stanford.

David: It is an annual machine vision algorithm competition. What it was was Fei-Fei had assembled a database of 14 million images that were hand-labeled. Famously, she used Mechanical Turk on Amazon (I think) to get them all hand-labeled.

Ben: Yes, I think that’s right.

David: The competition was what team can write the algorithm, that without looking at the labels—just seeing the images—could correctly identify the largest percentage. The best algorithms that would win the competitions year over year, were still getting more than a quarter of the images wrong, so 75% success rate. Great. Way worse than a human.

Ben: Can’t use it for much in a production setting when quarter of the time you’re wrong.

David: So then the 2012 competition, along comes AlexNet. Its error rate was 15%. Still high, but a 10% leap from the previous best being a 25% error rate, all the way down to 15% in one year. A leap like that had never happened before.

Ben: It’s 40% better than the next best on a relative basis. And why is it so much better, David? What did they figure out that would create a $4 trillion company in the future?

David: What Geoff, Alex, and Ilya did is they knew—we’ve been talking about all episode—that deep neural networks had all this potential, and Moore’s law advanced enough that you could use CPUs to create a few layers.

They had the aha moment of, what if we re-architected this stuff not to run on CPUs, but to run on a whole different class of computer chips that were by their very nature, highly, highly, highly parallelizable? Video game graphics cards, made by the leading company in the space at the time, NVIDIA. Not obvious at the time, and especially not obvious that this highly advanced, cutting edge, academic computer science research…

Ben: That was being done on supercomputers, usually.

David: …that was being done on super computers with incredible CPUs, would use these toy video game cards.

Ben: That retail for $1000.

David: Less at that point, like a couple of hundred dollars. The team in Toronto go out to the local Best Buy or something. They buy two NVIDIA GeForce GTX 580s, which were NVIDIA’s top of the line gaming cards at the time. The Toronto team rewrites their neural network algorithms in CUDA, NVIDIA’s programming language. They train it on these two off-the-shelf GTX 580s. This is how they achieve their deep neural network and do 40% better than any other entry in the ImageNet competition.

When Jensen says that this was the big bang moment of artificial intelligence, (a) he’s right. This shows everybody that, holy crap, if you can do this with two off-the-shelf GTX 580s, imagine what you could do with more of them or with specialized chips. And (b) this event is what sets NVIDIA on the path from a somewhat struggling PC gaming accessory maker to the leader of the AI wave and the most valuable company in the world today.

Ben: And this is how AI research tends to work, is there’s some breakthrough that gets you this big step change function. Then there’s actually a multi-year process of optimizing from there, where you get these diminishing returns curves on breakthroughs, where the first half of the advancement happens all at once, and then the second half takes many years after that to figure out. But it’s rare and amazing, and it must be so cool when you have an idea, you do it, and then you realize, oh my God, I just found the next giant leap in the field.

David: It’s like, I unlocked the next level, to use the video game analogy. I leveled up. So after AlexNet, the whole computer science world is abuzz.

Ben: People are starting to stop doubting neural networks at this point.

David: After AlexNet, the three of them from Toronto—Geoff Hinton, Alex Krizhevsky, and Ilya Sutskever—do the natural thing. They start a company called DNN (Deep Neural Network) Research. This company does not have any products. This company has AI researchers.

Ben: Who just won a big competition.

David: And predictably, as you might imagine, it gets acquired by Google almost immediately.

Ben: Oh, are you intentionally shortening this?

David: That’s what I thought the story was.

Ben: Oh, it is not immediately.

David: Oh, okay.

Ben: There’s a whole crazy thing that happens where the first bid is actually from Baidu.

David: Oh, I did not know that.

Ben: Baidu offers $12 million. Geoff Hinton doesn’t really know how to value the company and doesn’t know if that’s fair. He does what any academic would do to best determine the market value of the company. He says, thank you so much. I’m going to run an auction now. I’m going to run it in a highly structured manner where every time anybody wants to bid the clock resets and there’s another hour where anybody else can submit another bid.

David: No way. I didn’t know this. This is crazy.

Ben: He gets in touch with everyone that he knows from the research community who is now working at a big company who he thinks, hey, this would be a good place for us to do our research. That includes Baidu, that includes Google, that includes Microsoft, and there’s one other.

David: Facebook, of course,

Ben: It’s a two-year-old startup.

David: Oh, wait. It does not include Facebook?

Ben: It does not include Facebook. Think about the year. This is 2012, so Facebook’s not really in the AI game yet. They’re still trying to build their own AI lab.

David: Because Yann LeCun and FAIR would start in 2013. Is it Instagram?

Ben: Nope. It is the most important part of the end of this episode.

David: Wait. Well it can’t be Tesla because Tesla is older than that.

Ben: Nope.

David: Well, OpenAI wouldn’t get founded for years. Wow. Okay, you really got me here.

Ben: What company slightly predated OpenAI doing effectively the same mission?

David: Oh, of course. Hiding in plain sight. DeepMind. Wow.

Ben: DeepMind, baby. They are the fourth bidder in a four-way auction for DNN Research. Now of course, right after the bidding starts, DeepMind has to drop out. They’re a startup. They don’t actually have the cash to be able to buy.

David: Didn’t even cross my mind because my first question was where the hell would they get the money? Because they had no money.

Ben: But Geoff Hinton already knows and respects Demis, even though he is just doing this at-the-time startup called DeepMind.

David: That’s amazing. Wait, how is DeepMind in the auction but Facebook is not?

Ben: Isn’t that wild?

David: That’s wild.

Ben: The timing of this is concurrent with the, it was then called NIPS. Now it’s called NeurIPS Conference. Geoff Hinton actually runs the auction from his hotel room at the Harris Casino in Lake Tahoe.

David: Oh my God. Amazing.

Ben: So the bids all come in, and we got to thank Cade Metz, the author of Genius Makers, great book on the whole history of AI that we’re actually going to reference a lot in this episode. The bidding goes up and up and up. At some point Microsoft drops out, they come back in, told you DeepMind drops out. So it’s Baidu and Google really going at the end.

Finally at some point, the researchers look at each other and they say, where do we actually want to land? We want to land at Google. They stop the bidding at $44 million and just say, Google, this is more than enough money. We’re going with you.

David: Wow. I knew it was about $40 million. I did not know that old story. It’s almost like Google itself, the Dutch auction IPO process. How fitting.

Ben: That’s a perfect DNA. The three of them were supposed to split it 33% each. Alex and Ilya go to Geoff and say, I really think you should have a bigger percent. I think you should have 40% and we should each have 30%. That’s how it ends up breaking down.

David: Wow. What a team. Well that leads to the three of them joining Google Brain directly and turbocharging everything going on there. Spoiler alert, a couple of years later, Astro Teller who would take over running Google X after Sebastian Thrun left, would get quoted in the New York Times in a profile of Google X, that the gains to Google’s core businesses and search and ads and YouTube from Google Brain have way more than funded all of the other bets that they have made within Google X and throughout the company over the years.

Ben: It’s one of these things that if you make something a few percent better that happens to do tens or hundreds of billions of dollars in revenue, you find quite a bit of loose change in those couch cushions.

David: Quite a bit of loose change. But that’s not where the AI history ends within Google. There is another very important piece of the Google AI story that is an acquisition from outside of Google, the AI equivalent of Google’s acquisition of YouTube. That’s what we talked about a minute ago, DeepMind.

But before we tell the DeepMind story, now is a great time to thank a new partner of ours, Sentry.

Ben: Listeners, that is like someone’s standing guard.

David: Sentry helps developers debug everything from errors to latency and performance issues, pretty much any software problem and fix them before users get mad. As their homepage puts it, they are considered “not bad by over four million software developers.”

Ben: And today, we’re talking about the way that Sentry works with another company in the Acquired universe, Anthropic. Anthropic used to have some older monitoring systems in place, but as they scaled and became more complex, they adopted Sentry to find and fix issues faster.

David: When you’re building AI models like we’re talking about all episode here, small issues can ripple out into big ones fast. Let’s say you’re running a huge compute job, like training a model. If one node fails, it can have massive downstream impacts, costing huge amounts of time and money. Sentry helped Anthropic detect bad hardware early so they could reject it before causing a cascading problem and taking debugging down to hours instead of days for them.

Ben: One other fun update from Sentry. They now have an AI debugging agent called Seer. Seer uses all the context that Sentry has about your app usage to run root cause analysis as issues are detected. It uses errors, span data logs and tracing, and your code to understand the root cause, fix it, and get you back to shipping. It even creates pull requests to merge code fixes in.

David: And on top of that, they also recently launched agent and MCP server monitoring. AI tooling tends to offer limited visibility into what’s going on under the hood, shall we say. Sentry’s new tools make it easy to understand exactly what’s going on. This is everything from actual AI tool calls to performance across different models, and interactions between AI and the downstream services.

Ben: We’re pumped to be working with Sentry. We’re big fans of the company, and of all the great folks we’re working with there, they have an incredible customer lists, including not only Anthropic but Cursor, Versel, Linear and more.

Actually, if you’re in San Francisco or the Bay Area, Sentry is hosting a small invite-only event with Dave and I in San Francisco for product builders on October 23rd. You can register your interest at sentry.io/acquired, and just tell them that Ben and David sent you.

All right, David, DeepMind. I like your framing. The YouTube of AI.

David: The YouTube of AI for Google.

Ben: They bought this thing for, we’ll talk about the purchase price, but it’s worth what, $500 billion today? This is as good as Instagram or YouTube in terms of greatest acquisitions of all time.

David: 100%. I remember when this deal happened, just like I remember when the Instagram deal happened.

Ben: Because the number was big at the time.

David: It was big, but I remember it for a different reason. It was like when Facebook bought Instagram, like oh my God, this is wow. What a tectonic shift in the landscape of tech. In January 2014, I remember reading on TechCrunch this random news.

Ben: You’re like, Deep what?

David: That Google is spending a lot of money to buy something in London that I’ve never heard of, that’s working on artificial intelligence?

Ben: This really illustrates how outside of mainstream tech AI was at the time.

David: And then you dig in a little further, and this company doesn’t seem to have any products. It also doesn’t even really say anything on its website about what DeepMind is. It says it is a “cutting edge artificial intelligence company.”

Ben: Wait. Did you look this up on the Wayback machine?

David: I did.

Ben: Ah, nice.

David: To build general purpose learning algorithms for simulations, e-commerce, and games. This is 2014 and this does not compute, does not register.

Ben: Simulations, e-commerce, and games. It’s a random spattering of…

David: Exactly. It turns out though, not only was that description of what DeepMind was fairly accurate, this company and this purchase of it by Google was the butterfly flapping its wings equivalent moment that directly leads to OpenAI, ChatGPT, Anthropic, and basically everything.

Ben: Certainly Gemini.

David: That we know. Yeah, Gemini directly in the world of AI today.

Ben: And probably xAI, given Elon’s involvement.

David: Of course xAI.

Ben: In a weird way, iIt leads to Tesla’s self-driving too, Karpathy.

David: Definitely. So what is the story here? DeepMind was founded in 2010 by a neuroscience PhD named Demis Hassabis.

Ben: Who previously started a video game company?

David: Oh yeah, and a postdoc named Shane Legg at University College London, and a third co-founder who was one of Demis’ friends from growing up, Mustafa Suleyman. This was unlikely, to say the least.

Ben: This would go on to produce a Knight and Nobel Prize winner.

David: Demis the CEO was a childhood chess prodigy-turned-video game developer who, when he was age 17 in 1994, had gotten accepted to the University of Cambridge, but he was too young. The university told him, hey, take a gap year, come back.

He decided that he was going to go work at a video game developer at a video game studio called Bullfrog Productions for the year. While he’s there, he created the game Theme Park, if you remember that. It was like a theme park version of Sim City. This was a big game. This was very commercially successful. Rollercoaster Tycoon would be a clone of this that would have many, many sequels over the years.

Ben: Oh, I played a ton of that.

David: It sells 15 million copies in the mid-90s. Wow, wild. Then after this, he goes to Cambridge, studies computer science there. After Cambridge, he gets back into gaming, founds another game studio called Elixir that would ultimately fail. Then he decides, you know what? I’m going to go get my PhD in neuroscience. That is how Demis ends up at University College London.

There he meets Shane Legg who’s there as a postdoc. Shane is a self-described (at the time) member of the Lunatic Fringe in the AI community, in that he believes—this is 2008, 09 10—that AI is going to get more and more and more powerful every year, and that it will become so powerful that it will become more intelligent than humans. Shane is one of the people who actually popularizes the term artificial general intelligence (AGI).

Ben: Oh, interesting. Which of course lots of people talk about now and approximately zero people were afraid of that. You had the Nick Bostrom type folks, but very few people were thinking about super intelligence or the singularity or anything like that.

For what it’s worth, not Elon Musk, he’s not included in that list because Demis would be the one who tells Elon about this.

David: Yes, we’ll get to it. Demis and Shane hit it off. They pull in Mustafa, Demis’ childhood friend, who is himself extremely intelligent. He had gone to the University of Oxford and then dropped out (I think) at age 19 to do other startupy type stuff.

The three of them decide to start a company, DeepMind, the name (of course) being a reference to deep learning, Geoff Hinton’s work and everything coming out of the University of Toronto. The goal that the three of these guys have of actually creating an intelligent mind with deep learning, like Geoff, Ilya, and Alex aren’t really thinking about this yet. As we said, this is lunatic fringe-type stuff.

Ben: AlexNet, the cat paper, that whole world is about better classifying data. Can we better sort into patterns? It’s a giant leap from there to say, oh, we’re going to create intelligence.

David: I think probably some people, probably almost certainly at Google were thinking, oh, we can create narrow intelligence that’ll be better than humans at certain tasks.

Ben: A calculator is better than humans at certain tasks.

David: But I don’t think too many people were thinking, oh, this is going to be general intelligence smarter than humans. So they decide on the tagline for the company is going to be solve intelligence and use it to solve everything else.

Ben: Ooh, I like it.

David: I like it, yeah. They’re good marketers, too, these guys. There’s just one problem. To do what they want to do…

Ben: Money. Just say it. Money. Money is the problem.

David: Right, money is the problem for lots of reasons. But even more so than any other given startup in the 2010 era, it’s not like they can just go spin up an AWS instance, build an app, and deploy it to the app store. They want to build really, really, really, really, really big deep learning neural networks that requires Google-size levels of compute.

Ben: Well, it’s interesting. Actually, they don’t require that much funding yet. The AI of the time was go grab a few GPUs. We’re not training giant LLMs. That’s the ambition eventually, but right now what they just need to do is raise a few million dollars. But who’s going to give you a few million dollars when there’s no business plan, when you’re just trying to solve intelligence? You need to find some lunatics.

David: It’s a tough sell to VCs.

Ben: Except for the exact right—

David: As you say, they need to find some lunatics.

Ben: Oh, I chose my words carefully.

David: We use the term lunatic in a…

Ben: It’s endearing-ish.

David: …most endearing possible way here, given that they were all basically right. So in June 2010, Demis and Shane managed to get invited to the Singularity Summit in San Francisco, California.

Ben: Because they’re not raising money for this in London.

David: Definitely not. I think they tried for a couple of months and learned that that was not going to be a viable path. The Singularity Summit, organized by Ray Kurzweil—a future Google employee (I think), Chief Futurist, noted futurist—Eliezer Yudkowsky, and Peter Thiel. Demis and Shane are excited about getting this invite. They’re like, this is probably our one chance to get funded.

Ben: But we probably shouldn’t just walk in guns ablazed and say, Peter, can we pitch you?

David: Yeah. So they finagle their way into Demis getting to give a talk on stage at the summit.

Ben: Always the hack.

David: This is great. This is going to be the hack. The talk is going to be our pitch to Peter and Founders Fund—Peter has just started Founder’s Fund at this point, member of the PayPal Mafia, very wealthy.

Ben: I think he had a big Roth IRA at this point is the right way to frame it.

David: Big Roth IRA that he had invested in Facebook, first investor in Facebook. He is the perfect target, the architect, the presentation at the summit to be a pitch directly to Peter. Essentially a thinly veiled pitch.

Shane has a quote in Parmy Olson’s great book Supremacy that we used as a source for a lot of the DeepMind story. Shane says, “We needed someone crazy enough to fund an AGI company. Somebody who had the resources not to sweat a few million and liked super ambitious stuff.” They also had to be massively contrarian because every professor that he would go talk to would certainly tell him, absolutely do not even think about funding this. That Venn diagram sure sounds a lot like Peter Thiel.

So they show up at the conference, Demis is going to give the talk, goes out on stage, he looks out into the audience, Peter is not there. Turns out Peter wasn’t actually that involved in the conference.

Ben: No, he is a busy guy. He’s a co-founder, co-organizer, but is a busy guy.

David: Guy’s like, shoot, oh, we missed our chance. What are we going to do? And then fortune turns in their favor. They find out that Peter is hosting an after party that night at his house in San Francisco.

They get into the party, Demis seeks out Peter. Demis is very, very, very smart. As anybody who’s ever listened to him talk would immediately know. He’s like, rather than just pitching Peter head on, I’m going to come about this obliquely.

He starts talking to Peter about chess because he knows as everybody does that Peter Thiel loves chess. Demis had been the second highest ranked player in the world as a teenager in the under 14 category.

Ben: Good strategy.

David: Great strategy. The man knows his chess moves. Peter’s like, hmm, I like you. You seem smart. What do you do? And Demis explains he’s got this AGI startup, they’re actually here, he give a talk on stage as part of the conference, people are excited about this. And Peter says, oh, okay. All right, come back to Founders Fund tomorrow and give me the pitch.

They do. They make the pitch. It goes well. Founders Fund leads DeepMind’s seed round of about $2 million. My, how times have changed for AI company seed rounds these days.

Ben: Oh yes.

David: Imagine leading DeepMind’s seed round with less than a $2 million check. Through Peter and Founders Fund, they get introduced…

Ben: Hey Elon, you should meet this guy.

David: …to another member of the PayPal Mafia. Elon Musk.

Ben: It’s teed up in a pretty low key way. Hey Elon, you should meet this guy. He’s smart. He’s thinking about artificial intelligence. Elon says, great, come over to SpaceX. I’ll give you the tour of the place.

Demis comes over for a launch and a tour of the factory. Of course, Demis thinks it’s very cool, but really he’s trying to reorient the conversation over to artificial intelligence.

I’ll read this great excerpt from an article in the Guardian. “Musk told Hassabis his priority was getting to Mars as a backup planet in case something went wrong here. I don’t think he’d thought much about AI at this point.

Hassabis pointed out a flaw in his plan. I said, what if AI was the thing that went wrong here, then being on Mars wouldn’t help you, because if we got there, then it would obviously be easy for an AI to get there through our communication systems or whatever it was.

He hadn’t thought about that. He sat there for a minute without saying anything, just thinking, hmm, that’s probably true. Shortly after, Musk too became an investor in DeepMind.”

David: Yes. Yes. Yes.

Ben: I think it’s crazy that Demis is the one that woke Elon up to this idea of, we might not be safe from the AI on Mars either.

David: I hadn’t considered that.

Ben: So this is the first time the bit flips for Elon of, we really need to figure out a safe, secure AI for the good of the people, that seed being planted in his head. Which of course is what DeepMind’s ambition is. We are here doing research for the good of humanity like scientists in a peer reviewed way.

David: I think all that is true. Also in the intervening months to year after this meeting between Demis and Elon and Elon investing in DeepMind, Elon also starts to get really, really excited and convinced about the capabilities of AI in the near term, and specifically the capabilities of AI for Tesla.

Ben: Like with everything else in Elon’s world, once the bit flips and he becomes interested, he completely changes the way he views the world, completely sheds all the old ways and actions that he was taking, and it’s all about what do I most do to embrace this new worldview that I have.

David: And other people have been working on for a while already by this point. AI driving cars. That sounds like it would be a pretty good idea for Tesla.

Ben: It does.

David: So Elon starts trying to recruit as many AI researchers as he possibly can and machine vision and machine learning experts into Tesla. And then AlexNet happens. Man, AlexNet’s really, really, really good at identifying and classifying images and cat videos on YouTube and the YouTube recommender feed. Well, is that really that different from a live feed of video from a car that’s being driven and understanding what’s going on there?

Ben: Can we process it in real time and look at differences between frames?

David: Perhaps controlling the car? Not all that different. Elon’s excitement channeled initially through DeepMind and Demis about AI, and AI for Tesla starts ratcheting up big time.

Meanwhile, back in London, DeepMind is getting to work. They’re hiring researchers, they’re getting to work on models, they’re making some vague noises about products to their investors. Maybe we could do something in shopping, maybe something in gaming like the description on the website at the time of acquisition said. But mostly what they really, really want to do is just build these models and work on intelligence.

Then one day in late 2013, they get a call from Mark Zuckerberg. He wants to buy the company. Mark has woken up to everything that’s going on at Google after AlexNet and what AI is doing for social media feed recommendations at YouTube, the possibility of what it can do at Facebook and for Instagram. He’s gone out and recruited Yann LeCun, Geoff Hinton’s old postdoc who, together with Geoff, is one of the godfathers of AI and deep learning.

Ben: And really popularized the idea of convolutional neural networks, the next hot thing in the field of AI at this point in time.

David: And so with Yann, they have created FAIR (Facebook AI Research), which is a Google Brain rival within Facebook. Remember who the first investor in Facebook was, who’s still on the board…

Ben: Peter Thiel.

David: …and is also the lead investor in DeepMind. Where do you think Mark learned about DeepMind? Peter Thiel.

Ben: Was it? Do you know for sure that it was from Peter?

David: No, I don’t know for sure, but how else could Mark have learned about this startup in London?

Ben: I’ve got a great story of how Larry Page found out about it.

David: Oh, okay. Well we’ll get to that in one sec. Mark calls and offers to buy the company. There are various rumors of how much Mark offered, but according to Parmy Olson in her book Supremacy, the reports are that it was up to $800 million. Company with no products and a long way from AGI.

Ben: That squares with what Cade Metz has in his book that the founders would’ve made about twice as much money from taking Facebook’s offer versus taking Google’s offer.

David: Yup. So Demis of course takes this news to the investor group.

Ben: Which by the way is against everything the company was founded on. The whole aim of the company and what he’s promised the team is that DeepMind is going to stay independent, do research, publish in the scientific community. We’re not going to be captured and told what to do by the whims of a capitalist institution.

David: Yup. Definitely some deal point negotiating that has to happen with Mark and Facebook if this offer is going to come through.

Ben: But Mark is so desperate at this point, he is open to these very large deal point negotiations, such as Yann LeCun gets to stay in New York. Yann LeCun gets to stay operating his lab at NYU.

Yann LeCun is a professor. He’s flexible on some things. Turns out, Mark is not flexible on letting Demis keep control of DeepMind if he buys it. Demis argued for, we need to stay separate and carved out and we need this independent oversight board with his ability to intervene if the mission of DeepMind is no longer being followed, and Mark’s like, no. You’ll be a part of Facebook.

David: And you’ll make a lot of money. As this negotiation is going on, of course the investors in DeepMind get wind of this. Elon finds out about what’s going on. He immediately calls up Demis and says, I’ll buy the company right now with Tesla stock.

This is late 2013, early 2014. Tesla’s market cap is about $20 billion. Tesla stock from then to today is about a 70x runup. Demis, Shane, and Mustafa are like, wow, okay. There’s a lot going on right now. But to your point, they have the same issues with Elon and Tesla that they had with Mark. Elon wants them to come in and work on autonomous driving for Tesla. They don’t want to work on autonomous driving.

Ben: Or at least exclusively.

David: At least exclusively, yup. Then Demis gets a third call from Larry Page.

Ben: Do you want my story of how Larry knows about the company?

David: I absolutely want your story of how Larry knows about the company.

Ben: All right. This is still early in DeepMind’s life. We haven’t progressed all the way to this acquisition point yet. Apparently Elon Musk is on a private jet with Luke Nosek, who’s another member of the PayPal Mafia and an angel investor in DeepMind.

They’re reading an email from Demis with an update about a breakthrough that they had, where DeepMind AI figured out a clever way to win at the Atari game, Breakout. The strategy it figured out with no human training was that you could bounce the ball up around the edges of the bricks and then without needing to intervene, it could bounce around along the top and win the game faster without you needing to have a whole bunch of interactions with the paddle down at the bottom.

They’re watching this video of how clever it is and flying with them on the same private plane is Larry Page.

David: Of course, because Elon and Larry used to be very good friends.

Ben: Yes, and Larry is like, what? Wait, what are you watching? What company is this? And that’s how he finds out.

David: Wow. Elon must have been so angry about all this.

Ben: And the crazy thing is this kinship between Larry and Demis is (I think) the reason why the deal gets done at Google

David: Once the two of them get together, they are like peas in a pod. Larry has always viewed Google as an AI company. Demis of course, views DeepMind so much as an AI company that he doesn’t even want to make any products until they can get to AGI.

Ben: And Demis, in fact—we should share with listeners—told us this when we were talking to him to prep for this episode, just felt like Larry got it. Larry was completely on board with the mission of everything that DeepMind was doing.

David: And there’s something else very convenient about Google. They already have Brain. Larry doesn’t need Demis, Shane, Mustafa and DeepMind to come work on products within Google. Brain is already working on products within Google. Demis can really believe Larry when Larry says, nah, stay in London. Keep working on intelligence. Do what you’re doing. I don’t need you to come work on products within Google.

Ben: Brain is actively going and engaging with the product groups trying to figure out, hey, how can we deploy neural nets into your product to make it better? That’s they’re reason for being, so they’re happy to agree to this.

David: And it’s working. Brain and neural nets are getting integrated into search, into ads, into Gmail, into everything. It is the perfect home for DeepMind. Home away from home, shall we say.

And there’s a third reason why Google’s the perfect fit for DeepMind—infrastructure. Google has all the compute infrastructure you could ever want right there on tap.

Ben: At least with CPUs so far. So how’s the deal actually happen? Well, after buying DNN Research, Alan Eustace, who David you spoke with, was Google’s head of engineering at the time. He makes up his mind that he wanted to hire all the best deep learning research talent that he possibly could, and he had a clear path to do so. A few months earlier, Larry Page held a strategy meeting on an island in the South Pacific. In Cade Metz’ book, it’s an undisclosed island.

David: Of course he did.

Ben: Larry thought that deep learning was going to completely change the whole industry, so he tells his team, “Let’s really go big,” which effectively gave Alan a blank check to go secure all the best researchers that he possibly could. So in 2013, he decides, I’m going to get on a plane in December before the holidays and go meet DeepMind.

Crazy story about this. Geoff Hinton, who’s at Google at the time, had a thing with his back where he couldn’t sit down. He either has to stand or lay. So a long flight across the ocean is not doable. But he needs to be there as a part of the diligence process. You have Geoff Hinton, you need to use him to figure out if you’re going to buy a deep learning company.

Alan Eustace decides he’s going to charter a private jet, and he’s going to build this crazy custom harness rig so that Geoff Hinton won’t be sliding around when he’s laying on the floor during takeoff and landing.

David: Wow. I was thinking the first part of this, I’m pretty sure Google has planes. They could just get into a Google plane.

Ben: For whatever reason, this was a separate charter.

David: But it’s not solvable just with a private plane. You also need a harness.

Ben: And Alan is the guy who set the record for jumping out of the world’s highest, was it a balloon? I actually don’t know. The highest free fall jump that anyone has ever done. Even higher than that Red Bull stunt a few years before. He’s very used to designing these custom rigs for airplanes. He’s like, oh, no problem. You just need a bed and some straps. I jumped out of the atmosphere in a scuba suit. I think we’ll be fine.

David: That is amazing.

Ben: So they fly to London, they do the diligence, they make the deal. Demis has true kinship with Larry, and it’s done. $550 million US dollars. There’s an independent oversight board that is set up to make sure that the mission and goals of DeepMind are actually being followed. This is an asset that Google owns today that again, I think is worth half a trillion dollars if it’s independent.

David: Do you know what other member of the PayPal Mafia gets put on the ethics board after the acquisition?

Ben: Reid Hoffman?

David: Reid Hoffman.

Ben: Has to be, given the OpenAI tie later.

David: We are going to come back to Reid in just a little bit here.

Ben: After the acquisition, it goes very well, very quickly. Famously, the data center cooling thing happens where DeepMind carved off some part of the team to go and be an emissary to Google and look for ways to use DeepMind. One of them is around data center cooling.

Very quickly, July of 2016, Google announces a 40% reduction in the energy required to cool data centers. Google’s got a lot of data centers, a 40% energy reduction. I actually talked with Jim Gao, who’s a friend of the show and actually led a big part of this project. It was just the most obvious application of neural networks inside of Google right away. Pays for itself.

David: Imagine that paid for the acquisition pretty quickly there.

Ben: David, should we talk about AlphaGo on this episode?

David: Yeah, yeah, yeah.

Ben: I watched the whole documentary that Google produced about it. It’s awesome. This is actually something that you would enjoy watching, even if you’re not researching a podcast episode, and you’re just looking to pull something up and spend an hour or two. I highly recommend it. It’s on YouTube.

It’s the story of how DeepMind, post-acquisition from Google, trained a model to beat the world Go champion at Go. Everyone in the whole Go community coming in thought there’s no chance. This guy, Lee Sedol, is so good that there’s no way that an AI could possibly beat him.

It’s a five game thing. It just won the first three games straight, completely cleaned up and with inventive new creative moves that no human has played before. That’s the big crazy takeaway.

David: There’s a moment in one of the games where it makes a move of people like, is that a mistake? [...] move 37, then 100 moves later it plays out.

Ben: It was completely genius. And humans are now learning from DeepMind’s strategy of playing the game and discovering new strategies.

A fun thing for Acquired listeners who are like, why is it Go? Go is so complicated. Compared to chess, chess has 20 moves that you can make at the beginning of the game in any given turn, then mid-game there are 30–40 moves that you could make.

Go, on any given turn, has about 200. If you think combinatorially, the number of possible configurations of the board is more than the number of atoms in the universe.

That’s a great Demis quote, by the way. He says, “Even if you took all the computers in the world and ran them for a million years as of 2017, that wouldn’t be enough compute power to calculate all the possible variations.”

It’s cool because it’s a problem that you can’t brute force. You have to do something like neural networks, and there is this white space to be creative and explore. It served as this amazing breeding ground for watching a neural network be creative against a human.

David: And of course it’s totally in with Demis’ background and the DNA of the company of playing games. Demis was chess champion. And then after Go, they play StarCraft, right?

Ben: Oh really? I actually didn’t know that.

David: That was the next game that they tackled. StarCraft, a realtime strategy game against an opponent. That’ll come back up in a sec with another opponent here in OpenAI.

Ben: Yes, David. But before we talk about the creation of the other opponent, should we thank another one of our friends here at Acquired?

David: Yes, we should.

Ben: All right, listeners. We are here today to tell you about a new friend of the show. We are very excited about WorkOS.

David: If you’re building software that is used in enterprises, you’ve probably felt the pain of integrating things like SSO, SCIM, permissions, audit logs, and all the other features that are required by big customers. If you haven’t felt this pain yet, just wait until you get your first big enterprise customer, and trust us you will.

Ben: WorkOS turns these potential deal blockers into simple drop-in APIs. While WorkOS had great product/market fit a few years ago with developers who just want to save on some headache, they really have become essential in the AI era.

David: I was shocked when they sent over their latest customer list. Almost all the big AI companies use WorkOS today as the way that they’ve been able to rapidly scale revenue so fast. Companies like OpenAI, Anthropic, Cursor, Perplexity, Sierra, Replit, Vercel, hundreds of other AI startups all rely on WorkOS as their auth solution.

Ben: I called the founder to ask why, and he said it’s basically two things: (1) in the AI era, these companies scale so much faster that they need things like authorization, authentication, and SSO quickly to become enterprise-ready and keep up with customer demand even early in life, unlike older SaaS companies of yesteryear.

And (2) unlike that world where you could bring your own little SaaS product just for you and your little team, these AI products reach deep into your company’s systems and data to become the most effective. IT departments are scrutinizing heavier than ever to make sure that new products are compliant before they can adopt them.

David: It’s this like second order effect of the AI era that the days of, oh, just swipe a credit card, bring your own SaaS solution for your product team. You actually need to be enterprise-ready a lot sooner than you did before.

Ben: It’s not just about picking up that big potential customer for the revenue itself, either. It’s about doing it so your competitors don’t. Enterprise readiness has become so table stakes for companies no matter their stage, and WorkOS is basically the weapon of choice for the best software companies to shortcut this process and get back to focusing on what makes their beer taste better—building the product itself.

David: Amen, Amen. So if you’re ready to get started with just a few lines of code for SAML SCIM, ARBAC, SSO, authorization, authentication, and everything else to please IT admins and their checklists, check out WorkOS. It’s the modern software platform to make all this happen. That’s workos.com, and just tell them that Ben and David sent you.

Ben: All right, David. What are the second-order effects of Google buying DeepMind?

David: Well, there’s one person who is really, really, really upset about this, and maybe two people if you include Mark Zuckerberg, but Mark tends to play his cards a little closer to the vest. Of course, Elon Musk is very upset about this acquisition.

When Google buys DeepMind out from under him, Elon goes ballistic. As we said, Elon and Larry had always been very close. Now here’s Google, who Elon has already started to sour on a little bit as he’s now trying to hire AI researchers. You’ve got Alan Eustace flying around the world, sucking up all of the AI researchers into Google, and Elon’s invested in DeepMind, wanted to bring DeepMind into his own AI team at Tesla, and gone out from under him.

This leads to one of the most fateful dinners in Silicon Valley’s history, organized in the summer of 2015 at the Rosewood Hotel on Sandhill Road. Of course, where else would you do a dinner in Silicon Valley but the Rosewood, by two of the most leading figures in the valley at the time, Elon Musk and Sam Altman. Sam of course, being president of Y Combinator at the time. What is the purpose of this dinner? They are there to make a pitch to all of the AI researchers that Google, and to a certain extent, Facebook have sucked up and basically created this duopoly status on.

Ben: Again, Google’s business model and Facebook’s business model, these feed recommenders or these classifiers turn out to be unbelievably valuable so they can—it’s funny in hindsight saying this—pay tons of money to these people.

David: Tons of money, millions of dollars.

Ben: Take them out of academia and put them into their dirty capitalist research labs inside the companies.

David: Selling advertising. How dirty could you be? And the question in the pitch that Elon and Sam have for these researchers gathered at this dinner is, what would it take to get you out of Google for you to leave? And the answer to go around the table from almost everybody is nothing. You can’t. Why would we leave? We’re getting paid way more money than we ever imagined, many of us get to keep our academic positions and affiliations, and we get to hang out here at Google—

Ben: With each other?

David: —with each other.

Ben: Iron sharpens iron. These are some of the best minds in the world getting to do cutting edge research with enormous amount of resources and hardware at their disposal. It’s amazing.

David: It’s best infrastructure in the world. We’ve got Jeff Dean here. There is nothing you could tell us that would cause us to leave Google. Except there’s one person who is intrigued, and to quote from an amazing Wired article at the time by Cade Metz who would later write Genius Makers, right?

Ben: Exactly.

David: The quote is, “The trouble was so many of the people most qualified to solve these problems were already working for Google. No one at the dinner was quite sure that these thinkers could be lured into a new startup, even if Musk and Altman were behind it. But one key player was at least open to the idea of jumping ship. Then there’s a quote from that key player. ‘I felt like there were risks involved, but I also felt like it would be a very interesting thing to try.’”

Ben: It’s the most Ilya quote of all time.

David: The most Ilya quote of all time because that person was Ilya Sutskever (of course) of AlexNet, DNN Research, and Google, and about to become founding chief scientist of OpenAI. The pitch that Elon and Sam are making to these researchers is let’s start a new nonprofit AI research lab where we can do all this work out in the open. You can publish.

Ben: Free of the forces of Facebook and Google and independent of their control.

David: You don’t have to work on products, you can only work on research. You can publish your work. It will be open, it will be for the good of humanity. All of these incredible advances, this intelligence that we believe is to come will be for the good of everyone, not just for Google and Facebook.

Ben: And for one of the researchers, it seemed too good to be true. They basically weren’t doing it because they didn’t think anyone else would do it. It’s an activation energy problem. By the way, Google came back with a big counter, something like double the offer. I think it was delivered from Jeff Dean personally. Ilya said, nope, I’m doing this. That was massive for getting the rest of the top researchers to go with him.

David: And it was nowhere near all of the top researchers who left Google to do this, but it was enough. It was a group of seven or so researchers who left Google and joined Elon, Sam, and Greg Brockman from Stripe who came over to create OpenAI. Because that was the pitch. We’re all going to do this in the open.

Ben: And that’s totally what it was.

David: It totally is what it was. The stated mission of OpenAI was to “Advance digital intelligence in the way that is most likely to benefit humanity as a whole, unconstrained by a need to generate financial return.”

Ben: Which is fine as long as the thing that you need to fulfill your mission doesn’t take tens of billions of dollars. Here’s how they would fund it, originally. There was a billion dollars pledged. That came from famously Elon Musk, Sam Altman, Reid Hoffman, Jessica Livingston—who (I think) most people don’t realize was part of that initial tranche—and Peter Thiel. Founders Fund (of course) would go on to put massive amounts of money into OpenAI itself later as well.

The funny thing is, it was later reported that a billion dollars was not actually collected. Only about $130 million of it was actually collected to fund this nonprofit. For the first few years, that was plenty for the type of research they were doing, the type of compute they needed.

David: Most of that money was going to paying salaries to the researchers. Not as much as they could make at Google and Facebook, but still $1 million or $2 million for these folks.

Ben: That really worked until it really didn’t. David, what were they doing in the early days?

David: In the first days, it was all hands on deck recruiting and hiring researchers. There was the initial crew that came over. Then pretty quickly after that in early 2016, they get a big, big win when Dario Amodei leaves Google comes over, joins Ilya and crew at OpenAI. Dream team assembling here.

Ben: And was he on Google Brain before this?

David: He was on Google Brain. He, along with Ilya, would run large parts of OpenAI for the next couple of years, before (of course) leaving to start Anthropic.

But we’re still a couple of years away from Anthropic, Clause, ChatGPT, Gemini, everything today, for at least the first year or two. Basically, the plan at OpenAI is let’s look at what’s happening at DeepMind and show the research community that we can do as a new lab, the same incredible things that they’re doing and maybe even do them better.

Ben: Is that why it looks so game-like and game-focused?

David: Yes. They start building models to play games. Famously, the big one that they do is Dota 2 (Defense of the Ancients 2), the massively online battle arena video game, they’re like, all right. Well DeepMind, you’re playing StarCraft. Well, we’ll go play Dota 2. That’s even more complex, more real time.

Ben: And similar to the emergent properties of Go, the game would devise unique strategies that you wouldn’t see humans trying. It clearly wasn’t humans coded their favorite strategies and rules in. It was emergent. They did other things.

They had a product called Universe, which was around training computers to play thousands of games, from Atari games to open world games like Grand Theft Auto. They had something where they were teaching a model how to do a Rubik’s Cube. It was a diverse set of projects that didn’t seem to coalesce around, one of these is going to be the big thing.

David: It was research stuff. It was what DeepMind was doing.

Ben: It was a university research.

David: It was like DeepMind. If you think back to Elon being an investor in DeepMind, being really upset about Google acquiring it out from under him, makes sense.

Ben: And I think Elon deserves a lot of credit for having his name and his time attached to OpenAI at at the beginning. A lot of the big heavy hitter recruiting was Elon throwing his weight behind this ‘I’m willing to take a chance.’

David: Absolutely.

Ben: Okay, so that’s what’s going on over at OpenAI. Doing a lot of DeepMind-like stuff, bunch of projects, not one single obvious big thing they’re coalescing around. It’s not ChatGPT time, let’s put it that way.

Let’s go back to Google. Because last we checked in on them, yeah they bought DeepMind, but they had their talent raided. I don’t want you to get the wrong impression about where Google is sitting, just because some people left to go to OpenAI.

Back in 2013 when Alex Krizhevsky arrives at Google with Geoff Hinton and Ilya Skiver, he was shocked to discover that all their existing machine learning models were running on CPUs. People had asked in the past for GPUs since machine learning workloads were well-suited to run in parallel, but Google’s infrastructure team had pushed back and said the added complexity in expanding and diversifying the fleet, let’s keep things simple. That doesn’t seem important for us.

David: We’re a CPU shop here.

Ben: To quote from Genius Makers, “In his first days at the company, he went out and bought a GPU machine,” this is Alex, “from a local electronic store, stuck it in the closet down the hall from his desk, plugged it into the network, and started training his neural networks on this loan piece of hardware. Just like he did in academia, except this time Google’s paying for the electricity.

Obviously one GPU was not sufficient, especially as more Googlers wanted to start using it too. Jeff Dean and Alan Eustace had also come to the conclusion that Distbelief, while amazing, had to be re-architected to run on GPUs and not CPUs.”

Spring of 2014 rolls around. Jeff Dean and John Giannandrea, who we haven’t talked about this episode…

David: Yeah, JG.

Ben: …yes, you might be wondering. Wait, isn’t that the Apple guy? Yes. He went on to [...] Apple’s head of AI, who at this point in time was at Google and oversaw Google Brain 2014. They sit down to make a plan for how to actually formally put GPUs into the fleet of Google’s data centers, which is a big deal. It’s a big change. But they’re seeing enough reactions to neural networks that they know to do this.

David: After AlexNet is just a matter of time.

Ben: Yeah. So they settle on a plan to order 40,000 GPUs from NVIDIA.

David: Of course. Who else are you going to order them from?

Ben: For a cost of $130 million. That’s a big enough price tag that the request gets elevated to Larry Page, who personally approves it, even though finance wanted to kill it, because he goes, look, the future of Google is deep learning. As an aside, let’s look at NVIDIA at the time. This is a giant, giant order. Their total revenue was $4 billion. This is one order for $130 million.

David: NVIDIA’s primarily a consumer graphics card company at this point.

Ben: Yes, and their market cap is $10 billion. It’s almost like Google gave NVIDIA a secret that hey, not only does this work in research like the ImageNet competition, but neural networks are valuable enough to us as a business to make $100 million-plus investment in right now, no questions asked.

We got to ask Jensen about this at some point. This had to be a tell. This had to really give NVIDIA the confidence, oh we should way forward invest on this being a giant thing in the future.

So all of Google wakes up to this idea. They start really putting it into their products. Google Photos happened. Gmail starts offering typing suggestions. David, as you pointed out earlier, Google’s giant AdWords business started finding more ways to make more money with deep learning.

In particular when they integrated it, they could start predicting what ads people would click in the future. So Google started spending hundreds of millions more on GPUs on top of that $130 million, but very quickly paying it back from their ad system. It became more and more of a no-brainer to just buy as many GPUs as they possibly could.

But once neural nets started to work, anyone using them, especially at Google scale had this problem. Well now we need to do giant amounts of matrix multiplications anytime anybody wants to use one. The matrix multiplications are effectively how you do that propagation through the layers of the neural network. So you have this problem.

David: Totally. There’s the inefficiency of it, but then there’s also the business problem of, wait a minute. It looks like we’re just going to be shipping hundreds of millions soon to be billions of dollars over to NVIDIA every year for the foreseeable future.

Ben: There’s this amazing moment right after Google rolls out speech recognition—their latest use case for neural nets—just on Nexus phones. Because again, they don’t have the infrastructure to support it on all Android phones. It becomes a super popular feature.

Jeff Dean does the math, and figures out if people use this for (I don’t know call it) three minutes a day and we roll it out to all billion Android phones, we’re going to need twice the number of data centers that we currently have across all of Google just to handle it.

David: Just for this feature.

Ben: There’s a great quote where Jeff goes to Urs Hölzle and goes, we need another Google. Or, David as you were hinting at, the other option is we build a new type of chip customized for just our particular use case.

David: Matrix multiplication. Tensor multiplication. A tensor processing unit, you might say.

Ben: Wouldn’t that be nice? So conveniently, Jonathan Ross, who’s an engineer at Google, has been spending his 20% time at this point in history working on an effort involving FPGAs. These are essentially expensive but programmable chips that yield really fantastic results. They decide to create a formal project to take that work, combine it with some other existing work, and build a custom ASIC (application-specific integrated circuit).

Enter, David as you said, the tensor processing unit made just for neural networks, that is far more efficient from GPUs at the time, with the trade-off that you can’t really use it for anything else. It’s not good for graphics processing. It’s not good for lots of other GPU workloads. Just matrix multiplication and just neural networks. But it would enable Google to scale their data centers without having to double their entire footprint.

The big idea behind the TPU, if you’re trying to figure out what was the core insight, they use reduced computational precision. So it would take numbers like 4586.8272 and round it just to 4586.8, or maybe even just 4586 with nothing after the decimal point. This sounds counterintuitive at first. Why would you want less precise rounded numbers for this complicated math? The answer is efficiency.

If you can do the heavy lifting in your software architecture or what’s called quantization to account for it, you can store information as less precise numbers, then you can use the same amount of power and the same amount of memory and the same amount of transistors on a chip to do far more calculations per second. You can either spit out answers faster or use bigger models. The whole thing is quite clever behind the TPU.

The other thing that has to happen with the TPU is it needs to happen now. Because it’s very clear speech to text is a thing. It’s very clear some of these other use cases at Google.

David: Demand for all of this stuff that’s coming out of Google Brain is through the roof immediately.

Ben: And we’re not even two LLMs yet. It’s just like everyone expects some of this, whether it’s computer vision in photos or speech recognition. It’s just becoming a thing that we expect, and it’s going to flip Google’s economics upside down if they don’t have it.

The TPU was designed, verified, built, and deployed into data centers in 15 months. It was not like a research project that could just happen over several years. This was a hair on fire problem that they launched immediately. One very clever thing that they did was they used the FPGAs as a stop gap. Even though they were too expensive on a unit basis, they could get them out as a test fleet and just make sure all the math worked before they actually had the ASICs printed at, I don’t know if it was a TSMC, but fabbed and Ready.

The other thing they did is they fit the TPU into the form factor of a hard drive, so it could actually slot into the existing server racks. You just pop out a hard drive and you pop in a TPU without needing to do any physical re-architecture.

David: Wow. That’s amazing. That’s the most Googly infrastructure story since the corkboards.

Ben: Exactly. Also, all of this didn’t happen in Mountain View. It was at a Google satellite office in Madison, Wisconsin.

David: Whoa. Why Madison, Wisconsin?

Ben: There was a particular professor out of the university, and there were a lot of students that they could recruit from. It was probably them or Epic. Where are you going to go work?

David: Yeah. Wow.

Ben: They also then just kept this a secret.

David: Right. Why would you tell anybody about this?

Ben: Because it’s not like they’re offering these in Google Cloud, at least at first. Why would you want to tell the rest of the world what you’re doing? The whole thing was a complete secret for at least a year before they announced it at Google IO. So really crazy.

The other thing to know about the TPUs is they were done in time for the AlphaGo match. That match ran on a single machine with four TPUs in Google Cloud. Once that worked, obviously that gave Google a little bit of extra confidence to go really, really real production.

That’s the TPU. V1 by all accounts was not great. They’re on V7 or V8 now. It’s gotten much better. TPUs and GPUs look a lot more similar than they used to, than they’ve adopted features from each other. But today, Google it’s estimated has 2–3 million TPUs. For reference NVIDIA shipped—people don’t know for sure—somewhere around 4 million GPUs last year.

People talk about AI chips like it’s this one horse race with NVIDIA. Google has an almost NVIDIA-scale internal thing making their own chips at this point for their own and for Google Cloud customers. The TPU is a giant deal in AI in a way that I think a lot of people don’t realize.

David: This is one of the great ironies and maddening things to OpenAI and Elon Musk, is that OpenAI gets founded in 2015 with the goal of, hey, let’s shake all this talent out of Google and level the playing field, and Google just accelerates.

Ben: They also build TensorFlow. That’s the framework that Google Brain built to enable researchers to build, train, and deploy machine learning models. They built it in such a way that it doesn’t just have to run on TPUs. It’s super portable without any rewrites to run on GPUs or even CPUs too. This would replace the old Distbelief system and be their internal and external framework for enabling ML researchers going forward.

David: Somewhat paradoxically during these years after the founding of OpenAI, yes some amazing researchers are getting siphoned off from Google and Google Brain, but Google Brain is also firing on all cylinders during this timeframe.

Ben: Delivering on the business purposes for Google left and right.

David: And pushing the state-of-the-art forward in so many areas. Then in 2017, a paper gets published from eight researchers on the Google brand team.

Ben: Kind of quietly.

David: These eight folks were obviously very excited about the paper, what it described and the implications of it, and they thought it would be very big. Google itself, cool. This is the next iteration of our language model work. Great.

Ben: Which is important to us. But are we sure this is the next Google? No.

David: No. There are a whole bunch of other things we’re working on that seem more likely to be the next Google. But this paper and its publication would actually be what gave OpenAI the opportunity…

Ben: To build the next Google.

David: …to grab the ball and run with it, and build the next Google, because this is the Transformer paper.

Ben: Okay. Where did the Transformer come from? What was the latest thing that language models had been doing at Google?

David: Coming out of the success of Franz Och’s work on Google Translate and the improvements that happened there.

Ben: In the late 2000s-ish? 2007?

David: Yeah, mid- to late-2000s. They keep iterating on Translate, and then once Geoff Hinton comes on board and AlexNet happens, they switch over to a neural network–based language model for Translate.

Ben: Which was dramatically better and a big, crazy cultural thing, because you’ve got these researchers parachuting in, again led by Jeff Dean, saying, I’m pretty sure our neural networks can do this way better than the classic methods that we’ve been using for the last 10 years. What if we take the next several months and do a proof of concept?

They end up throwing away the entire old code base, and just completely wholesale switching to this neural network. There’s actually this great New York Times magazine story that ran in 2016 about it. I remember reading the whole thing with my jaw on the floor, like wow, neural networks are a big effing deal. And this was the year before the Transformer paper would come out.

David: Before the Transformer paper, yes. They do the rewrite of Google Translate, make it based on recurrent neural networks, which were state-of-the-art at that point in time, and it’s a big improvement.

But as teams within Google Brain and Google Translate keep working on it, there are some limitations. In particular a big problem was that they “forgot” things too quickly. I don’t know if it’s exactly the right analogy, but you might say in today’s Transformer-world–speak, you might say that their context window was pretty short.

Ben: As these language models progressed through text, they needed to remember everything they had read so that when they need to change a word later or come up with the next word, they could have a whole memory of the body of text to do that.

David: One of the ways that Google tries to improve this is to use something called long short-term memory networks or LSTMs as the acronym that people use for this. Basically what LSTMs do is they create a persistent or long short-term memory—you got to use your brain a little bit here—for the model so that it can keep context as it’s going through a whole bunch of steps.

Ben: And people were pretty excited about LSTMs at first.

David: People are thinking like, oh LSTMs are what are going to take language models and large language models mainstream. And indeed in 2016, they incorporated into Google Translate these LSTMs. It reduces the error rate by 60%. Huge jump.

The problem with LSTMs though, they were effective but they were very computationally intensive, and they didn’t parallelize that great. All the efforts that are coming out of AlexNet and then the TPU project of paralyzation—this is the future, this is how we’re going to make AI really work—LSTMs are a bit of a roadblock here.

So a team within Google Brain starts searching for a better architecture that also has the attractive properties of LSTMs, that it doesn’t forget context too quickly but can parallelize and scale better

Ben: To take advantage of all these new architectures.

David: And a researcher named Jakob Uszkoreit had been toying around with the idea of broadening the scope of “attention” in language processing. What if rather than focusing on the immediate words, instead what if you told the model, hey, pay attention to the entire corpus of text, not just the next few words. Look at the whole thing, and then based on that entire context and giving your attention to the entire context, give me a prediction of what the next translated word should be.

Now by the way, this is actually how professional human translators translate text. You don’t just go word by word. I actually took a translation class in college, which was really fun. You read the whole thing of the original in the original language, you get and understand the context of what the original work is, and then you go back and you start to translate it with the entire context of the passage in mind.

It would take a lot of computing power for the model to do this, but it is extremely parallelizable. Jakob starts collaborating with a few other people on the Brain team. They get excited about this. They decide that they’re going to call this new technique the Transformer, because: (a) that is literally what it’s doing. It’s taking in a whole chunk of information, processing, understanding it, and then transforming it. And (b) they also love Transformers as kids. It’s not not why they named it the Transformer

Ben: It’s taking in the giant corpus of text and storing it in a compressed format, right?

David: Yeah.

Ben: I bring this up because that is exactly how you pitch the micro kitchen conversation with Noam Shazeer in 2000–2001, 17 years earlier, who is a co-author on this paper.

David: Speaking of Noam Shazeer, he learns about this project and he decides, hey, I’ve got some experience with this. This sounds pretty cool. LSTMs definitely have problems. This could be promising. I’m going to jump in and work on it with these guys.

It’s a good thing he did because before Noam joined the project, they had a working implementation of the Transformer, but it wasn’t actually producing any better results than LSTMs. Noam joins the team, basically pulls a Jeff Dean, rewrites the entire code base from scratch, and when he is done, the Transformer now crushes the LSTM-based Google Translate solution. It turns out that the bigger they make the model, the better the results get. It seems to scale really, really, really well.

Stephen Levy wrote a piece in Wired about the history of this, and there are all sorts of quotes from the other members of the team just littered all over this piece with things like, “Noam is a magician.” “Noam is a wizard.” “Noam took the idea and came back and said it works now.”

Ben: And you wonder why Noam and Jeff Dean are the ones together working on the next version of Gemini now.

David: Noam and Jeff Dean are definitely two peas in a pod here.

Ben: So we talked to Greg Cprrado from Google Brain, one of the founders of Google Brain, and it was a really interesting conversation because he underscored how elegant the Transformer was. He said it was so elegant that people’s response was often this can’t work. It’s too simple. Transformers are barely a neural network architecture.

David: It was another big change from the AlexNet, Geoff Hinton lineage neural networks.

Ben: It actually has changed the way that I look at the world, because he pointed out that in nature—this is Greg—the way things usually work is the most energy-efficient way they could work. Almost from an evolution perspective that the most simple elegant solutions are the ones that survive because they are the most efficient with their resources.

You can port this idea over to computer science, too, that he said he’s developed a pattern recognition inside of the research lab to realize that you’re probably onto the right solution when it’s really simple and really efficient versus a complex idea.

It’s very clever. I think it’s very true. You know how when you sit and you have a thorny problem and you debate and you whiteboard and you come up with all, and then you’re like, oh my God. It’s so simple. And that ends up being the right answer.

David: There’s an elegance to the Transformer.

Ben: And that other thing that you touched on there, this is the beginning of the modern AI. Just feed it more data. The famous piece, The Bitter Lesson by Rich Sutton wouldn’t be published until 2019. For anyone who hasn’t read it, it’s basically, we always think as AI research, we’re so smart and our job is to come up with another great algorithm, but effectively in every field, from language to computer vision to chess, you just figure out a scalable architecture, and then the more data wins. Just these infinitely scaling…

David: More data, more compute, better results.

Ben: Yes, and this is really the start of when that starts to be like, oh, we have found the scalable architecture that will go at so far for, I don’t know, close to a decade of just more data in, more energy, more compute, better results.

David: So the team and Noam like, yo, this thing has a lot of potential.

Ben: This is more than better Translate. We can really apply this.

David: This is going to be more than better Google Translate. The rest of Google, though, definitely slower to wake up to the potential.

Ben: They build some stuff. Within a year, they build BERT the large language model.

David: Absolutely true. It is a false narrative out there that Google did nothing with the Transformer after the paper was published. They actually did a lot.

Ben: In fact, BERT was one of the first LLMs.

David: They did a lot with Transformer-based large language models. After the paper came out, what they didn’t do was treat it as a wholesale technology platform change.

Ben: They were doing things like BERT and MoM, this other model. They could work it into search results quality. I think that did meaningfully move the needle, even though Google wasn’t bragging about it and talking about it. They got better at query comprehension. They were working it into the core business, just like every other time Google Brain came up with something great.

David: Perhaps one of the greatest decisions ever for value to humanity and maybe one of the worst corporate decisions ever for Google, Google allows this group of eight researchers to publish the paper under the title, Attention Is All You Need, obviously a nod to the classic Beatles song about love.

As of today in 2025, this paper has been cited over 173,000 times in other academic papers, making it currently the 7th most cited paper of the 21st century. I think all of the other papers above it on the list have been out much longer.

Also, of course, within a couple of years, all eight authors of the Transformer paper had left Google to either start or join AI startups, including OpenAI.

Ben: Brutal. Of course, Noam starting Character.AI, which, what do we call it? A hackquisition? He would end up back at Google via some strange licensing, IP, and hiring agreement on the few billion dollars order.

David: Very, very expensive mistake on Google’s part.

Ben: It is fair to say that 2017 begins the five year period of Google not sufficiently seizing the opportunity that they had created.

David: With the Transformer, yes. Speaking of seizing opportunities, what is going on at OpenAI during this time?

Ben: And does anyone think the Transformer’s a big deal over there?

David: Yes, they did. But here’s where history gets really, really crazy. Right after Google publishes the Transformer paper in September of 2017, Elon gets really, really fed up with what’s going on at OpenAI.

Ben: There are seven different strategies. Are we doing video games? Are we doing competitions? What’s the plan?

David: What is happening here as best as I can tell, all you’re doing is just trying to copy DeepMind. Meanwhile, I’m here building SpaceX and Tesla. Self-driving is becoming more and more clear as critical to the future of Tesla. I need AI researchers here, and I need great AI advancements to come out to help what we’re doing at Tesla. OpenAI isn’t cutting it.

He makes an ultimatum to Sam and the rest of the OpenAI board. He says, “I’m happy to take full control of OpenAI and we can merge this into Tesla.” I don’t even know how that would be possible to merge a nonprofit into Tesla.

Ben: But in Elon land, if he takes over as CEO of OpenAI, it almost doesn’t matter. We’re just treating it as if it’s the same company anyway, just like we do with the deals with all of my companies.

David: Or he’s out completely along with all of his funding. Sam and the rest of the board are like, no.

Ben: And as we know now, they’re culling capital into the business. It’s not like they actually got all the cash up front.

David: They’re only $130 million-ish into the billion dollars of commitment. They don’t reach a resolution, and by early 2018, Elon is out. Along with him, the main source of OpenAI’s funding.

Either this is just a really, really, really bad misjudgment by Elon, or the panic that this throws OpenAI into is the catalyst that makes them reach for the Transformer and say, all right, we got to figure things out. Necessity’s the mother of invention. Let’s go for it.

Ben: It’s true. I don’t know if during this personal tension between Elon and Sam, if they had already decided to go all in on Transformers or not. Because the thing you very quickly get to, if you decide Transformers language models, were going all in on that, you do quickly realize you need a bunch of data. You need a bunch of compute, you need a bunch of energy, and you need a bunch of capital.

If your biggest backer is walking away, the 3D chess move is, oh, we got to keep him because we’re about to pivot the company, and we need his capital for this big pivot we’re doing. The 4D chess is, if he walks away, maybe I can turn it into a for-profit company, then raise money into it, and eventually generate enough profits to fund this extremely expensive new direction we’re going in. I don’t know which of those it was.

David: I don’t know either. I suspect the truth is it’s some of both.

Ben: But either way, how nuts is it that: (a) these things happened at the same time, and (b) the company wasn’t burning that much cash and then they decided to go all in on, we need to do something so expensive that we need to be a for-profit company in order to actually achieve this mission, because it’s just going to require hundreds of billions of dollars for the far foreseeable future.

David: Yup. So in June of 2018, OpenAI releases a paper describing how they have taken the Transformer and developed a new approach of pre-training them on very large amounts of general text on the Internet, then fine-tuning that general pre-training to specific use cases.

They also announced that they have trained and run the first proof of concept model of this approach, which they are calling GPT-1 (Generatively Pre-trained Transformer version 1).

Ben: Which we should say is right around the same time as BERT, and right around the same time as another large language model based on the Transformer out of here in Seattle, the Allen Institute.

David: Yes, indeed.

Ben: So it’s not as if this is heretical and a secret. Other AI labs, including Google’s own, are doing it. But from the very beginning, OpenAI seemed to be taking this more seriously, given the cost of it would require betting the company if they continued down this path.

David: Or betting the nonprofit, betting the entity. We’re going to need some new terminology here.

Elon just walked out the door. Where are they going to get the money for this? Sam turns to one of the other board members of OpenAI, Reid Hoffman. Reid, just a year or so earlier, had sold LinkedIn to Microsoft. Reid is now on the board of Microsoft. Reid says, hey, why don’t you come talk to Satya about this?

Ben: Do you know where he actually talks to Satya?

David: Oh, I do. In July of 2018, they set a meeting for Sam Altman and Satya Nadella to sit down while they’re both at the Allen & Company Sun Valley Conference in Sun Valley, Idaho.

Ben: It’s perfect.

David: And while they’re there, they hash out a deal for Microsoft to invest $1 billion into OpenAI in a combination of both cash and Azure Cloud credits. In return, Microsoft will get access to OpenAI’s technology, get an exclusive license to OpenAI’s technology for use in Microsoft’s products.

The way that they will do this is OpenAI, the nonprofit will create a captive for-profit entity called OpenAI LP, controlled by the nonprofit OpenAI Inc. Microsoft will invest into the captive for-profit entity. Reid Hoffman joins the board of this new structure, along with Sam, Ilya, Greg Brockman, Adam D’Angelo, and Tasha McCauley. Thus the modern OpenAI for-profit, non-profit? is created.

Ben: The thing that’s still being figured out even today here in 2025, is created. This is the complete history of AI. This is not just the Google AI episode.

David: Well, these things are totally inextricable. I was just going to say, this is the Google Part III episode. Microsoft, they’re back. Microsoft is Google’s mortal enemy. That, in our first episode on the founding of Google and Search, then in the second episode on Alphabet and all the products that they made, the whole strategy at Google was always about Microsoft. They finally beat them on every single front. And here they are—

Ben: Showing up again saying, what was Satya’s line? We just want to see them dance.

David: I think the line that would come a couple of years later is we want the world to know that we made Google dance. Oh man.

Ben: But this is all still pre-ChatGPT. This is just Sam lining up the financing he needs for what appears to be a very expensive scaling exercise they’re about to embark on with GPT-2 and onward.

David: And this is the right time to talk about why, from OpenAI’s perspective, Microsoft is the absolute perfect partner. It’s not just that they have a lot of money.

Ben: Although that helps.

David: That helps a lot. But more important than money, they have a really, really great public cloud—Azure.

Ben: OpenAI is not going to go buy a bunch of NVIDIA GPUs and then build their own data center here at this point in 2018. That’s not the scale of company that they are. They need a cloud provider in order to actually do all the compute that they want to do. If they were back at Google and these researchers are doing it, great. Then they have all the infrastructure. But OpenAI needs to tie themselves to someone with the infrastructure.

David: And there’s basically only two non-Google options. They’re both in Seattle, and hey, one of them in Microsoft is really interested, also has a lot of cash. It seems like a great partnership.

Ben: That’s true. I wonder if they did talk to AWS at all about it. Because I think—this is a crazy Easter egg; I hesitate to say it out loud—AWS was actually in the very first investment with Elon in OpenAI.

David: Oh wow.

Ben: And I don’t know if it was in the form of credits or what the deal was, but I’d seen it reported a couple of places that AWS actually was in that nonprofit round.

David: In the nonprofit funding, the donations to the early OpenAI.

Ben: Anyway, Microsoft’s OpenAI, they end up tying up.

David: A match made in heaven.

Ben: Satya and Sam are on stage together talking about how this amazing partnership in marriage has come together, and they’re off to model training.

David: And this paves the way for the GPT era of OpenAI. But before we tell that story…

Ben: …now is a great time to thank one of our favorite companies, Shopify.

David: And this is really fun because we have been friends and fans of Shopify for years. We just had Toby on ACQ2 to talk about everything going on in AI, and everything that has happened at Shopify in the six years now since we covered the company on Acquired.

Ben: It’s been a pretty insane transformation for them.

David: Back at their IPO, Shopify was the go-to platform for entrepreneurs and small businesses to get online. What’s happened since is that it’s still true, and Shopify has also become the world’s leading commerce platform for enterprises of any size, period.

Ben: What’s so cool about the company is how they’ve managed to scale without losing their soul. Even though companies like Everlane and Vori, even older established companies like Mattel are doing billions of revenue on Shopify, the company’s mission is still the same as the day Toby founded it—to create a world where more entrepreneurs exist.

David: Ben, you got to tell everyone your favorite enterprise brand that is on Shopify.

Ben: Oh, I’m saving that for next episode. I have a whole thing planned for episode two of this season.

David: Okay, great. Anyway, the reason enterprises are now also using Shopify is simple: because businesses of all sizes just sell more with Shopify. They built this incredible ecosystem where you can sell everywhere. Obviously your own site—that’s always been true—but now with Shopify, you can easily sell on Instagram, YouTube, TikTok, Roblox, Roku, ChatGPT, Perplexity, anywhere. Plus with Shop Pay (their accelerated checkout), you get amazing conversion and it has a built-in user base of 200 million people who have their payment information already stored with it.

Ben: Shopify is the ultimate example of not doing what doesn’t make your beer taste better. Even if you’re a huge brand, you’re not going to build a better e-commerce platform for your product. But that is what Toby and Shopify’s entire purpose is, so you should use them.

David: So whether you’re just getting started or already at huge scale, head on over to shopify.com/acquired, and just tell them that Ben and David sent you.

Ben: All right. What are we, in GPT-2? Is that what’s being trained right here?

David: Yes, GPT-2.

Ben: This was the first time I heard about it. Data scientists around Seattle were talking about this. Cool.

David: After the first Microsoft partnership, the first billion dollar investment, in 2019 OpenAI releases GPT-2, which is still early but very promising, that can do a lot of things.

Ben: A lot of things, but it required an enormous amount of creativity on your part. You had to be a developer to use it, and if you were a consumer there was a very heavy load put on you. You had to go write a few paragraphs, paste those few paragraphs into the language model, and then it would suggest a way to finish what you were writing based on the source paragraphs. But it wasn’t interactive.

David: It was not a chat interface.. There was no interface (essentially) for it.

Ben: It was an API.

David: But it can do things like obviously translate text. Google’s been doing that for a long time. But GPT-2, you could do stuff like make up a fake news headline, give it to GPT-2, and it would write a whole article. You would read it and you’d be like, sounds like it was written by a bot. But again, there was no front door to it for normal people. You had to really be willing to wait in the muck to use this thing.

Then the next year in June of 2020, GPT-3 comes out. Still no front door user interface to the model, but it’s very good. GPT-2 showed the promise of what was possible. GPT-3, it’s starting to be in the conversation of can this thing pass the Turing test? You have a hard time distinguishing between articles that GPT wrote and articles that humans wrote. It’s very good, and there starts to be a lot of hype around this thing.

Ben: Even though consumers aren’t really using it, the broader awareness is that there’s something interesting on the horizon. I think the number of AI pitch decks that VCs are seeing is starting to tick up around this time.

David: As is the NVIDIA stock price. Then in the next year, in the summer of 2021, Microsoft releases GitHub Copilot using GPT-3. This is the first, not just Microsoft product that comes out with GPT baked into it, but first…

Ben: Productization.

David: …product anywhere, yeah. First productization of GPT.

Ben: Of any OpenAI technology.

David: It’s baked. This starts a massive change in how software gets written in the world.

Ben: Slowly, then all at once. It’s one of these things where at first just a few software engineers and there was a lot of whispers of how cool is this? It makes me a little bit more efficient. And now you get all these comments, like 75% of all companies code is written with AI.

David: After that, Microsoft invests another $2 billion in OpenAI, which seemed like a lot of money at the time. That takes us to the end of 2021. There’s an interesting context shift that happens around here.

Ben: The bottom falls out on tech stocks, crypto, the broader markets, really, everyone suddenly goes from risk on to risk off. Part of it was war on Ukraine, but a lot of it was interest rates going up. Google gets hit really hard. The high watermark was November 19th of 2021. Google was right at $2 trillion of market cap. About a year after that slide began, they were worth a trillion dollars, nearly a 50% drawdown.

David: Wow, so towards the end of 2022, leading up to the launch of ChatGPT.

Ben: People (I think) are starting to realize Google’s slow. They’re slow to react to things. It feels like they’re an old crusty company. Are they like the Microsoft of the 2000s where they haven’t had a breakthrough product in a while. People are not bright on the future of Google. And then ChatGPT comes out.

David: Which means if you were bullish on Google back then and contrarian, you could have invested at a trillion dollar market cap.

Ben: Which is interesting. In October of 2021, the market was saying that the forthcoming AI wave will not be a strength for Google. Or maybe what it was saying is we don’t even know anything about a forthcoming AI wave because people are talking about AI.

But they’ve been talking about VR. They’ve been talking about crypto. They’ve been talking about all this frontier tech, and that’s not the future at all. This company just feels slow and unadaptive. Slow and unadaptive at that point in history (I think) would’ve been a fair characterization. They had an internal chat bot, right?

David: Yes, they did. Before we talk about ChatGPT, Google had a chat bot. Noam Shazeer, incredible engineer, re-architected the Transformer, made it work, one of the lead authors of the paper, storied career within Google, has all of this sway, should have all of this sway within the company. After the Transformer paper comes out, he and the rest of the team are like, guys, we can use this for a lot more than Google Translate. In fact, the last paragraph of the paper—

Ben: Are you about to read the Transformer paper?

David: Yes, I am. “We are excited about the future of attention-based models and plan to apply them to other tasks. We plan to extend the Transformer to problems involving input and output modalities other than text, and to investigate large inputs and outputs such as images, audio, and video.” This is in the paper.

Ben: Wow.

David: Google obviously does not do any of that for quite a while. Noam though, immediately starts advocating to Google leadership. “Hey, I think this is going to be so big, the Transformer that we should actually consider just throwing out the search index and the 10 blue links model and go all in on transforming all of Google into one giant Transformer model.” Then Noam actually goes ahead and builds a chatbot interface to a large Transformer model.

Ben: Is this Lambda?

David: This is before Lambda. Mina is what he calls it. There is a chat bot in the late teens–2020 timeframe that Noam built within Google, that arguably is pretty close to ChatGPT. Now it doesn’t have any of the post training safety that ChatGPT does, so it would go off the rails.

Ben: Someone told us that you could just ask it who should die, and it would come up with names for you of people that should die. It was not a shippable product.

David: It was a very raw, not safe, not post trained chatbot and model. But it existed within Google and they didn’t ship it.

Ben: And technically not only did it not have post training, it didn’t have RLHF either. This very core component of the models today, the reinforcement learning with human feedback, that ChatGPT, I don’t know if it had it in 3, but it did in 3.5, and it did for the launch of ChatGPT. Realistically, it wasn’t launchable, even if it was an OpenAI thing because it was so bad. But a company of Google stature certainly could not take the risk. So strategically they have this working against them.

But aside from the strategy thing, there are two business model problems here. One, if you’re proposing drop the 10 blue links and just turn google.com into a giant AI chat bot. Revenue drops when you provide direct answers to questions versus showing advertisers and letting people click through to websites that upsets the whole apple cart. Obviously they’re thinking about it now, but until 2021, that was an absolute non-starter to suggest something like that.

Two, there were legal risks of sitting in between publishers and users. Google at this point had spent decades fighting the public perception and court rulings that they were disintermediating publishers from readers. There was a very high bar internally, culturally, to clear if you were going to do something like this. Even those info boxes that popped up that took until the 20-teens to make it happen, those really were mostly on non-monetizable queries anyway. Anytime that you were going to say, hey, Google’s going to provide you an answer instead of 10 blue links, you had to have a bulletproof case for it.

David: There was also a brand promise and trust issue, too. Consumers trusted Google so much. For us even today when I’m doing research for Acquired, we need to make sure we get something right. I’m going to Google.

Ben: I look something up in Claude. It gives me an answer. I’m like, that’s a really good answer. Then I verify by searching Google that I can find those facts, too, if I can’t click through the sources on Claude. That’s my workflow.

David: Which sounds funny today, but it’s important. If you’re going to propose replacing the 10 blue links with a chat bot, you need to be really sure that it’s going to be accurate. And in 2020–2021, that was definitely not the case. Arguably still isn’t the case today.

There also wasn’t a compelling reason to do it because nobody was really asking for this product. Noam knew and people in Google knew that you could make a chatbot interface to a Transformer-based LLM and that was a really compelling product. The general public didn’t know. OpenAI didn’t even really know. GPT was out there.

Ben: Do you know the story of the launch of ChatGPT?

David: Well, I think I do. I have it in my notes here.

Ben: All right. So they’ve got GPT-3.5. It’s becoming very, very useful.

David: This is late 2022. They’ve got 3.5.

Ben: But there’s still this problem of how am I supposed to actually use it? How does it productized? And Sam just says we should make a chat bot. That seems like a natural interface for this. Can someone just make a chat and within a week, internally…

David: Someone makes a chat.

Ben: They just turn calls to the ChatGPT 3.5 API into a product where you’re just chatting with it, and every time you kick off a chat message, it just calls GPT 3.5 on the API. And that turns out to be this magic product.

I don’t think they expected it. Servers are tipping over. They’re working with Microsoft to try to get more compute. They’re cutting deals with Microsoft in real time to try to get more investment, to get more Azure credits or get advances on their Azure credits, in order to handle the incredible load in November of 2022 that’s coming in of people wanting to use this thing.

They also just throw up a paywall randomly because they thought that the business was going to be an API business. They thought that the projections were all about how much revenue they were going to do through B2B licensing deals. Then they just realized, oh, there are all these consumers trying to use this. Put up a paywall to at least dampen the most expensive use of this thing, so we can offset the costs or slow the rollout.

David: This isn’t a Google Search 89% gross margin stuff here.

Ben: They end up having incredibly fast revenue takeoff just from the quick stripe paywall that they threw up over a weekend to handle all the demand. To say that OpenAI had any idea what was coming would also be completely false. They did not get that this would be the next big consumer product when they launched it.

David: Ben Thompson loves to call OpenAI the accidental consumer tech company. It was definitely accidental. Now there is actually another slightly different version of the motivation for launching the chat.

Ben: Is this the Dario…

David: Interface? Yeah, the Dario and Anthropic version. Anthropic was working on what would become Claude. Rumors were out there and people at OpenAI got wind of like, oh hey, Anthropic and Dario are working on a chat interface. We should probably do one too. If we’re going to do one, we should probably launch it before they launch theirs.

I think that had something to do with the timing. But again, I don’t think anybody, including OpenAI, realized what was going to happen, which is, Ben you alluded to it but to give the actual numbers.

On November 30th, 2022…

Ben: Basically Thanksgiving.

David: …OpenAI launches a research preview of an interface to the new GPT-3.5 called ChatGPT. That morning on the 30th, Sam Altman tweets, “Today we launched ChatGPT. Try talking with it here,” and then a link to chat.openai.com.

Within a week, less than a week, actually, it gets one million users. By the end of the year, one month later, December 31st, 2022, it has 30 million users. By the end of the next month, by the end of January 2023, so two months after launch, it crosses 100 million registered users, the fastest product in history to hit that milestone. Completely insane.

Before we talk about what that unleashes within Google, which is the famous code red, to rewind a little bit back to Noam in the chat bot within Google Mina, Google does keep working on Mina. They develop it into something called Lambda, which is also a chatbot, also internal.

Ben: I think it was a language model. At this point in time, they still differentiated between the underlying model brand name and the application name.

David: Lambda was the model, then there also was a chat interface to Lambda that was internal for Google use only. Noam is still advocating to leadership. We got to release this thing. He leaves in 2021 and found a chat bot company, Character AI, that still exists to this day and they raise a lot of money, as you would expect. Then Google ultimately in 2024 after ChatGPT launches pays $2.7 billion (I think) to do a licensing deal with Character AI, the net of which Noam comes back to Google.

Ben: I think Larry and Sergey were like, hmm, if we’re going to compete seriously, we need Noam back and blank check to go get him.

David: So throughout 2021–2022, Google’s working on the Lambda model and then the chat interface to it. In May of 2022, they do release something that is available to the public called AI Test Kitchen, which is an AI product test area where people can play around with Google’s internal AI products, including the Lambda chat interface.

Ben: And all fairness predates ChatGPT.

David: Do you know what they do to NeRF chat so that it doesn’t go too far off the rails? This is amazing.

Ben: No.

David: For the version of Lambda Chat that is in AI Test Kitchen, they stop all conversations after five turns. You can only have five turns of conversation with the chat bot, then it’s just, and we’re done for today. Thank you. Goodbye.

Ben: Oh wow.

David: The reason they did that was for safety. The more turns you had with it, the more likely it would start to go off the rails.

Ben: Honestly, it was a fair concern. This thing was not for public consumption. If you remember back a few years before, Microsoft released Tay, which was this crazy, racist chatbot.

David: They launched it as a Twitter bot, right? It was going off the rails on Twitter. This was in 2016 (I think).

Ben: Maximal impact of badness. Despite Google all the way back in 2017, Sundar declared we are an AI-first company, is being understandably very cautious in real public AI launches, especially on consumer-facing things.

David: And as far as anyone else is concerned before ChatGPT, they are an AI-first company, and they’re launching all this amazing AI stuff. It’s just within the vector of their existing products.

So ChatGPT comes out, becomes the fastest product in history to 100 million users. It is immediately obvious to Sundar, Larry, Sergey, all of Google leadership, that this is an existential threat to Google. ChatGPT is a better user experience to do the same job function that Google Search does.

Ben: And to underscore this, if you didn’t know it in November of 2022, you sure knew it by February of 2023 because good old Microsoft, our biggest scariest enemy. announces a new Bing powered by OpenAI.

Satya has a quote, “It’s a new day for search. The race starts today.” There is an announcement of a new AI-powered search page. He says, “We want to rethink what search was meant to be in the first place. In fact, Google’s success in the initial days came by re-imagining what could be done in search. I think the AI era we’re entering gets us to think about it.”

This is the worst possible thing that could happen to Google, that now Microsoft can actually challenge Google on their own turf, intent on the Internet with a legitimately different, better, differentiated product vector. Not what Bing was trying to do copycat. This is the full leapfrog, and they have the technology partnership to do it.

David: Or so everybody thinks at the moment.

Ben: Oh my God, terrifying.

David: This is when Satya says the quote in an interview around this launch with Bing. “I want people to know that we made Google dance.” Oh boy. Well hey, if you come at the king, you’d best not Miss. And this big launch misses.

So what happens in Google? December 2022, even before the big launch but after the ChatGPT moment, Sundar issues a code red within the company.

Ben: What does that mean?

David: Up until this point, Google and Sundar. and Larry and everyone had been thinking about AI as a sustaining innovation in Clay Christensen’s terms. This is great for Google, this is great for our products. Look at all these amazing things that we’re doing.

Ben: It further entrenches incumbents.

David: It further is entrenching our lead in all of our already leading products.

Ben: We can deploy more capital in a predictable way to either drive down costs or make our product experiences that much better than any startup could make.

David: Got more monetized that much better all the things. Once ChatGPT comes out, on a dime overnight, AI shifts from being a sustaining innovation to a disruptive innovation. It is now an existential threat. And many of Google’s strengths from the last 10, 15, 20 years of all the AI work that’s happened in the company are now liabilities. They have a lot of existing castles to protect.

Ben: That’s right. They have to run everything through a lot of filters before they can decide if it’s a good idea to go try to out-OpenAI OpenAI.

David: This code red that Sundar issues to the company is actually a huge moment. Because what it means and what he says is we need to build and ship real native AI products ASAP.

This is actually what you need to do, the textbook response to a disruptive innovation as the incumbent. You need to not bury your head in the sand and you need to say, okay, we need to actually go build and ship products that are comparable to these disruptive innovators.

Ben: And you need to be laser operationally in all the details to try and figure out where is it that the new product is actually cannibalizing our old product, and where is it that the new product can be complimentary. And just lean into all the ways in which you can be complimentary in all the different little scenarios.

Really what they’ve been trying to do this ballet from 2022 onward is protect the growth of Search while also creating the best AI experiences they can. It’s very clever the way that they do AI overviews for some but not all queries. They have AI mode for some, but not all users. And then they have Gemini, the full AI app, but they’re not redirecting google.com to Gemini. It’s this very delicate dance of protecting the existing franchise while also building a hopefully non-cannibalizing as much as we can new franchise.

David: And you see them really going hard (I think) building leading products in non-search cannibalizing categories like video.

Ben: Veo 3 or Nano Banana. These are things that don’t in any way cannibalize the existing franchise. They in fact use some of Google strength, all the YouTube training data and stuff like that.

David: So what happens next? As you might expect, it gets worse before it gets better. Code red goes out December 2022.

Ben: Bard, baby. Launch Bard.

David: Oh boy. Well, even before that, January 23 when OpenAI hits 100 million registered users for ChatGPT, Microsoft announces they’re investing another $10 billion in OpenAI, and says that they now own 49% of the for-profit entity.

Incredible in and of itself, but then now think about this from the Google lens of Microsoft, our Enemy. They now arguably own, obviously in retrospect here, they don’t own OpenAI, but it seems at the time, oh my God. Microsoft might now own OpenAI, which is our first true existential threat in our history as a company.

Ben: Not great, Bob.

David: So then February 2023, the big integration launches. Satya has the quote about wanting to make Google dance. Meanwhile, Google is scrambling internally to launch AI products as fast as possible. The first thing they do is they take the Lambda model and the chatbot interface to it. They rebrand it as Bard.

Ben: They ship that publicly and—

David: They release it immediately. February 2023, ship it publicly. available GA to anyone.

Ben: Which maybe was the right move, but God, it was a bad product.

David: It was really bad.

Ben: I didn’t know the term at the time RLHF, but it was clear it was missing a component of some magic that ChatGPT had. This reinforcement learning with human feedback where you could really tune the appropriateness, the tone, the voice, the correctness of the responses, it just wasn’t there.

David: To make matters worse, in the launch video for Bard—this is a choreographed, prerecorded video where they’re showing conversations with Bard—Bard gives an inaccurate factual response to one of the queries that they include in the video.

Ben: This is one of the worst keynotes in history.

David: After the Bard launch and this keynote, Google stock drops 8% on that day. Then, like we were saying, once the actual product comes out, it becomes clear it’s just not good. It pretty quickly becomes clear, it’s not just that the chat bot isn’t good, the model isn’t good.

In May, they replaced Lambda with a new model from the Brain team called PaLM. It’s a little bit better, but it’s still clearly behind not only GPT-3.5, but in March of 2023, OpenAI comes out with GPT-4, which is even better. You can access that now through ChatGPT.

Here is where Sundar makes two really, really big decisions. Number one, he says, “We cannot have two AI teams within Google anymore. We’re merging Brain and DeepMind into one entity called Google DeepMind.”

Ben: Which is a giant deal. This is in full violation of the original deal terms of bringing DeepMind in.

David: And the way he makes it work is he says, Demis, you are now CEO of the AI division of Google, Google DeepMind. This is all hands on deck and you and DeepMind are going to lead the charge, you’re going to integrate with Google Brain, and we need to change all of the past 10 years of culture around building and shipping AI products within Google.

Ben: To further illustrate this, when Alphabet became Alphabet, they had all these separate companies, but things that were really core to Google, like YouTube actually stayed a part of Google. DeepMind was its own company. That’s how separate this was. They’re working on their own models. In fact, those models are predicated on reinforcement learning. That was the big thing that DeepMind had been working on the whole time.

Reading in between the lines, it’s Sundar looking at his two AI labs and going, look, I know you two don’t actually get along that well, but look. I don’t care that you had different charters before. I am taking the responsibility of Google Brain and giving it to DeepMind, and DeepMind is absorbing the Google Brain team. I think that’s what you should read into it, because as you look at where the models went from here, they came from DeepMind.

David: There’s a little bit of interesting backstory to this too. Mustafa Suleyman the third co-founder of DeepMind, at some point before this…

Ben: He became the head of Google AI policy or something?

David: …he had already shifted over to Brain and to Google. He stayed there for a little while and then he ended up getting close with who else? Reid Hoffman. Remember Reid is on the ethics board for DeepMind, and Mustafa and Reid leave and go found Inflection AI, which fast forward now into 2024 after the absolute insanity that goes down at OpenAI in Thanksgiving 2023 when Sam Altman gets fired over the weekend during Thanksgiving, and then brought back by Monday when all the team threatened to quit and go to Microsoft.

Ben: OpenAI loves Thanksgiving. Can’t wait for this year.

David: They love Thanksgiving. Yeah, gosh. After all that, which certainly strains the Microsoft relationship. Remember again, Reid is on the board of Microsoft. Microsoft does one of these acquisition type deals with Inflection AI, and brings Mustafa in as the head of AI for Microsoft.

Ben: Crazy,

David: Just wild.

Ben: Crazy turn of events.

David: Okay. That first big decision that Sundar makes is unifying DeepMind and Brain. That was huge. Equally big. he says, “I want you guys to go make a new model, and we’re just going to have one model. That is going to be the model for all of Google internally, for all of our AI products externally. It’s going to be called Gemini. No more different models, no more different teams. Just one model for everything. This is also a huge deal.

Ben: It’s a giant deal. And it’s twofold. It’s push and it’s pull. It’s saying, hey, if anyone’s got a need for an AI model, you got to start using Gemini. But two, it’s actually the Google Plus thing where they go to every team and they start saying, Gemini is our future. You need to start looking for ways to integrate Gemini into your product.

David: I’m so glad you brought up Google Plus. This came up with a few folks I spoke to in the research. Obviously, this is all playing out realtime, but the point a lot of people at Google made is the Gemini situation is very different than the Google Plus situation.

This is a technical thing: (a) which has always been Google’s wheelhouse, but (b) even more importantly, this is the rational business thing to do in the age of these huge models. Even for a company like Google, there are massive scaling laws to models.

Ben: The more data you put in, the better it’s going to get, the better all the outputs are going to be.

David: And because of scaling laws, you need your models to be as big as possible in order to have the best performance possible. If you’re trying to maintain multiple models within a company, you’re repeating multiple huge costs to maintain huge models. You definitely don’t want to do that. You need to centralize on just one model.

Ben: It’s interesting. There’s also something to read into where at first it was the Gemini model underneath the Bard product. Bard was still the consumer name. Then at some point they said, no, we’re just calling it all Gemini. Gemini became the user-facing name.

Also, this pulls in my Quintessence from the Alphabet episode. I know it’s a little bit woo-woo, but with Google saying we’re actually going to name the consumer service the name of the AI model, they’re admitting to themselves, this product is nothing but technology. There isn’t product in this to do on top of it.

It’s just like Gmail. Gmail was technology. It was fast search. It was lots of storage. It was used in the web. The product in this wasn’t particular the way that Instagram was all about the product.

Gemini the model, Gemini the chat bot says, we’re just exposing our amazing breakthrough technology to you all, and you get to interface directly with it. Anthropologically looking from afar, it feels like it’s that principle at work.

David: I totally agree. I think it’s actually a really important branding point and rallying point to Google and Google culture to do this.

Ben: All right, so this is all the stuff going on in Google 2023-ish in AI. Before we catch up to the present, I have a whole other branch of Alphabet that has been a real bright spot for AI. Can I go there? Can I take this off ramp, if you will?

David: Can you take the wheel, so to speak?

Ben: May I take the wheel? May I investigate another bet?

David: Please tell us the Waymo story.

Ben: Awesome. We got to rewind back all the way to 2004, the DARPA Grand Challenge, which was created as a way to spur research into autonomous ground robots for military use. Actually what it did for our purposes here today is create the seed talent for the entire self-driving car revolution 20 years later.

The competition itself is really cool. There is a 132 mile race course. Now, mind you, this is 2004 in the Mojave Desert that the cars have to race on. It is a dirt road. No humans are allowed to be in or interact with the cars. There are monitored 100% remotely, and the winner gets $1 million.

David: $1 million.

Ben: Which was a break from policy. Normally these are grants, not prize money. So this needs to be authorized by an act of Congress. The $1 million eventually felt comical. The second year they raised the pot to $2 million. It’s crazy thinking about what these researchers are worth today, that that was the prize for the whole thing.

The first year in 2004 went fine. There were some amazing tech demonstrations on these really tight budgets, but ultimately zero of the 100 registered teams finished the race.

But the next year in 2005 was the real special year. The progress that the entire industry made in those first 12 months from what they learned is totally insane.

Of the 23 finalists that were entering the competition, 22 of them made it past the spot where the furthest team the year before had made it. The amount that the field advanced in that one year is insane.

Not only that. Five of those teams actually finished all 132 miles. Two of them were from Carnegie Mellon, and one was from Stanford led by a name that all of you will now recognize, Sebastian Thrun.

David: Indeed.

Ben: This is Sebastian’s origin story before Google. Now, as we said, Sebastian was kind enough to help us with prep for this episode, but I actually learned most of this from watching a 20-year-old Nova documentary that is available on Amazon Prime video. Thanks to Bret Taylor for giving us the tip on where to find this documentary.

David: Yes, the hot research tip.

Ben: What was special about this Stanford team? Well one, there’s a huge problem with noisy data that comes out of all of these sensors. It’s in a car in the desert getting rocked around. It’s in the heat, it’s in the sun. So common wisdom, and what Carnegie Mellon did was to do as much as you possibly can on the hardware to mitigate that. So things like custom rigging, gimbals, and giant springs to stabilize the sensors. Carnegie Mellon would essentially buy a Hummer and rip it apart and rebuild it from the wheels up. We’re talking welding and real construction on a car.

The Stanford team did the exact opposite. They viewed any new piece of hardware as something that could fail. So in order to mitigate risks on race day, they used all commodity cameras and sensors that they just mounted on a nearly unmodified Volkswagen. They only innovated in software and they figured they would just come up with clever algorithms to help them clean up the messy data later. Very googly, right?

David: Very googly.

Ben: The second thing they did was an early use of machine learning to combine multiple sensors. They mounted laser hardware on the roof, just like what other teams were doing, and this is the way that you can measure texture and depth of what is right in front of you. And the data, it’s super precise, but you can’t drive very fast because you don’t really know much about what’s far away since it’s this fixed field of view, it’s very narrow. Essentially, you can’t answer that question of how fast can I drive or is there a turn coming up?

So on top of that, the way they solved it was they also mounted a regular video camera. That camera can see a pretty wide field of view just like the human eye, and it can see all the way to the horizon just like the human eye. And crucially, it could see color.

What it would do—this is really clever—they would use a machine learning algorithm in real time. In 2005, this computer is sitting in the middle of the car. They would overlay the data from the lasers on top onto the camera feed. From the lasers, you would know if the area right in front of the car was okay to drive or not. Then the algorithm would look up in the frames coming off the camera, overlaid what color that safe area was, and then extrapolate by looking further ahead at other parts of the video frame to see where that safe area extended to, so you could figure out your safe path through the desert.

David: That’s awesome.

Ben: It’s so awesome.

David: I’m imagining a Dell PC sitting in the middle of this car in 2005.

Ben: It’s not far off. In the email that we send out, we’ll share some photos of it. It could then drive faster with more confidence, and it knew when turns were coming up. Again, this is real time, onboard the camera, 2005, is wild on that tech.

Ultimately, both of these bets worked and the Stanford team won in super dramatic fashion. They actually passed one of the Carnegie Mellon teams autonomously through the desert. It’s this big dramatic moment in the documentary.

You would think, so then Sebastian goes to Google and builds Waymo. No. As we talked about earlier, he does join Google through that crazy, please don’t raise money from Benchmark and Sequoia and we’ll just hire you instead. But he goes and works on Street View and Project Ground Truth, and co-found Google X.

David, as you were alluding to earlier, this Project Chauffeur that would become Waymo is the first project inside Google X.

David: And I think the story is that Larry came to Sebastian and was like yo, that’s self-driving car stuff, do it. Sebastian was like, no, come on. That was a DARPA challenge. Larry was like, no, no, you should do it.

Ben: He’s like, no, no, that won’t be safe. There are people running around cities. I’m not just going to put multi ton killer robots on roads and go and potentially harm people. Larry finally comes to him and says, why? What is the technical reason that this is impossible? And Sebastian goes home, has a sleep on it. He comes in the next morning and he goes, I realize what it was. I’m just afraid.

David: Such a good moment.

Ben: So they start. He’s like, there’s not a technical reason. As long as we can take all the right precautions and hold a very high bar on safety, let’s get to work. Larry then goes, great. I’ll give you a benchmark so that way if you’re succeeding. He comes up with these 10 stretches of road in California that he thinks will be very difficult to drive. it’s about a thousand miles, and the team starts calling it the Larry 1000. It includes driving to Tahoe, Lombard Street in San Francisco, Highway 1 to Los Angeles, the Bay Bridge. This is the bogey.

David: If you can autonomously drive these stretches of road, pretty good indication that you can probably do anything.

Ben: So they start the project in 2009 within 18 months. This tiny team, I think they hired, it’s like a dozen people or something. They’ve driven thousands of miles autonomously, and they managed to succeed in the full Larry 1000 within 18 months.

David: Totally unreal how fast they did it. Then also totally unreal how long it takes after that to productize and create the Waymo that we know today.

Ben: It’s like the first 99% and then the second 99% that takes 10 years. Self-driving is one of these really tricky types of problems where it’s surprisingly easy to get started, even though it seems like it would be an impossible thing.

But then there are edge cases everywhere, weather, road conditions, other drivers, novel road layouts, night driving. It takes this massive amount of work for a production system to actually happen.

Then the question is, what business do we build? What is the product here? And there was what Sebastian wanted, which was highway assist, the lowest stakes, most realistic, let’s make a better cruise control.

There’s what Eric Schmidt wanted, which is crazy. He proposed, oh, let’s just go buy Tesla and that’ll be our starting place. Then we’ll just put all of our self-driving equipment on all the cars. David, do you know what it would’ve cost to buy Tesla at the time?

David: I think at the time that negotiations were taking place between Elon and Larry and Google, this was in the depths of the Model S production scaling woes. I think Google could have bought the company for $5 billion. That’s what I remember.

Ben: It was $3 billion.

David: $3 billion, oh my goodness.

Ben: Obviously that didn’t happen, but what a crazy alternative history that could have been.

David: I think if that had happened, DeepMind would not have gone down in the same way, and probably OpenAI would not have gotten founded.

Ben: That’s probably right.

David: I think that is obviously unprovable.

Ben: The counterfactuals that we always come up with on this show, you can’t know.

David: Seems more likely than not to me, that at a minimum OpenAI would not exist.

Ben: Then there was what Larry wanted to do, option three, build robo taxis. Ultimately, that is at least right now, what they would end up doing.

We could do a whole episode about this journey, but we will just hit some of the major points for the sake of time. The big thing to keep in mind here, neither Google nor the public really knew if self-driving was something that could happen in the next two years from any given point or take another 10.

Just to illustrate it, for the first five years of Project Chauffeur, it did not use deep learning at all. They did the Larry 1000 without any deep learning and then went another 3½ years.

David: Wow, that’s crazy. And yet totally illustrates, you never know how far away the end goal is.

Ben: And this is a field that comes from the only way progress happens is through these series of breakthroughs. You don’t know: (a) how far the next breakthrough is because at any given time, there are lots of promising things in the field, most of which don’t work out. Then (b) when there is a breakthrough, actually how much lift that will give you over existing methods. So anytime people are forecasting, oh, and AI, we’re going to be able to do X, Y, Z, and X years, it’s a complete fool’s errand. Even the experts don’t know.

Here are the big milestones. 2013, they started using convolutional neural nets, they could identify objects, they got much better perception capabilities. This 2013–2014 period is when Google found religion around deep learning. This is like right after the 40,000 GPUs rolled out. They’ve actually got some hardware to start doing this on now.

2016, they’ve seen enough technology proof that they think, let’s commercialize this. We can actually spin this out into a company. So Waymo becomes its own subsidiary inside of Alphabet. It’s no longer a part of Google X anymore.

2017, obviously the Transformer comes out. They incorporate some learnings from the Transformer, especially around prediction and planning.

March of 2020, they raise $3.2 billion from folks like Silver Lake, Canada Pension and Investment Board, Mubadala, Andreessen Horowitz, and of course the biggest check (I think) Alphabet. I think they’re always the biggest check because Alphabet is still the majority owner, even after a bunch more fundraisers.

In October of 2020, they launched the first public commercial, no human behind the driver’s seat thing in Phoenix. It’s the first in the world. This is 11 years after succeeding in the Larry 1000. And this is nuts. I had given up at this point. I was like, that’s cute that Waymo and all these other companies are trying to do self-driving, seems like it’s never going to happen. Then they actually were doing a large volume of rides safely with consumers and charging money for it in Phoenix.

David: Then they bring it to San Francisco, where for me and lots of people in San Francisco, it is a huge part of life in the city here now. It’s amazing.

Ben: Yeah. Every time I’m down, I love taking them. They’re launching in Seattle soon; I’m pumped. Interestingly, they don’t make the hardware, so they use a Jaguar vehicle, that from what I can tell, is only in Waymo’s. I don’t know if anybody else drives that Jaguar or if you can buy it, but they’re working on a van next. They have some next generation hardware.

For anyone who hasn’t taken it, it’s an Uber but with no driver. That launched in June of 2024. Along the way there, they raised their “Series B,” another $2.5 billion. Then after the San Francisco rollout, they raised their “Series C,” $5.6 billion. This year in January, they were reportedly doing more in gross bookings than Lyft in San Francisco.

David: Wow. I totally believe it. It is the number one option in San Francisco that I and everybody I know too always goes for a ride hailing. It’s like, try to get Waymo. If there’s not a Waymo available anytime soon, then go down the stack.

Ben: We’re living in the future and how quickly we fail to appreciate it.

David: And what’s cool (I think) for people who, it hasn’t come to their city and is not part of their lives yet, it’s not just that it’s a cool experience to not have a driver behind them. Pretty quickly that just fades. It’s actually a different experience.

If I need to go somewhere with my older daughter, I don’t mind hailing a Waymo, bringing the car seat, installing the car seat in the Waymo and driving with my daughter. And she loves it. We call it a robot car and she’s like a robot car. I’m so excited. I would never do that with an Uber.

Ben: That’s interesting.

David: To my dog. Whenever I need to go with my dog, it’s super awkward to hail an Uber and be like, hey I got my dog. Can the dog committed? Not a big deal with a Waymo. Then when you’re in town…

Ben: We can actually have sensitive conversations in the car.

David: You can have phone calls. It really is a different experience.

Ben: That’s so true. May as well catch up to today. They’re operating in five cities—Phoenix, San Francisco, LA Austin, and Atlanta. They have hundreds of thousands of paid rides every week. They’ve now driven over 100 million miles with no human behind the wheel, growing at 2 million every week. There are over 10 million paid rides across 2000 vehicles in the fleet.

They’re going to be opening a bunch more cities in the US next year. They’re launching in Tokyo, their first international city, slowly and then all at once. That’s the lesson here. The technology, they really continued with that multi-sensor approach all the way from the DARPA Grand Challenge—camera, lidar, they added radar, and actually they use audio sensing as well. Their approach is basically any data that we can gather is better because that makes it safer.

They have 13 cameras, 4 lidar, 6 radar, and the array of external microphones. This is obviously way more expensive of a solution than what Tesla is just doing with cameras. But Waymo’s party line is, they believe it is the only path to full autonomy to hit the safety bar and regulatory bar that they’re aiming for. It seems like a really big line in the sand for them anytime you talk to somebody in that organization.

David: And as a regular user of both products, happy owner and driver of a model Y in addition to a regular Waymo user, at least with the current instantiation of full self-driving on my Tesla, vastly different products.

Full self-driving on my Model Y is great. I use it all the time on the freeway, but I would never not pay attention. Whereas every time I get in a Waymo, it’s almost like Google search. I just trust that oh, this is going to be completely and totally safe. I’m sitting in the backseat and I can totally tune out.

Ben: I think I trust my Model Y FSD more than you do, but I get what you’re saying. And frankly, regulatorily, you are required to still pay attention in Tesla and not in the Waymo.

The safety thing is super real, though. If you look at the numbers, over a million motor vehicle crashes cause fatalities every year, or there are over a million fatalities. In the US alone, over 40,000 deaths occur per year. If you break that down, that’s 120 every day. That’s a giant cause of death.

The study that Waymo just released last month showed that they have 91% fewer crashes with serious injuries or worse compared to the average human driver, even controlled for the fact that Waymos right now are only driving on city surface streets. They controlled it apples to apples with human driving data, and it’s a 91% reduction in those serious either fatality or serious injury things. Why aren’t we all talking about this all the time every day? This is going to completely change the world in a giant cause of death.

So while we’re in Waymo land, what do you think about doing some quick analysis? Because I’ve been scratching my head here of what is this business. And then I promise we’ll go back to the rest of Google AI and catch up to today.

It is super expensive to operate, especially at early scale. The training is high, the inference is high, the hardware is high, et cetera, et cetera, et cetera.

David: Also, the operations are expensive.

Ben: In fact, they’re experimenting. Some cities they actually outsource the operations. There’s a rental car company in Texas that manages it. Or they’ve partnered (I believe) with Lyft and Uber. They’re trying all sorts of O&O versus partnership models to operate it.

David: And the operations are like, these are electric cars. They need to be charged, they need to be cleaned, they need to be returned to depots, they need to be checked out, they need to have sensors replaced.

Ben: So the question is what is the potential market opportunity? How big could this business be? And there are a few different ways you could try to quantify it.

One total market size thing you could do is try to sum the entire automaker market cap today. That would be $2.5 trillion globally if you include Tesla or $1.3 trillion without. But Waymo’s not really making cars. That’s probably the wrong way to slice it.

You could look at all the ride sharing companies today, which might be a better comp because that’s the business that Waymo is actually in today. That’s on the order of $300 billion, most of which is Uber. That’s addressable market cap today with ride sharing.

Waymo’s ambitions though are bigger than that. They want to be in the cars that you own. They want to be in long haul trucking. They believe they can grow the share of transportation because there are blind people that could own a car. There are elderly people who could get where they need to go on their own without having a driver, that sort of thing.

The most squishy but I think the most interesting way to look at it, is what is the value from all of the reduction in accidents? Because that’s really what they’re doing. It’s a product to replace accidents with non accidents.

David: I think that’s viable. But again, I would say as a regular user of the product, it is a different and expanding product to human ride share.

Ben: So your argument is whatever number I come up with for reducing accidents, it’s still a bigger market than that because there’s additional value created in the product experience itself.

David: Yeah. Scoping just to ride share now that we have Waymo in San Francisco, I use Waymo in scenarios where I would never use an Uber or a Lyft.

Ben: Makes sense. Here’s the data we have. The CDC released a report saying deaths from crashes in 2022 in the US resulted in $470 billion in total costs, including medical costs and the cost estimates for lives lost. Which is crazy that the CDC has some way of putting the costs on human life, but they do.

If you reduce crashes 10x, which is what Waymo seems to be saying in their data, at least for the serious crashes, that’s over $420 billion a year in total costs that we would save as a nation.

Now it’s not totally apples to apples—I recognize this—but that cost savings is more than Google does today in revenue in their entire business. You could see a path to a Google-sized opportunity for Waymo as a standalone company just through this analysis as long as they figure out a way to get costs down to the point where they can run this as a large and profitable business.

David: It is an incredible 20+ year success story within Google.

Ben: The way I want to close it is the investment so far actually hasn’t been that large when you consider this opportunity. They have burned somewhere in the neighborhood of $10–$15 billion. That’s why I was listing all the investments to get to this point.

David: Chump change compared to foundational models.

Ben: Dude, also, let’s just keep it scoped in this sector. That’s one year of Uber’s profits.

David: Wow. Seems like a good bet.

Ben: I used to think this was some wild goose chase. It now looks really, really smart.

David: Totally agree.

Ben: Also, that cost $10–$15 billion is the profits that Google made last month.

David: Well, speaking of Google, should we catch us up to today with Google AI?

Ben: Yes. I think where you were is the Gemini launch.

David: So Sundar makes these two decrees mid-2023: (1) We’re merging brain and DeepMind into one team for AI within Google. (2) We’re going to standardize on one model, the future Gemini and DeepMind/Brain team. You go build it, and then everybody in Google you’re going to use it.

Ben: Not to mention, apparently Sergey Brin is now back as an employee working on Gemini.

David: Employee number…

Ben: Got his Newton badge back.

David: Yeah, got his badge back. Once Sundar makes these decisions, Jeff Dean and Oriol Vinyals from Brain go over and team up with the DeepMind team and they start working on Gemini.

Ben: I’m a believer now, by the way. You got Jeff Dean working on it. I’m in.

David: If you got Jeff Dean on it, it’s probably going to work. If you weren’t a believer yet, wait until I’m going to tell you next. Once they get Noam back, when they do the deal with Character AI, bring him back into the fold, Noam joins the Gemini team and Jeff and Noam are the two co-technical leads for Gemini now. So…

Ben: Let’s go.

David: …let’s go. They actually announce this very quickly at the Google IO keynote in May 2023. They announced Gemini, they announced the plans, they also launch AI overviews in search, first as a Labs product and then later that becomes just standard for everybody using Google Search.

Ben: Which is crazy, by the way. The number of Google searches that happen is unfathomably large. I’m sure there’s a number for it, but just think about that’s about the highest level of computing scale that exists, other than high bandwidth things like streaming. But just think about the instances of Google searches that happen. They are running an LLM inference on all of those, or at least as many as they’re willing to show AI overviews on, which I’m sure is not every query but many.

David: A subset, but still a large, large number of Google search. I see them all the time. This is really Google immediately deciding to operate at AI speed. ChatGPT happened in November 30th, 2022. We’re now in May 2023. All of these decisions have been made, all of these changes have happened, and they’re announcing things at IO.

Ben: And they’re really flexing the infrastructure that they’ve got. The fact that they can go like, oh yeah, sure, let’s do inference on every query or Google, we can handle it.

David: So a key part of this new Gemini model that they announced in May 2023 is it’s going to be multimodal. Again, this is one model for everything—text, images, video, audio, one model. They release it for early public access in December 2023. Still, also crazy. Six months, they build it, they trade it, they release it.

Ben: That is amazing.

David: Wild. February 2024, they launched Gemini 1.5 with a 1 million token context window. Much, much larger context window than any other model on the market,

Ben: Which enables all sorts of new use cases. There are all these people who were like, oh, I tried to use AI before but it couldn’t handle my X, Y, Z use case. Now they can.

David: The next year, February 2025, they released Gemini 2.0. March of 2025, one month later they launched Gemini 2.5 Pro in experimental mode and then that goes GA in June.

Ben: This is like NVIDIA pace how often they’re shipping.

David: Yeah, seriously. Also in March of 2025, they launch AI mode. You can now switch over on google.com to chatbot mode.

Ben: And they’re split testing auto opting some people into AI mode to see what the response is. This is the golden goose.

David: The elephant is tap dancing here. Then there are all the other AI products that they launched. NotebookLM comes out during this period, AI generated podcasts.

Ben: Which, does that sound like us to you? It feels a little trained.

David: The number of texts that we got when that came out of, this must be trained on Acquired. I do know that a bunch of folks on the NotebookLM team are Acquired fans. I don’t know if they trained on us. Then there’s the video, the image stuff, Veo 3, Nano Banana, Genie 3 that just came out recently. Genie, this is insane. This is a world builder based on prompts and videos.

Ben: You haven’t actually used it yet, right? You watched that hype video?

David: Yeah, I watched the video. I haven’t actually used it.

Ben: If it does that, that’s unbelievable. It’s a real time generative…

David: World builder.

Ben: …world builder, yeah. You look right and it invents stuff to your right. You combine that with a Vision Pro hardware, you’re just living in a fantasy land.

David: They announced there are now 450 million monthly users of Gemini. Now that includes everybody who’s accessing Nano Banana.

Ben: I can’t believe this stat. This is insane. Even with recently being number one in the app store, it still feels hard to believe Google’s saying it, so it must be true. But I just wonder what are they counting as use cases of the Gemini app?

David: Certainly everybody who’s using Nano Banana is using Gemini.

Ben: But is it counting AI overviews or is it counting AI mode or is it counting something where I accidentally, like Meta said, that crazy high number of people using Meta AI? That was complete garbage. That was people searching Instagram who accidentally hit a Llama model that made some things happen and they were like, ugh, go away. I actually am just looking for a user. Is it really 450 million or is it 450 million?

David: Good question. Either way, going from zero is crazy impressive in the amount of time that they have done.

Ben: Especially given revenue’s at an all time high, they seem to so far—at least in this squishy early phase—be able to figure out how to keep the core business going while doing well as a competitor in the cutting-edge of AI.

David: And to foreshadow a little bit to we’re going to do a bull and bear here in a minute, as we talked about in our Alphabet episode, Google does have a history of navigating platform shifts incredibly well in the transition of mobile.

Ben: It’s true.

David: Definitely a rockier start here in the AI platform shift, much rockier, but hey look, if you were to lay out a recipe for how to respond given the rocky start, it’d be hard to come up with a much better slate of things than what they’ve done over the last two years.

Ben: All right, should I give us the snapshot of the business today?

David: Give us the snapshot of the business today. Oh yeah. Also, by the way, the federal government decided they were a monopoly and then decided not to do anything about it because of AI.

Ben: Between the time when we shipped our Alphabet episode and here with our Google AI episode, or our Part II and Part III for those who prefer simpler naming schemes, there was a US versus Google antitrust case. The judge first ruled that Google was a monopoly in internet search, and then did not come up with any material remedies. There are some, but I would call them immaterial. They did not need to spin off Chrome and they did not need to stop sending tens of billions of dollars to Apple and others.

In other words, yes, Google’s a monopoly and the cost of doing anything about that would have too many downstream consequences on the ecosystem. So we’re just going to let them keep doing what they’re doing. One of the reasons that the judge cited of why they weren’t going to really take these actions is because of the race in AI. That because tens of billions of dollars of funding have gone into companies like OpenAI and Anthropic and Perplexity, Google essentially has this new war to fight and we’re going to leave it to the free market to do its thing where it creates viable competition on its own, and we’re not going to hamstring Google.

Personally, I think this argument is a little bit silly. None of these AI companies are generating net income, and just because they’ve raised a huge amount of money, it doesn’t mean that will last forever. They’ll all burn through their existing cash in a pretty short period of time. If the spigots ever dry up, Google doesn’t have any self-sustaining competition right now, whether in their old search business or in AI. It is all dependent on people believing that the opportunity is so large that they keep pouring tens of billions of dollars into these competitors.

David: Plenty of other folks have made the glib comment, but there’s merit to it of, hey, as flatfooted as Google was when ChatGPT happened, if the outcome of this is they avoid a Microsoft-level distraction and damage to their business from a US federal court monopoly judgment, worth it.

Ben: Well, there’s a funny meme here that you could draw. You know that meme of someone pushing the domino and it knocking over some big wall later?

David: Yeah.

Ben: There’s the domino of Ilya leaving Google to start OpenAI and the downstream effect is Google is not broken up.

David: Exactly.

Ben: It actually saves Google.

David: It actually saves Google.

Ben: It’s totally wild.

David: Totally wild.

Ben: All right, here’s the business today. Over the last 12 months, Google has generated $370 billion in revenue. On the earnings side, they’ve generated 140 billion over the last 12 months, which is more profit than any other tech company. The only company in the world with more earnings is Saudi Aramco. Let’s not forget Google is the best business ever.

David: And we also made the point at the end of the Alphabet episode, even in the midst of all of this AI era and everything that’s happened over the last 10 years, the last five years, Google’s core business has continued to grow 5x since the end of our Alphabet episode in 2015–2016.

Ben: Market cap, Google’s surge passed their old peak of $2 trillion and just hit that $3 trillion mark earlier this month. They’re the fourth most valuable company in the world behind NVIDIA, Microsoft, and Apple. It’s just crazy.

On their balance sheet—actually, I think this is pretty interesting; I normally don’t look at balance sheets as a part of this exercise, but it’s useful and here’s why in this case—they have $95 billion in cash and marketable securities. I was about to stop there and make the point, wow. Look at how much cash and resources they have.

David: I’m actually surprised it’s not more.

Ben: It used to be $140 billion in 2021, and over the last four years they’ve massively shift from this mode of accumulating cash to deploying cash. A huge part of that has been the CapEx of the AI data center build out. They’re very much playing offense in the way that meta Microsoft and Amazon are in deploying that CapEx.

But the thing that I can’t quite figure out is the largest part of that was actually buybacks, and they started paying a dividend. If you’re not a finance person, the way to read into that is yes, we still need a lot of cash for investing in the future of AI and data centers, but we still actually had way more cash than we needed and we decided to distribute that to shareholders. That’s crazy.

David: Best business of all time, right?

Ben: That illustrates what a crazy business their core search ads business is. If they’re saying the most capital-intense race in business history is happening right now, we intend to win it. And we have tons of extra cash lying around on top of what we think, plus a safety cushion for investing in that CapEx race.

David: Yeah. Wow.

Ben: So there are two businesses that are worth looking at here: (1) Gemini to try to figure out what’s happening there, and (2) is a brief history of Google Cloud. I want to tell you the cloud numbers today, but it’s probably worth actually understanding how did we get here on cloud.

First on Gemini. Because this is Google and they have (I think) the most obfuscated financials of any of the companies we’ve studied, they anger me the most in being able to hide the ball in their financial statements. Of course, we don’t know Gemini-specific revenue.

What we do know is there are over 150 million paying subscribers to the Google One bundle. Most of that is on a very low tier. It’s on the $5–$10 a month. The AI stuff kicks in on the $20 a month tier where you get the premium AI features, but I think that’s a very small fraction of the 150 million today.

David: I think that’s what I’m on.

Ben: But two things to note: (1) It’s growing quickly that 150 million is growing almost 50% year over year. But (2) is Google has a subscription bundle that 150 million people are subscribed to. I’ve had in my head that AI doesn’t have a future as a business model that people pay money for. That it has to be ad-supported like Search.

David: But hey, that’s not nothing. That’s like—

Ben: That’s almost half of America.

David: How many subscribers does Netflix have?

Ben: Netflix is in the hundreds of millions. Spotify is now a quarter billion, something like that. We now live in a world where there are real scaled consumer subscription services.

I owe this insight to Shishir Mehrotra. We chatted actually last night because I name dropped him in the last episode and then he heard it, so reached out and we talked, and that’s made me do a 180. I used to think if you’re going to charge for something, your total addressable market shrunk by 90%–99%. But he has this point that if you build a really compelling bundle and Google has the digital assets to build a compelling bundle…

David: Oh my goodness. Youtube Premium, NFL Sunday ticket.

Ben: Yes, stuff in the Play Store, YouTube Music, all the Google One storage stuff, they could put AI in that bundle and figure out through clever bundle economics a way to make a paid AI product that actually reaches a huge number of paying subscribers.

David: Totally.

Ben: So we really can’t figure out how much money Gemini makes right now. Probably not profitable anyway, so what’s the point of even analyzing it?

David: But okay, tell us the Cloud story. We intentionally did not include Cloud in our Alphabet episode.

Ben: Google Part II, effectively.

David: Google Part II, yes, because it is a new product and now very successful one within Google that was started during the same time period as all the other ones that we talked about during Google Part II. But it’s so strategic for AI.

Ben: It is a lot more strategic now in hindsight than it looked when they launched it. Just quick background on it. It started as Google App Engine. It was a way in 2008 for people to quickly spin up a backend for a web or soon after a mobile app. It was a Platform as a Service, so you had to do things in this very narrow googly way.

It was very opinionated. You had to use this SDK, you had to write it in Python or Java, you had to deploy exactly the way they wanted you to deploy. It was not a thing where they would say, hey developer, you can do anything you want. Just use our infrastructure. It was opinionated. Super different than what AWS was doing at the time.

What they’re still doing today, which the whole world eventually realized was right, which is cloud should be Infrastructure as a Service. Even Microsoft pivoted Azure to this reasonably quickly, where it was like, you want some storage? We got storage for you. You want a VM? We got a VM for you. You want some compute? You want a database?

David: We got you.

Ben: Fundamental building blocks. Eventually, Google launches their own infrastructure as a service in 2012, it took four years. They launched Google Compute Engine that they would later rebrand Google Cloud platform. That’s the name of the business today.

The knock on Google is that they could never figure out how to possibly interface with the enterprise. Their core business, they made really great products for people to use that they loved polishing. They made them all as self-service as possible. Then the way they made money was from advertisers. And let’s be honest, there’s no other choice but to use Google Search.

David: It didn’t necessarily need to have a great enterprise experience for their advertising customers because they were going to come anyway.

Ben: So they’ve got this self-serve experience. Meanwhile, the cloud is a knife fight. These are commodities

David: All about the enterprise.

Ben: It’s the lowest possible price. It’s all about enterprise relationships and clever ways to bundle and being able to deliver a full solution.

David: You say solution, I hear gross margin. But yes, Google out of their natural habitat in this domain.

Ben: And early on they didn’t want to give away any crown jewels. They viewed their infrastructure as this is our secret thing. We don’t want to let anybody else use it. And the best software tools that we have on it that we’ve written for ourselves like Bigtable or Borg, how we run Google or Distbelief, these are not services that we’re making available on Google Cloud.

David: These are competitive advantages. Then they hired the former president of Oracle, Thomas Kurian.

Ben: And everything changed. 2017, two years before he comes in, they had $4 billion in revenue, 10 years into running this business. 2018 is their first very clever strategic decision. They launched Kubernetes. The big insight here is if we make it more portable for developers to move their applications to other clouds, the world is wanting multi-cloud here.

David: We’re the third place player, we don’t have anything to lose, so we can offer this tool, counter position against AWS and Azure.

Ben: We shift the developer paradigm to use these containers they orchestrate on our platform and then we have a great service to manage it for you. It was very smart. This becomes one of the pillars of their strategy is you want multi-cloud? We’re going to make that easy, and sure, you can choose AWS or Azure too. It’s going to be great.

David, as you said, the former president of Oracle, Thomas Kurian, is hired in late 2018. You couldn’t ask for a better person who understands the needs of the enterprise than the former president of Oracle. This shows up in revenue growth right away.

In 2020, they crossed $13 billion in revenue, which was nearly tripling in three years. They hired like 10,000 people into the go-to-market organization. I’m not exaggerating that. That’s on a base of 150 people when he came in, most of which were seated in California, not regionally distributed throughout the world. The funniest thing is Google was a cloud company all along. They had the best engineers building this amazing infrastructure.

David: They had the products, they had the infrastructure, they just didn’t have the go-to market organization.

Ben: And the productization was all googly. It was for us, for engineers. They didn’t really build things that lead enterprises build the way that they wanted to build. This all changes. 2022, they hit $26 billion in revenue. 2023, they’re a real viable third cloud. They also flipped to profitability in 2023. Today, they’re over $50 billion in annual revenue run rate. It’s growing 30% year over year. They’re the fastest growing of the major cloud providers, 5x in five years.

It’s really three things: (1) It’s finding religion on how to actually serve the enterprise, (2) it’s leaning into this multi-cloud strategy and actually giving enterprise developers what they want, and (3) AI has been such a good tailwind for all hyperscalers, because these workloads all need to run in the cloud, because it’s giant amounts of data and giant amount of compute and energy. But in Google Cloud you can use TPUs, which they make a ton of, and everyone else is desperately begging NVIDIA for allocations to GPUs. If you’re willing to not use Cuda and build on Google Stack, they have an abundant amount of TPUs for you.

David: This is why we saved Cloud for this episode. There are two aspects of Google Cloud that I don’t think they foresaw back when they started the business with App Engine, but are hugely strategically important to Google today.

One is just simply that cloud is the distribution mechanism for AI. If you want to play an AI today, you either need to have a great application, a great model, a great ship, or a great cloud. Google is trying to have all four of those. There is no other company that has (I think) more than one.

Ben: I think that’s the right call. Think about the big AI players. NVIDIA has a cloud, but not really. They just have chips. Made the best chips and the chips everyone wants, but chips.

Then you just look around the rest of the big tech companies. Meta right now, only an application. They’re completely out of the race for the frontier models at the moment. We’ll see what they’re hiring spree yields. You look at Amazon. Infrastructure, they have application. Maybe. I don’t actually know if amazon.com. I’m sure it benefits from LLMs in a bunch of ways.

David: Mainly, it’s cloud.

Ben: And cloud leader. Microsoft, it’s just cloud. They make some models, but…

David: They’ve got applications, but yeah, cloud.

Ben: Apple?

David: Nothing.

Ben: Nothing. AMD, just chips.

David: OpenAI, model. Anthropic, model.

Ben: These companies don’t have their own data centers. They’re making noise about making their own chips, but not really, and certainly not at scale. Google has scale data center, scale chips, scale usages of model. Even just from google.com, queries now on AI overviews.

David: And scale applications. They have all of the pillars of AI, and I don’t think any other company has more than one.

Ben: And they have the very most net income dollars to lose.

David: So then there’s the chip side specifically of this. If Google didn’t have a cloud, it wouldn’t have a chip business. It would only have an internal chip business. The only way that external companies—users, developers, model researchers—could use TPUs would be if Google had a cloud to deliver them. Because there’s no way in hell that Amazon or Microsoft are going to put TPUs from Google in their clouds.

Ben: We’ll see.

David: We’ll see, I guess.

Ben: I think within a year it might happen. There are rumors already that some neoclouds in the coming months are going to have TPUs.

David: Interesting.

Ben: Nothing announced, but TPUs are likely going to be available in neocloud soon, which is an interesting thing. Why would Google do that? Are they trying to build an NVIDIA-type business where they make money selling chips? I don’t think so. I think it’s more that they’re trying to build an ecosystem around their chips the way that Cuda does, and you’re only going to credibly be able to do that if your chips are accessible and anywhere that someone’s running their existing workloads.

David: It’d be very interesting if it happens. And look, you may be right. Maybe there will be TPUs in AWS or Azure someday, but I don’t think they would’ve been able to start there. If Google didn’t have a cloud and there weren’t any way for developers to use TPUs and start wanting TPUs, would Amazon or Microsoft be like, eh, you know? All right, Google. We’ll take some of your TPUs, even though no developer out there uses some.

David: All right, well with that, let’s move into analysis. I think we need to do bull and bear on this one.

Ben: We have to this time.

David: Got to bring that back.

Ben: For these episodes in the present, it seems like we need to paint the possible futures.

David: Bringing back bull and bear. I love it. Then we’ll do playbook, powers, quintessence. Bring it home.

Ben: Perfect. All right, so here’s my set of bull cases. Google has distribution to basically all humans as the front door to the Internet. They can funnel that however they want. You’ve seen it with AI overviews, you’ve seen it with AI mode. Even though lots of people use ChatGPT for lots of things, Google’s traffic (I assume) is still essentially an all time high and it’s a default behavior.

David: Yup, powerful.

Ben: That is a bet on implementation that Google figures out how to execute and build a great business out of AI, but it is still theirs to lose.

David: And they’ve got a viable product. It’s not clear to me that Gemini is any worse than OpenAI or Anthropics products.

Ben: No, I completely agree. This is a value creation, value capture thing. The value creation is there in spades. The value capture mechanism is still TBD. Google’s old value capture mechanism is one of the best in history. That’s the issue at hand. Let’s not get confused that it’s not a good experience. It’s a great experience.

We’ve talked about the fact that Google has all the capabilities to win in AI and it’s not even close. Foundational model, chips, hyperscaler, all this with self-sustaining funding. That’s the other crazy thing is you look at the clouds have self-sustaining funding, NVIDIA has self-sustaining funding, none of the model makers have self-sustaining funding, so they’re all dependent on external capital.

David: Google is the only model maker who has self-sustaining funding.

Ben: Isn’t that crazy? Basically all the other large scale usage foundational model companies are effectively startups. And Google’s is funded by a money funnel so large that they’re giving extra dollars back to shareholders for fun. Again, we’re in the bull case.

David: Well, when you put it that way, yeah.

Ben: A thing we didn’t mention, Google has incredibly fat pipes connecting all of their data centers. After the dot-com crash in 2000, Google bought all that dark fiber for pennies on the dollar, and they’ve been activating it over the last decade. They now have their own private backhaul network between data centers. No one has infrastructure like this. Not to mention that serves YouTube. They’re fat pipes.

David: Which in and of itself is its own bull case for Google in the future.

Ben: That’s a great point.

David: Ben Thompson had a big article about this yesterday at the time of recording.

Ben: That was a mega bowl case that Ben Thompson published this week that it was an interesting point. A text-based Internet is the old Internet. It’s the first instantiation of the Internet because we didn’t have much bandwidth. The user experience that is actually compelling is…

David: Video.

Ben: …high resolution video everywhere all the time.

David: We already live in the YouTube internet.

Ben: And not only can they train models on really the only scale source of UGC media across long form and short form, but they also have that as the number two search engine, this massive destination site, so they previewed things like you’ll be able to buy AI-labeled or AI-determined things that show up in videos.

And if they wanted to, they could just go label every single product in every single video and make it all instantly shoppable. It doesn’t require any human work to do it. They could just do it, and then run their standard ads model on it. That was a mind expanding piece that Ben published yesterday. Or I guess if you’re listening to this a few weeks ago about that.

David: And then there are also all the video AI applications that they’ve been building, like Flow and Veo. What is that going to do for generating videos for YouTube? That will increase engagement and ad dollars for YouTube. Going to work real well.

Ben: Yup. They still have an insane talent bench, even though they’ve bled talent here and there and lost people. They have also shown they’re willing to spend billions for the right people and retain them.

Unit economics. Let’s talk about unit economics of chips. Everyone is paying NVIDIA 75%–80% gross margins, implying something like a 4x or 5x markup on what it costs to make the chips.

A lot of people refer to this as the Jensen tax or the NVIDIA tax. You can call it that. You can call it good business. You can call it pricing power. You can call it scarcity of supply, whatever you want, but that is true. Anyone who doesn’t make their own chips is paying a giant, giant premium to NVIDIA.

Google has to still pay some margin to their chip hardware partner, Broadcom, that handles a lot of the work to actually make the chip, interface with TSMC. I have heard that Broadcom has something like a 50% margin when working with Google on the TPU versus NVIDIA’s 80%. But that’s still a huge difference to play with. A 50% gross margin from your supplier or an 80% gross margin from your supplier is the difference between a 2x markup and a 5x markup.

David: I guess that’s right.

Ben: When you frame it that way, it’s actually a giant difference of the impact to your cost. You might wonder appropriately, well are chips actually the big part of the total cost of ownership of running one of these data centers or training one of these models? Chips are the main driver of the cost. They depreciate very quickly. This is at best a five-year depreciation because of how fast we are pushing the limits of what we can do with chips, the needs of next generation models, how fast TSMC is able to produce.

David: Even that is ambitious. If you think you’re going to get five years of depreciation on AI chips five years ago, we were still two years away from ChatGPT.

Ben: Or think about what Jensen said at, we were at GTC this year. He was talking about Blackwell and he said something about Hopper. He was like, eh, you don’t want Hopper. My sales guys are going to hate me, but you really don’t want Hopper at this point. These were the H100s. This was the hot chip just when we were doing our most recent NVIDIA episode.

David: Things move quickly.

Ben: I’ve seen estimates that over half the cost of running an AI data center is the chips and the associated depreciation. The human cost, that R&D is actually a pretty high amount because hiring these AI researchers and all the software engineering is meaningful. Call it 25%–33%. The power is actually a very small part. It’s 2%–6%.

When you’re thinking about the economics of doing what Google’s doing, it’s actually incredibly sensitive to how much margin you are paying your supplier in the chips, because it’s the biggest cost driver of the whole thing.

I was sanity checking some of this with Gavin Baker, who’s the partner at Atreides Management to prep for this episode. He’s a great public equities investor who’s studied the space for a long time. We actually interviewed him at the NVIDIA GTC pregame show.

He pointed out normally, like in historical technology eras, it hasn’t been that important to be the low cost producer. Google didn’t win because they were the lowest cost surge engine. Apple didn’t win because they were the lowest cost. That’s not what makes people win.

But this era might actually be different because these AI companies don’t have 80% margins the way that we’re used to in the technology business, or at least in the software business. At best, these AI companies look like 50% gross margins.

Google being definitively the low-cost provider of tokens because they operate all their own infrastructure and because they have access to low markup hardware, it actually makes a giant difference and might mean that they are the winner in producing tokens for the world.

David: Very compelling bull case there.

Ben: That’s a weirdly winding analytical bull case, but if you want to really get down to it, they produce tokens.

David: I’ve got one more bullet point to add to the bull case for Google here. Everything that we talked about in Part II (the Alphabet episode), all of the other products within Google—Gmail, Maps, Docs, Chrome, Android—that is all personalized data about you that Google owns, that they can use to create personalized AI products for you that nobody else has.

Ben: Another great point. Really the question to close out the bull case is, is AI a good business to be in compared to Search? Search is a great business to be in. So far AI is not, but in the abstract—again, we’re in the bull case, so I’ll give you this—it should be. With traditional web search, you type in 2–3 words. That’s the average query length.

I was talking to Bill Gross, and he pointed out that in AI chat you’re often typing 20+ words, so there should be an ad model that emerges and ad rates should actually be dramatically higher because you have perfect precision.

David: You have even more intent.

Ben: You know the crap out of what that user wants, so you can really decide to target them with the ad or not. AI should be very good at targeting with the ad. So it’s all about figuring out the user interface, the mix of paid versus not, exactly what this ad model is. But in theory, even though we don’t really know what the product looks like now, it should actually lend itself very well to monetization.

Since AI is such an amazing transformative experience, all these interactions that were happening in the real world or weren’t happening at all, like the answers to questions and being on a time spent, is now happening in these AI chats. So it seems like the pie is actually bigger for digital interactions than it was in the search era. Again, monetization should increase because the pie increases there.

Then you’ve got the bull case of Waymo could be its own Google-size business.

David: I was just thinking of that. That’s scoping all of this to a replacement to the search market. Waymo and potentially other applications of AI beyond the traditional search market could add to that.

Ben: Then there’s the galaxy Brain bull case, which is if Google actually creates AGI, none of this even matters anymore. Of course, it’s the most valuable thing.

David: That feels out of the scope for an Acquired episode.

Ben: It’s disconnected. Yes, agree.

Bear case. So far, this is all fun to talk about, but then the product shape of AI has not lent itself well to ads. Despite more value creation, there’s way less value capture. Google makes something like $400-ish per user per year just based on some napkin math in the US. That’s a free service that everyone uses and they make $400-ish a year. Who’s going to pay $400 a year for access to AI? It’s a very thin slice of the population.

David: Some people certainly will, but not every person in America.

Ben: Some people will pay $10 million. If you’re only looking at the game on the field today, I don’t see the immediate path to value capture. Think about when Google launched in 1998. It was only two years before they had AdWords. They figured out an amazing value capture mechanism instantly.

David: Very quickly. Another bear case, think back to Google launch in 1998. It was immediately obviously the superior product. Definitely not the case today.

Ben: No. There are four, five great products.

David: Google’s dedicated AI offerings in chatbot was initially the immediately obviously inferior product, and now it’s arguably on par with several others.

Ben: They own 90% of the search market. I don’t know what they own of the AI market, but it’s not 90%. Is it 25%? I don’t know, but at steady state it probably will be something like 25%, maybe up to 50%. But this is going to be a market with several big players in it. So even if they monetized each user as great as they monetize it in Search, they’re just going to own way less of them.

David: Or at least it certainly seems that way right now.

Ben: AI might take away the majority of the use cases of Search, and even if it doesn’t I bet it takes away a lot of the highest value ones. If I’m planning a trip, I’m planning that in AI. I’m no longer searching on Google for things that are going to land Expedia ads in my face.

David: Or health. Another huge vertical.

Ben: Hey, I think I might have something that reminds me of mesothelioma. Is it that or not? Oh, where are you going to put the lawyer ads? Maybe you put them there. Maybe it’s just an ad product thing. But these are very high value…

David: Queries.

Ben: …former searches that those feel like some of the first things that are getting siphoned off to AI. Any other bear cases?

David: I think the only other bear case I would add is that they have the added challenge now of being the incumbent this time around, and people and the ecosystem isn’t necessarily rooting for them in the way that people were rooting for Google when they were a startup, and in the way that people were still rooting for Google in the mobile transition.

I think the startups have more of the hearts and minds these days. So I don’t think that’s quantifiable, but is just going to make it all a little harder path to row this time around.

Ben: You’re right. They had this incredible PR and public love tailwind the first time around.

David: And part of that’s systemic too. Like all of tech and all of big tech is just generally more out of favor with the country and the world now than it was 10 or 15 years ago.

Ben: It’s more important. It’s just big infrastructure. It’s not underdogs anymore.

David: And that affects the OpenAIs, the Anthropics, and the startups, too, but I think to a lesser degree.

Ben: They had to start behaving like big tech companies really early in their life compared to Google. Google gave a Playboy interview during their quiet period of their IPO. Times have changed.

David: Well, given all the drama at OpenAI, I don’t know that I’d characterize them as acting like a mature company.

Ben: Fair.

David: Company entity, whatever they are. But point taken.

Ben: Well, I worked most of my playbook into the story itself. You want to do power?

David: Great. Let’s move on and do power. Hamilton Helm’s seven powers analysis of Google here in the AI era. The seven powers are scale economies, network economies, counter positioning, switching costs, branding, quartered resource, and process power.

Ben: And the question is, which of these enables a business to achieve persistent differential returns? What entitles them to make greater profits than their nearest competitor sustainably? Normally we would do this on the business all up. I think for this episode we should try to scope it to AI products.

David: Agreed.

Ben: Usage of Gemini AI mode and AI overviews versus the competitive set of Anthropic, OpenAI, Perplexity, Grok, Meta AI.

David: Et cetera. Scale economies, for sure. Even more so in AI than traditional in tech.

Ben: They’re just way better. Look, they’re amortizing the cost of model training across every Google search. I’m sure it’s some super distilled down model that’s actually happening for AI overviews. But think about how many inference tokens are generated for the other model companies and how many inference tokens are generated by Gemini. They just are amortizing that fixed training cost over a giant, giant amount of inference, that I saw some crazy chart. We’ll send it out to email subscribers.

In April of 2024, Google was processing 10 trillion tokens across all their surfaces. In April of 2025, that was almost 500 trillion. That’s a 50x increase in one year of the number of tokens that they’re vending out across Google services through inference.

Between April of 2025 and June 2025, it went from a little under 500 trillion to a little under one quadrillion tokens. Technically 980 trillion, but they are now—because it’s later in the summer—definitely sending out maybe even multiple quadrillion tokens.

David: Wow.

Ben: So among all the other obvious scale economies things of amortizing all the costs of their hardware, they are amortizing the cost of training runs over a massive amount of value creation.

David: Scale economies must be the biggest one.

Ben: I find switching costs to be relatively low. I use Gemini for some stuff, then it’s really easy to switch away. That probably stops being the case when it’s personal AI, to the point that you’re talking about integrating with your calendar and your mail and all that stuff.

David: The switching costs have not really come out yet in AI products, although I expect they will.

Ben: They have within the enterprise for sure. Network economies? I don’t think if anyone else is a Gemini user, it makes it better for me because they are sucking up the whole internet whether anyone’s participating or not.

David: Agree. I’m sure AI companies will develop network economies over time. I can think of ways it could work, but yeah, right now, no. Arguably for the foundational model companies, can’t think of obvious reasons right now.

Ben: Where does Hamilton put distribution? Because that’s a thing that they have right now that no one else has. Despite ChatGPT having the Kleenex brand, Google distribution is still unbelievable. I don’t know, is that a cornered resource?

David: Cornered resource I guess? Yeah.

Ben: Definitely of that.

David: Google Search is a cornered resource for sure.

Ben: Certainly don’t have counter positioning. They’re getting counter positioned. I don’t think they have process power, unless they were coming up with the next Transformer reliably. But I don’t think we’re necessarily seeing that. There’s great research being done at a bunch of different labs. Branding they have.

David: Branding is a funny one. Well, I was going to say it’s a little bit to my bear case point about they’re the incumbent.

Ben: It cuts both ways, but I think it’s net positive.

David: Probably. For most people, they trust Google.

Ben: They probably don’t trust these, who knows AI companies, but I trust Google. I bet that’s actually stronger than any downsides as long as they’re willing to still release stuff on the cutting edge.

So to sum it up, it’s scale economies is the biggest one. It’s branding, and it’s a cornered resource.

David: And potential for switching costs in the future. Yup, sounds right to me.

Ben: But it’s telling that it’s not all of them. In Search, it was very obviously all of them or most of them.

David: Quite telling.

Ben: Well I’ll tell you, after hours and hours spending multiple months learning about this company, my quintessence when I boil it all down is just that this is the most fascinating example of the innovator’s dilemma ever.

Larry and Sergey control the company. They have been quoted repeatedly saying that they would rather go bankrupt than lose at AI. Will they really? If AI isn’t as good a business as Search—and it feels like of course it will be, of course it has to be; it’s just because of the sheer amount of value creation—if it’s not and they’re choosing between two outcomes, one is fulfilling our mission of organizing the world’s information and making it universally accessible and useful, and having the most profitable tech company in the world, which one wins? Because if it’s just the mission, they should be way more aggressive on AI mode than they are right now, and full flip over to Gemini. It’s a really hard needle a thread.

I’m actually very impressed at how they’re managing to currently protect the core franchise, but it might be one of these things where it’s being eroded away at the foundation in a way that just somehow isn’t showing up in the financials yet. I don’t know.

David: I totally agree. And in fact, perhaps influenced by you, I think my quintessence is a version of that too. I think if you look at all the big tech companies, Google, as unlikely as it seems given how things started, is probably doing the best job of trying to thread the needle with AI right now.

That is incredibly commendable to Sundar and their leadership. They’re making hard decisions, like we’re unifying DeepMind and Brain. We’re consolidating and standardizing on one model, and we’re going to ship this stuff real fast. While at the same time not making rash decisions.

Ben: It’s hard. Rapid but not rash, you know?

David: Yes. Obviously, we’re still in early innings of all this going on and we’ll see in 10 years where it all ends up.

Ben: Being tasked with being the steward of a mission and the steward of a franchise with public company shareholders is a hard dual mission. Sundar and the company is handling it remarkably well, especially given where they were five years ago. And I think this will be one of the most fascinating examples in history to watch it play out.

David: Totally agree. Well this concludes our Google series for now.

Ben: All right, let’s do some carve outs.

David: All right, let’s do some carve outs. Well first off, we have a very, very fun announcement to share with you all. The NFL called us.

Ben: We’re going to the Super Bowl, baby.

David: Acquired is going to the Super Bowl. This is so cool.

Ben: It’s the craziest thing ever.

David: The NFL is hosting an innovation summit the week at the Super Bowl, the Friday before Super Bowl Sunday. The Super Bowl is going to be in San Francisco this year in February. It’s only natural coming back to San Francisco with the Super Bowl that the NFL should do an innovation summit, and we’re going to host it.

Ben: That’s right. The Friday before, there’s going to be some great on-stage interviews and programming. Most of you, we can’t fit millions of people in a tidy auditorium in San Francisco the week of the Super Bowl when every other venue has tons of stuff too, so there’ll be an opportunity to watch that streaming online. And as we get closer to that date in February, we will make sure that you all know a way that you can tune in and watch the MC-ing, interviewing, and festivities at hand, Super Bowl week.

David: It’s going to be an incredible, incredible day leading up to an incredible Sunday.

Ben: Well speaking of sport, my carve out is I finally went and saw F1. It is great. I highly recommend anyone go see it, whether you’re an F1 fan or not. It is just beautiful cinema.

David: Amazing. Did you see it in the theater or…?

Ben: I did see it in the theater, yeah.

David: Wow. Nice.

Ben: I unfortunately missed the IMAX window, but it was great. It was my first time being in movie theater in a while, and whether you watch it at home or whether you watch it in the theater, I recommend the theater, but it’s going to be a great surround sound experience wherever you are.

David: I haven’t been to the movie theater since the Eras tour, which I think is just more about the current state of my family life with two young children.

Ben: My second one, so if you’re going to laugh, is the Travelpro suitcase.

David: This is the brand that pilots and flight attendants use, right?

Ben: Maybe. I think I’ve seen some of them use it. Usually they use something higher end like a Briggs and Riley or a Tumi or… Travelpro is not the most high-end suitcase, but I bought two really big ones for some international travel that we were doing with my 2-year-old toddler.

I must say they’re robust. The wheels glide really well. They’re really smooth. They have all the features you would want. They’re soft shell so you can really jam it full of stuff. But it’s also a thick amount of protection. so even if you do jam it full of stuff, it’s probably not going to break. This is approximately the most budget suitcase you could buy.

I’m looking at the big honking international check bag version. It’s $416 on Amazon right now. I’ve seen it cheaper. They have great sales pretty often. Everything about this suitcase checked lots of boxes for me, and I completely thought I would be the person buying the Rimowa suitcase or something very high end. This is just perfect. I think I may be investing in more Travelpro suitcases.

David: More Travelpro, nice. Well, hey look. For family travel. You don’t want nice stuff.

Ben: Yeah, I bought it thinking like I’ll just get something crappy for this trip, but it’s been great. I don’t understand why I wouldn’t have a full lineup of Travelpro gear, so…

David: Amazing.

Ben: This is my budget pick-gone-right that I highly recommend for all of you.

David: I love how Acquired is turning into the wire cutter here.

Ben: That’s it for me today.

David: Great. All right, I have two carve outs. I have one carve out and then I have an update in my ongoing Google Carve out saga. But first my actual carve out, it is the Glue Guys podcast.

Ben: Oh, that’s great. Those guys are awesome.

David: So great. Our buddy, Ravi Gupta, partner at Sequoia, and his buddies, Shane Battier, the former basketball player; and Alex Smith, the former quarterback for the 49ers, the Kansas City Chiefs, and the Redskins, their dynamic is so great. They have so much fun.

Half of their episodes, like us, are just them, and then half of their episodes are with guests. Ben and I, we went on it a couple of weeks ago. That was really fun. When we were on it, we were talking about this dynamic of some episodes do better than others and pressure for episodes and whatnot.

The guys brought up this interview they did with a guy named Wright Thompson. They said like, look, this is an episode. It’s got 5000 listens. Nobody’s listened to it. It’s so good. The mentality that we have about it is not that we’re embarrassed that nobody listened to it. It’s that we feel sorry for the people who have not yet listened to it because it’s so good. I was like, that is the way to think about your episodes.

Ben: So here you are. You’re giving everyone the gift of…

David: We’re giving everyone the gift because then I was like, all right, well I got to go listen to this episode.

Ray Thompson, I didn’t know anything about him before. I probably read his work in magazines over the years without realizing it. He’s the coolest dude. He has the same accent as Bill Gurley. Listening to him sounds like if Bill Gurley, instead of being a VC, only wrote about sports and basically dedicated his whole life to understanding the mentality and psychology of athletes and coaches, it’s so cool. It’s so cool. It’s a great episode. Highly, highly, highly recommend.

Ben: All right. Legitimately, I’m queuing that up right now.

David: Great. That’s my carve out. Then my ongoing family video gaming saga. In Google Part I, I said I was debating between the Switch 2 and the Steam Deck.

Ben: That’s right. First you got the Steam Deck because you decided your daughter actually wasn’t old enough to play video games with you, so you just got the thing for you.

David: The update was I went with the Steam Deck for that reason. I thought if it’s just for me, it would be more ideal. I have an update.

Ben: You also got a Switch?

David: No, not yet. But the most incredible thing happened. My daughter noticed this device that appeared in our house that dad plays every now and then. We were on vacation and I was playing the Steam Deck. She was like, what’s that? Well let me tell you.

I’ve been playing this really cool, indie, old school–style RPG called Sea of Stars. It’s like a Chrono Trigger–style, Super Nintendo–style RPG. I’m playing it, and when my daughter comes up she’s like, can I watch you play? And I’m like, hell yeah, you can watch me play. I get to play video games and you sit here and snuggle with me. Amazing.

Ben: I get to play video games and call it parenting.

David: Then it gets even better. Probably two weeks ago we’re playing and she’s like, hey dad, can I try? I’m like, absolutely you can try. I hand her the Steam Deck and it was one of the most incredible experiences I’ve had as a parent because she doesn’t know how to play video games. I’m watching her learn how to use a joystick and hit the button.

Ben: Wow. Supervised learning.

David: Yeah, yeah, yeah. Supervised learning. I’m telling her what to do, and then within two or three nights she got it. She doesn’t even know how to read yet, but she figured it out and I’m watching her in real time.

Now the last week, it’s turned to mostly she’s playing and I’m helping her asking questions of like, well what do you think you should do here? Should you go here? I think this is the goal. I think this is where… It’s so, so fun.

I think I might actually, pretty soon, her birthday’s coming up, end up getting a Switch so that we can play together on the Switch. But unintentionally the Steam Deck was the gateway drug for my soon-to-be 4-year-old daughter.

Ben: That’s awesome. There you go. Parent of the year right there. Getting to play video games and… Oh, honey. I got it. I’ll take it.

David: Oh yeah, I got it.

Ben: All right, well listeners, we have lots of thank yous to make for this episode. We talked to so many folks who are instrumental in helping put it together.

First to thank you to our partners this season. J.P. Morgan Payments. Trusted, reliable payments infrastructure for your business no matter the scale. That’s jpmorgan.com/acquired.

Sentry, the best way to monitor for issues in your software and fix them before users get mad. That’s sentry.io/acquired.

WorkOS, the best way to make your app enterprise-ready starting with single sign on in just a few lines of code. workos.com.

And Shopify, the best place to sell online, whether you’re a large enterprise or just a founder with a big idea. shopify.com/acquired.

The links are all in the show notes. As always, all of our sources for this episode are linked in the show notes.

David: First, Steven Levy at Wired and his great classic book on Google, In the Plex, which has been an amazing source for all three of our Google episodes. Definitely go by the book and read that.

Also to Parmy Olson at Bloomberg for her book, Supremacy, about DeepMind and OpenAI, which was a main source for this episode.

I guess also to Cade Metz, right?

Ben: For Genius Makers, yeah. Great book.

David: Our research thank yous: Max Ross, Liz Reid, Josh Woodward, Greg Corrado, Sebastian Thrun, Anna Patterson, Bret Taylor, Clay Bavor, Demis Hassabis, Thomas Kurian, Sundar Pichai.

A special thank you to Nick Fox, who is the only person we spoke to for all three Google episodes for research. We got the hat trick.

Ben: To Arvind Navaratnam at Worldly Partners for his great writeup on Alphabet, link in the show notes.

To Jonathan Ross, original team member on the TPU, and today the founder and CEO of Groq, making chips for inference.

To the Waymo folks, Dmitri Dolgov and Suzanne Philion.

To Gavin Baker from Atreides Management.

To M.G. Siegler, writer at Spyglass. M.G. is just one of my favorite technology writers and pundits.

David: OG TechCrunch writer.

Ben: That’s right. To Ben Eidelson, for being a great thought partner on this episode and his excellent recent episode on the Stepchange podcast on the history of data centers. I highly recommend it if you haven’t listened already. It’s only episode three for them of the entire podcast and they’re already getting, I don’t know, 30,000–40,000 listens on it. This thing is taken off.

David: Amazing. Dude, that’s way better than we were doing on episode three.

Ben: It’s way better than we were doing. And if you like Acquired, you will love the Stepchange podcast. Ben is a dear friend, so highly recommend checking it out.

To Koray Kavukcuoglu from the DeepMind team, building the core Gemini models.

To Shishir Mehrotra, the CEO of Grammarly, formerly ran product at YouTube.

To Jim Gao, the CEO of Phaidra and former DeepMind team member.

Chetan Puttagunta, partner at Benchmark.

To Akash Patel for helping me think through some of my conclusions to draw, and to Bryan Lawrence from Oakcliff Capital for helping me think about the economics of AI data centers.

If you like this episode, go check out our episode on the early history of Google in the 2010s with our Alphabet episode, and of course our series on Microsoft and NVIDIA. After this episode, go check out ACQ2 with Tobi Lütke, the founder and CEO of Shopify, and come talk about it with us in the Slack at acquired.fm/slack.

David: And don’t forget our 10th anniversary celebration Acquired. We are going to do a open Zoom call, an LP call just like the days of yore with anyone. Listeners, come join us on Zoom, it’s going to be on October 20th at 4:00 PM Pacific Time. Details are in the show notes.

Ben: And with that listeners, we’ll see you next time.

David: We’ll see you next time.

Note: Acquired hosts and guests may hold assets discussed in this episode. This podcast is not investment advice, and is intended for informational and entertainment purposes only. You should do your own research and make your own independent decisions when considering any financial transactions.

More Episodes

All Episodes > 

Thank you! You're now subscribed to our email list, and will get new episodes when they drop.

Oops! Something went wrong while submitting the form