Gavin Baker - Nvidia v. Google, Scaling Laws, and the Economics of AI
88m 19s
The transcription covers various topics related to technology, finance, and AI. Bramps AI streamlines expense reviews, allowing finance teams to focus on strategic tasks. Ridgeline is commended for its innovative asset management technology, aiding firms in scaling efficiently. Alpha Sense provides AI-powered channel checks for early insights into public companies. The "Invest Like The Best" podcast delves into market strategies and ideas. Gavin Baker discusses technology and investing on podcasts. Detailed discussions on Gemini 3, Blackwell chip, and scaling laws in AI are presented. A comparison between Google's TPU and Nvidia's GPU in the context of AI advancements is also highlighted.
Transcription
17082 Words, 95200 Characters
Here's an interesting question to think about, if your finance team suddenly had an extra
week every month, what would you have them work on? Most CFOs don't know because their
finance teams are grinding it out on lost expense reports, invoice coding, and tracking down
receipts until the last possible minute. That's exactly the problem that Bramps had out to solve.
Looking at the parts of finance, everyone quietly hates and asking why are humans doing any of this?
Turns out they don't need to. Bramps AI handles 85% of expense reviews automatically with 99%
accuracy, which means your finance team stops being the department that processes stuff and
starts being the team that thinks about stuff. Here's the real shift. Companies using ramp
aren't just saving time, they're reallocating it. While competitors spend two weeks closing their
books, you're already planning next quarter. While they're cleaning up spreadsheets, you're thinking
about new pricing strategy, new markets, and where the next dollar of ROI comes from. That difference
compounds. Go to ramp.com/invest to try ramp and see how much leverage your team gains
when the work you have to do stops getting in the way of the work that you want to do.
To me, Ridgeline isn't just a software provider, it's a true partner in innovation.
They're redefining what's possible in asset management technology, helping firms scale faster,
operates smarter, and stay ahead of the curve. I want to share a real-world example of how they're
making a difference. Let me introduce you to Brian. Brian, please introduce yourself and tell
us a bit about your role. My name is Brian Strang. I'm the Technical Operations Lead and I work at
Congress Asset Management. How would you describe your experience working with Ridgeline?
Ridgeline is a technology partner, not a software vendor, and the people really care. I get sales,
calls all the time and ignore them. Ridgeline sold me very quickly. We went from $7 billion to $23 billion
and the goal is $50 billion. Ridgeline was the clear frontrunner to help us scale.
In your view, what most distinguishes Ridgeline? They reimagined how this industry should work,
it's obvious that they were operating on another level. It's worth reaching out to Ridgeline to see
what the unlock can be for your firm. Visit RidgelineApps.com to schedule a demo.
One of the hardest parts of investing is seeing what shifting before everyone else does. Alpha
Sense is helping investors do exactly that. You may already know Alpha Sense is the market intelligence
platform trusted by 75% of the world's top hedge funds, providing access to over 500 million
premium sources from company filings and broker research to news, trade journals, and over 200,000
expert transcript calls. What you might not know is that they've recently launched something
game-changing, AI-powered channel checks. Channel checks give you a real-time expert-driven
perspective on public companies, weeks before they show in earnings or consensus revisions. Alpha
Sense uses an AI interviewer to run thousands of expert calls with real human experts every month,
asking consistent questions across experts so the signals are clean, comparable, and useful.
You get live updates as interviews come in, full transcript access, and coverage across
every major sector. Instantly compare insights across experts and analyze quarter-over-quarter trends
in sentiment and key performance indicators. For investors trying to stay ahead of the fast moving
markets, it's already table stakes. Hello and welcome everyone, I'm Patrick O'Shaughnessy and
this is Invest Like The Best. This show is an open-ended exploration of markets, ideas, stories,
and strategies that will help you better invest both your time and your money. If you enjoy these
conversations and want to go deeper, check out Colossus Review, our quarterly publication with
in-depth profiles of the people shaping business and investing. You can find Colossus Review along
with all of our podcasts at join Colossus.com. Patrick O'Shaughnessy is the CEO of Positive Some.
All opinions expressed by Patrick and podcast guests are solely their own opinions and do not
reflect the opinion of Positive Some. This podcast is for informational purposes only and should not
be relied upon as a basis for investment decisions. Clients of Positive Some may maintain positions
in the securities discussed in this podcast. To learn more, visit psum.vc.
I will never forget when I first met Gavin Baker in 2017. I find his interest in markets
his curiosity about the world to be as infectious as any investor that I've ever come across.
He is encyclopedic on what is going on in the world of technology today, and I've had the good fortune
to host him every year or two on this podcast. In this conversation, we talk about everything
that interests Gavin. We talk about Nvidia, Google, and its TPUs, the changing AI landscape, the math,
and business models around AI companies, and everything in between. We even discuss the crazy idea
of data centers in space, which he communicates with his usual passion and logic.
In closing, at the end of this conversation, because I've asked him my traditional closing question
before, I asked him a different question, which led to a discussion of his entire investing
origin story that I had never heard before. Because Gavin is one of the most passionate thinkers
and investors that I know, these conversations are always amongst my most favorite.
I hope you enjoy this latest in the series of discussions with Gavin Baker.
I would love to talk about how you, like in the nitty-gritty process, new things that come out
in this AI world, because it's happening so constantly. I'm extremely interested in it, and I find
it very hard to keep up, and I have a couple of blogs that I go read and friends that I call.
But maybe let's take Gemini 3 as a recent example. When that comes out, take me into your office,
what are you doing? How do you and your team process an update like that, given how often these things
are happening? I mean, I think the first thing is you have to use it yourself. And I would just say,
I'm amazed at how many famous and August investors are reaching really definitive conclusions about AI.
Well, now based on the free tier, the free tier is like you're dealing with the 10-year-olds,
and you're making conclusions about the 10-year-olds capabilities as an adult. And you could just pay,
and I do think, actually, you do need to pay for the highest tier, whether it's Gemini Ultra,
Super Grock, whatever it is, you have to pay the $200 per month tiers. Whereas those are like a
fully-fledged 30-35-year-old, it's really hard to extrapolate from a native or a 10-year-old to the
35-year-old, and yet a lot of people are doing that. And the second thing is, there was an insider
post about OpenAI, and they said to a large degree OpenAI runs on quitterfops. And I just think AI
happens on X. There have been some really memorable moments, like there was a giant fight between
the PyTorch team at Meta and the Jack's team at Google on X. And the leaders of each lab had to
step in publicly say, "No one from my lab is allowed to say bad things about the other lab,
and I respect them, and that is the end of that." The companies are all commenting on each other's
posts, you know, the research papers come out. There's a list of, you know, if on planet Earth,
there's 500 to 1,000 people who really, really understand this and are at the cut of edge of it.
And a good number of them live in China. I just think you have to follow those people closely,
and I think there is an incredible signal. Everything in AI is just downstream.
Of those people? Yeah. Everything Andre Karpathy writes. You have to read it three times.
Yeah. Minimum. Yeah, he's incredible. And then I would say any time one of those labs, the four labs
that matter, open AI, Jim and I, and throughout the Connected AI, which are clearly the four leading labs.
Any time somebody from one of those labs goes on a podcast, I just think it's so important to listen.
For me, one of the best use cases of AI is to keep up with all of this. Listen to a podcast,
and then if there are parts that I thought were interesting, just talk about it with AI. And I
think it's really important to have as little friction as possible. I'll bring it up. I can either
press this button and pull up rock, or I have this. It's so amazing. If people leave, we have this.
I know. It's like somebody said on X, you know, like we imbued these rocks with crazy spells.
And now we can summon super intelligent genies on our phones over the air. You know, it's crazy.
Crazy. Okay. So something like Gemini 3 comes out with the public interpretation was,
oh, this is interesting. It seems to say something about scaling laws and the pre-training stuff.
What is your frame on like the state of general progress in frontier models in general? Like what
are you watching most closely? Yeah. Well, I do think Gemini 3 was very important because it showed us
that scaling laws for pre-training are intact. They stated that unequivocally. And that's important
because no one on planet Earth knows how or why scaling laws for pre-training work. It's actually
not a law. It's an empirical observation. And it's an empirical observation that we've measured
extremely precisely in this held for a long time. But our understanding of scaling laws for pre-training,
and maybe this is a little bit controversial with 20% of researchers, but probably not more than that.
It's kind of like the ancient British people's understanding of the sun. Are the ancient
Egyptians understanding the sun? They can measure it so precisely that the east-west access
to the great pyramids are perfectly aligned with the equinoxes. And so are the east-west
accesses of Stonehenge? Perfect measurement. They didn't understand orbital mechanics.
They had no idea how or why it rose in the east, set in the west, and moved across the horizon.
There's the aliens in the ocean. Yeah. Yeah. Our god had a chariot. And so it's really important every
time we get a confirmation of that. So Gemini 3 was very important in that way. But I think there's
been a big misunderstanding of maybe in the public equity investing community, or the broader
more generalist community. Based on the scaling laws of pre-training, there really should have been
no progress in 24 and 25. And the reason for that is, is after XAI figured out how to get
200,000 hoppers coherent, you had to wait for the next generation of chips. Because you really can't
get more than 200,000 hoppers coherent. And coherent just means you could just think of it as
each GPU knows what every other GPU is thinking. They kind of are sharing memory. They're connected.
The scale of networks and the scale out. And they have to be coherent during the pre-training
process. The reason we've had all this progress, maybe we could show like the ARC, AGI slide where you
had 0 to 8 over four years, 0 to 8% intelligence. And then you went from 8% to 95% in three months
when the first reasoning model came out from OpenAI. We have these two new scaling laws of post-training,
which is just reinforcement learning with verified rewards. And verified is such an important
concept in AI. Like one of Carpathi's great things was with software. Anything you can specify,
you can automate with AI. Anything you can verify, you can automate. It's such an important concept.
And I think important distinction. And then test time compute. And so all the progress we've had,
we've had immense progress since October 24 through today was based entirely on these two new
scaling laws. And Jim and I-3 was arguably the first test since Hopper came out of the scaling
law for pre-training. And it held. And that's great because all these scaling laws are multiplicative.
So now we're going to apply these two new reinforcement learning, verified rewards and test time
compute to much better base models. There's a lot of misunderstanding about Jim and I-3 that I think
is really important. So the most important thing to conceptualize everything in AI has a struggle
between Google and Nvidia. And Google has the TPU. And Nvidia has their GPUs and Google only has
a TPU and they use a bunch of other chips for networking. And Nvidia has the full stack. Blackwell
was delayed. Blackwell was Nvidia's next generation chip. The first iteration of that was the Blackwell
200. A lot of different skews were canceled. And the reason for that is it was by far the most
complex product transition we've ever gone through in technology. Going from Hopper to Blackwell,
first you go from air cooled to liquid cooled. The rack goes from weighing round numbers a thousand
pounds to three thousand pounds. Goes from round numbers 30 kilowatts, which is 30 American homes to
130 kilowatts, which is 130 American homes. I analogize it to imagine if to get a new iPhone,
you had to change all the outlets in your house to 220 volts, put in a Tesla power wall, put in
a generator, put in solar panels, that's the power, you know, put in a whole home humidification
system and then reinforce the floor because the floor can't handle this. So it was a huge product
transition. And then just the rack was so dense, it was really hard for them to get the heat out.
So Blackwells have only really started to be deployed and really scaled deployments
over the last three or four months. Can you explain why it spent such an important thing that Blackwell
was delayed? Because Blackwell is so complicated and it was so hard for everyone to get these
exquisitely complex racks working consistently. Had reasoning not come along, there would have been
no AI progress from bid 2024 through essentially Jim and I three. There would have been none. Everything
would have stalled. And you can imagine what that would have meant to the markets. For sure,
we would have lived in a very different environment. So reasoning kind of bridged this 18-month gap,
reasoning kind of saved AI because it let AI make progress without Blackwell or the next generation
of TPU, which were necessary for the scaling loss for pre-trading to continue. Google came out
with the TPU V6 in 2024 and the TPU V7 in 2025. In semiconductor time, imagine like Hopper,
it's like a World War II era airplane. And it was by far the best World War II era airplane.
It's P51 Mustang with the Merlin engine. And two years later, in semiconductor time,
that's like you're an F4 phantom, okay? Because Blackwell was such a complicated product and so hard
to ram. Google was training Jim and I three on 24 and 25 era TPUs, which are like F4 phantoms.
Blackwell, it's like an F35. It just took a really long time to get it going. So I think Google for
sure has this temporary end of advantage right now from pre-training perspective. I think it's also
important that they've been the lowest cost producer of tokens. And this is really important because
AI is the first time in my career as a tech investor that being the low cost producers ever matter.
Apple is not worth trillions because they're low cost producer of phones. Microsoft is not worth
trillions because they're low cost producer of software. And video is not worth trillions because
they're the low cost producer of AI accelerators. It's never mattered. And this is really important
because what Google has been doing has the low cost producer is they have been sucking the economic
oxygen out of the AI ecosystem, which is an extremely rational strategy for them. And for anyone
is the low cost producer, let's make life really hard for our competitors. So what happens now,
I think it says pretty profound implications. One, we'll see the first models trained on Blackwell
in early 2026. I think the first Blackwell model will come from XAI. And the reason for that is just
according to Jensen, no one builds data centers faster than Eva. Jensen has said this on the record.
And even once you have the Blackwells, it takes six to nine months to get them performing at the
level of Hopper. Because Hopper is finally tuned. Everybody knows how to use it. The software is
perfect for it. The engineers know all its quirks. Everybody knows how to architect a Hopper data
service. And by the way, when Hopper came out, it took six to 12 months for it to really outperform
Ampere, which was generation before. So if you're Jensen or Nvidia, you need to get has many GPUs
deployed in one data center as fast as possible and a coherent cluster so you can work out the bugs.
And so this is what XAI effectively does for Nvidia, because they build the data centers
the fastest. They can deploy Blackwells at scale the fastest and they can help work with Nvidia
to work out the bugs for everyone else. So because they're the fastest, they'll have the first Blackwell
model. We know that scaling loss for pre-training are intact. And this means the Blackwell models
are going to be amazing. Blackwell is, I mean, it's not enough 35 first and last four phantom,
but from my perspective, it is a better chip. You know, maybe it's like an F 35 or so Raphael.
And so now that we know pre-skilling loss for holding, we know that these Blackwell models are going
to be really good based on the raw specs. They should probably be better. And then something even
more important happens. So the GB 200 was really hard to get it going. The GB 300 is a great chip.
It is drop in compatible in every way with those GB 200 racks. Now you're not going to replace the
GB 200s, but just any power walls. Yeah, just any data center that can handle those. You can
slot in the GB 300s and now everybody's good at making those racks and you know how to get the
heat out. You know how to cool them. You're going to put those GB 300s in. And then the companies
that use the GB 300s, they're going to be the low-cost producer of tokens, particularly if you're
vertically integrated. If you're paying a margin to someone else to make those tokens, you're probably
not going to be. And he just has pretty profound implications. I think it has to change Google's
strategic calculus. If you have a decisive cost advantage in your Google and you have search
and all these other businesses, why not run AI at a negative 30% margin? It is by far the rational
decision to take the economic oxygen out of the environment. You eventually make it hard for
your competitors who need funding unlike you to raise the capital they need. And then on the other
side of that, maybe you have an extremely dominant sheer position. While that calculus changes,
once Google is no longer the low-cost producer, which I think will be the case. The Blackwells are now
being used for training. And then when that model is trained, you start shifting Blackwell clusters
over to inference. And then all these cost calculations and these dynamics change.
It's very interesting, like during the strategic and economic calculations between the players,
I've never seen anything like it. Everyone understands their position on the board.
What the prize is, what play their opponents are running. And it's really interesting to watch.
If Google changes its behavior, because it's going to be really painful for them as a
higher-cost producer to run that negative 30% margin, it might start to impact their stock.
That has pretty profound implications for the economics of AI. And then when Ruben comes out,
the gap is going to expand significantly. Versus TPUs. Versus TPUs and all other ASICs.
Now, I think training of three is probably going to be pretty good and training four are going to
be good. Why is that the case? Why won't TPU be V9 be everybody's good?
A couple of things. So one, for whatever reason, Google made more conservative design decisions.
I think part of that is so Google, let's say the TPU. So there's front-end and back-end
of semiconductor design and then there's dealing with Taiwan Semi. And you can make an ASIC in a lot of
ways. What Google does is they do mostly the front-end for the TPU and then Broadcom does the back-end
and manages Taiwan Semi and everything. It's a crude analogy but the front-end is like the
architect of a house. They design a house. The back-end is the person who builds the house. And the
managing Taiwan Semi is like stamping out that house like Lunar or D.R. Horton. And for doing those
two lot of parts, Broadcom runs a 50 to 55% gross margin. We don't know what on TPUs. Let's say in
2027, TPU, I think it's its estimates, maybe somewhere around 30 billion. Again, who knows? 30 billion,
I think is a reasonable estimate. 50 to 55% gross margins. So Google is paying Broadcom 15 billion dollars.
That's a lot of money. At a certain point, it makes sense to bring a semiconductor program entirely
in house. So in other words, Apple does not have an ASIC partner for their chips. They do the front-end
themselves, the back-end and they manage Taiwan Semi. And the reason is they don't want to pay that 50%
margin. So at a certain point, it becomes rational to renegotiate this and just just perspective.
The entire op-ex of Broadcom Semi Conductor Division is around numbers 5 billion dollars. So it'd be
economically rational. Now that Google's paying, if it's 30 billion or paying them 15, Google can go
to every person who works in Broadcom Semi, double their comp and make an extra 5 billion. In 2028,
let's just say it does 50 billion. Now it's 25 billion. You can triple their comp. And by the way,
you don't need them all. And of course, they're not going to do that because of competitive concerns.
But with TPU-V8 and V9, all of this is beginning to have an impact because Google is bringing in media
tech. This is maybe the first way you send a warning shot to Broadcom. We're really not happy
about all this money we're paying. But they did bring media tech in in the Taiwanese ASIC companies,
have much lower gross margins. So this is kind of the first shot against the bowel. And then there's
all this stuff people say, but Broadcom has the best 30s. Broadcom has really good 30s and 30s is
like an extremely foundational technology because it's how the chips communicate with each other.
You have to serialize it, do serialize. But there are other good 30s providers in the world.
A really good 30s is maybe it's worth 10 or 15 billion a year, but it's probably about 25 billion a year.
So because of that friction, and I think conservative design choices on the part of Google,
and maybe the reason they made those conservative design choices is because they were going to a
bifurcated supply. TPU is slowing down. I would say as the GPUs are accelerating, this is the first
competitive response of Lisa and Jensen to everybody saying we're going to have our own ASIC,
is hey, we're just going to accelerate. We're going to do a GPU every year and you cannot keep up with
us. And then I think what everybody is learning is like, oh wow, that's so cool. You made your own
accelerator has an ASIC. Wow, what's the NIC going to be? What's the CPU going to be?
What's the scale up switch going to be? What's the scale of protocol? What's the scale out switch?
What kind of objects are you going to use? What's the software that's going to make all this work
together? And then it's like, oh shit, I made this tiny little chip. And you know, like
whether it's admitted or not, I'm sure the GPU makers don't love it when their customers make
ASICs to try and compete with them. And like, whoops, what did I do? I thought this was easy. You know,
it takes at least three generations to make a good chip. Like the TPU V1, I mean, it was an achievement
and then they made it. It was really not till TPU V3 or V4 that the TPU started to become like
even vaguely competitive. Is that just a classic like learning by doing thing? 100%. And even if you've
made from my perspective, the best ASIC team at any semiconductor company is actually the Amazon ASIC team.
They're the first one to make the graph on the CPU. They have this Nitro, it's called SuperNIC.
They've been extremely innovative, really clever. And like, training them an infantry one,
maybe they're a little better than the TPU V1, but only a little. Training them too, you get a little
better. Training them three. I think the first time, it's like, okay. And then, you know, I think
training four will probably be good. I will be surprised if there are a lot of ASICs other than
training them in TPU. And by the way, and training them in TPU will both run on customer-owned tooling
at some point. We can debate when that will happen, but the economics of success that I just described
mean it's inevitable. Like, no matter what the companies say, just the economics make it and
reasoning from first principles make it absolutely inevitable. If I were to zoom all the way out on
this stuff, I find these details unbelievably interesting and it's like the grandest game that's
ever been played. It's so crazy and so fun to follow. Sometimes I forget to zoom out and say,
well, so what? Like, okay. So, project this forward three generations past Rubin or whatever.
What is like the global human dividend of all this crazy development?
Will we keep making the loss lower on these pre-training scaling models? Like, who cares? Like,
it's been a while since I've asked this thing something that I wasn't kind of blown away by the
answer for me personally. What are the next couple of things that all this crazy infrastructure
war allows us to unlock because they're so successful? If I were to pause it like an event path,
I think the Blackwell models are going to be amazing. The dramatic reduction in per token cost
enabled by the GP300 and the prime more the MI450 than the MI355 will lead to these models being
allowed to think for much longer, which means they're going to be able to do new things. I was very
impressed. My three made me a restaurant reservation. It's the first time it's done something for me.
And I mean, other than like go research something and teach me stuff, if you can make a restaurant
reservation, you're not that far from being able to make a hotel reservation and an airplane reservation
and order me an Uber and- All of a sudden, you got an assistant. Yeah, and you can just imagine
everybody talks about that. You can just imagine it's on your phone. I think that's pretty near-term.
But some big companies that are very tech forward, 50% plus of customer support is already done
by AI and that's a $400 billion industry. And then if you know what AI is great about is persuasion,
that sales and customer support. And so of the functions of a company, if you think about them,
they're to make stuff, sell stuff, and it support the customers. So right now maybe in late 26,
you're going to be pretty good at two of them. I do think it's going to have a big impact on media.
Like I think robotics, you know, we talked about the last time, are going to finally start to be real.
You know, there's an explosion and kind of exciting robotics startups. I do still think that the
main battle is going to be- we in Tesla's Optimus in the Chinese because, you know, it's easy to make
prototypes. It's hard to mass produce them. But then it goes back to that what Andre Carpathi
said about AI can automate anything that can be verified. So any function where there's a right
or wrong answer or right or wrong outcome, you can apply reinforcement learning and make the AI
really good at that. What are your favorite examples of that so far or theoretically?
I mean, just does the model balance. They'll be really good at making models. Do all the books
globally reconciled. They'll be really good at accounting, double entry bookkeeping and haste
balance. There's a verifiable you got it right or wrong. Supporters say, did you make the sale or not?
That's just like AlphaGo. Did you win or you lose? Did the guy convert or not? Did the customer ask
for an escalation during customer support or not? It's most important functions are important because
they can be verified. So I think if all of this starts to happen and starts to happen in 26,
they'll be in ROI on Blackwell and then all this will continue and then we'll have Ruben.
And then that'll be another big quantum spin Ruben in the MI450 and the TPU V9. And then I do think
just the most interesting question is what are the economic returns to artificial superintelligence?
Because all of these companies in this great game, they've been in a prisoner's dilemma. They're
terrified that if they slow down gone forever and their competitors don't, it's an existential risk.
And you know, Microsoft blinked for like six weeks earlier this year. Yeah.
But I think they would say they regret that. But with Blackwell and for sure with Ruben,
the economics are going to dominate the prisoner's dilemma from a decision making and
spending perspective just because the numbers are so big. And this goes to kind of the ROI on AI
question. And the ROI on AI has empirically, factually, unambiguously been positive. I just always
find it strange that there's any debate about this because the largest bidder just on GPUs or public
companies, they report something called audited quarterly financials and you can use those things
to calculate something called a return on investment capital. And if you do that calculation,
the ROIC of the big public spenders on GPUs is higher than it was before they ramped spinning.
And you could say, well, part of that is, you know, up savings. Well, at some level,
that is part of what you expect the ROI to be from AI. And then you say, well, a lot of is actually
just applying to use moving the big recommender systems that power the advertising and the recommendation
systems from CPUs to GPUs and you've had massive efficiency gains. And that's why all the revenue
growth at these companies is accelerated. But like, so what? The ROI has been there. And it is
interesting, like every big internet company, the people who are responsible for the revenue
are intensely annoyed at the amount of GPUs that are being given to the researchers.
It's a very linear question. If you give me more GPUs, I will drive more revenue. Yeah.
Give me those GPUs. We'll have more revenue, more gross profit. And then we get to everybody.
So it's as constant fight at every company. One of the factors in the prisoners still,
Emma, as everybody has this like religious belief that we're going to get to ASI. And at the end of
the day, what do they all want? Almost all of them want to live forever. And they think that ASI
is going to help them with that. Right. Good return. That's good. But we don't know. And if
as humans, we have pushed the boundaries of physics, biology, and chemistry, the natural laws that
govern the universe, then maybe the economic returns to ASI aren't that high. I'm very curious
about your favorite sort of throw cold water on this stuff. Type takes that you think about sometimes.
One would be like the things that would cause, I'm curious what you think, the things that would
cause this demand for compute to change or even the trajectory of it to change. There's one
really obvious bear case that it is just edge AI. And it's connected to the economic returns to ASI
in three years on a bigger and bulkier phone to fit the amount of DRAM necessary. You know,
in the battery won't part less as long. You will be able to probably run like a pruned down version
of something like Jim and I five or GROC for GROC 4.1 or chat GPT at 30 60 tokens per second.
And then that's free. And this is clearly Apple's strategy. It's just we're going to be a distributor
of AI and we're going to make it privacy safe and run on the phone. And then you can call one of the
big models, you know, the God models in the cloud, whatever you have a question. And if that happens,
if like 30 60 tokens a second at a 115 IQ is good enough, I think that's a bear case other than
just the scaling laws break. But in terms of if we assume scaling laws continue and we now
know they're going to continue for pre-training for at least one more generation. And we're very early
in the two new scaling laws for post-training, mid-training, RLVR, whatever people want to call it,
and then test time computed inference. We're so early in those and we're getting so much better
at helping the models hold more and more context in their minds as they do this test time compute.
And that's really powerful because everybody's like, well, how's the model going to know this? Well,
eventually you can get through, you can hold enough context. You can just hold every slack message
and outlook message and company manual. And in a company, in your context, and then you can compute
the new task and compare it with your knowledge of the world, what you think with the model things,
all this context and maybe that like just really, really long context windows are the solution to a
lot of the current limitations. And that's enabled by all these cool tricks like KV Cash Offload and
stuff. But I do think other than scaling laws slowing down other than there being low economic
returns to ASI, edge AI is to me by far the most plausible and scariest bear case.
I like to visualize like different S-curve you invested through the iPhone. And I love to like see
the visual of the iPhone models as it sort of went from this clunky, bricky thing up to the what we
have now. We're like, each one's like a little bit, you know, obviously we've sort of petered out
on its form factor. If you picture something similar for the frontier models themselves,
does it feel like it's at a certain part of that natural technology paradigm progression? If you're
paying for human eye ultra or super rock and you're getting the good AI, it's hard to see differences.
Like I have to go really deep on something like do you think PCI Express or Ethernet is a better
protocol for scale up networking and why show me the scientific vapors. And if you shift between
models and you ask a question like that, where you know it really deeply, then you see differences.
I do play fantasy football, winnings are turning into charity, but it is like, you know,
these new models are quite a bit better at helping who should I play. They think in much more
sophisticated ways. If you're a historically good fantasy football player and you're having a bad
season, this is why this is why because you're not using it, you know, and I think we'll see that
in more and more domains. But I do think they're already at a level where unless you're a true expert
or just have an intellect that is beyond mind, it's hard to see the progress. And that's why I do think
we just shift from getting more intelligent to more useful unless more intelligence starts leading
to these massive scientific breakthroughs and we're hearing cancer in 26 and 27. I don't know
that we're going to be hearing cancer, but I do think from almost an ROIS curve, we need to kind of
hand off from intelligence to usefulness and then usefulness will then have to hand off to scientific
breakthrough just that creates whole new industries. What are the building blocks of usefulness in your
mind? Just being able to do things consistently and reliably. And a lot of that is keeping all the
context. Like, there's a lot of context if someone wants to plan a trip for me. Like, you know, I've
acquired these strange preferences. Like I follow that guy, Andrew Cuperman. So I like to have an
East facing balcony. So I think I'm warning Sun, you know, so the AI has to remember, here's how I like
to fly. Here are my preferences for that. Being out of plain with Starlink is important to me. Okay,
here are the resorts I've historically liked. Here are the kinds of areas I've liked. Here are the
rooms that I would really like at each. That's a lot of context and to keep all of that and kind of
wait those, it's a hard problem. So I think context windows are a big part of it. You know, there's this
meter task evaluation thing how long it could work for and you could think of that as being related
in some way to context, not precisely, but that just task length needs to keep expanding because
booking a restaurant and booking is economically useful, but it's not that economically useful.
But booking me an entire vacation and knowing the preferences of my parents, my sister,
my niece and my nephew, that's a much harder problem and that's something that like a human
might spend three or four hours on optimizing that. And then if you can do that, that's amazing.
But then again, I just think it has to be good at sales and customer support relatively soon.
And then after that, it has to be and I think it is already here. I do think we're going to see
in a kind of an acceleration and the awesomeness of various products engineers are using AI to make
products better and faster. We both invested in Fortell, the here and the company, which is just
absolutely remarkable, like something I never would have thought of. And we're going to see I think
something like that in every vertical and that's AI being used for the most core function of any
company, which is designing the product. And then it will be, you know, there's already lots of
examples of AI being used to help manufacture the product and distribute it more efficiently,
whether it's optimizing a supply chain, having a vision system, watch a production line. A lot of
stuff is happening. I think it's really interesting in this whole ROI part is 4 to 500 companies are
always the last to adopt a new technology. They're conservative. They have lots of regulations, lots of
lawyers. Startups are always the first. So let's think about the cloud, which was the last truly
transformative new technology for enterprises, being able to have all of your compute and the cloud
and usas. So it's always upgraded. It's always great, et cetera, et cetera. You can get it on every
device. I think the first AWS reinvent, I think it was in 2013. And by 2014, every startup on
Planet Earth ran on the cloud. The idea that you would buy your own server and storage box and
router was ridiculous. And that probably happened like even earlier. That probably already happened
before the first reinvite. The first big fortune 500 companies started to standardize on it like maybe
five years later. You see that with AI. I'm sure you've seen this in your startups. And I think one
reason VCs are more broadly bullish on AI than public market investors is VCs see very real productivity
gains. There's all these charts that for a given level of revenue, a company today
has significantly lower employees than a company of two years ago. And the reason is AI is doing
a lot of the sales, the support, and helping to make the product. I mean, there's, you know,
iconic has some charts, a 16z. By the way, David George is a good friend. Great guy. He has his model
busters thing. So there's very clear data that this is happening. So people who have a lens
into the world of venture see this. And I do think it was very important in the third quarter. This is
the first quarter where we had fortune 500 companies outside of the tech industry give specific
quantitative examples of AI driven uplift. So see, Robinson went up something like 20% on earnings.
Should I tell people what sees Robinson does? Yeah. Like, let's just say a truck goes from Chicago
to Denver. And then the trucker lives in Chicago. So it's going to go back from Denver to Chicago.
There's an empty load. It's the a Robinson has all these relationships with these truckers
and trucking companies. And they match shippers demand with that empty load supply to make the
trucking more efficient. You know, they're a free foreritor. You know, there's actually lots of
companies like this, but they're the biggest and most dominant. So one of the most important things
they do is they quote price and availability. So somebody a customer calls them up and says,
hey, I urgently need three 18 wheelers from Chicago to Denver. But in the past, they said it would take
them, you know, 15 to 45 minutes. And they only quoted 60% of inbound requests with AI. They're
quoting 100% and doing it in seconds. And so they printed a great quarter and stock went up 20%
and it was because of AI driven productivity that's impacting the revenue line, the cost line,
everything. I was actually very worried about the idea that we might have this black well ROI
air gap because we're spending so much money on black well. Those black wells are being used for
training. And there's no ROI on training. Training is you're making the model. The ROI comes from
inference. So I was really worried that, you know, we're going to have maybe this three quarter period
where the cap access unimaginably high. Those black wells are only being used for training.
Right. Arsting, flat eyes going off. Exactly. So ROIC goes down and you can see like meta, meta they
printed, you know, because meta has not been able to make a frontier model. Meta printed a quarter
where ROIC declined. And that was not good for the stocks. I was really worried about that. I do think
that those data points are important in terms of suggesting that maybe we'll be able to navigate
this potential air gap and ROIC. Yeah. It makes me wonder about in this market, I'm like everybody
else. It's the 10 companies at the top that are all the market cap more than all of the attention.
There's 49 of the other companies, the S&P 500, you study those too. Like, what do you think about
that group? Like, what is interesting to you about the group that now nobody seems to talk about?
And no one really seems to care about because they haven't driven returns and they're a smaller
percent of the overall index. I think that people are going to start to care if you have more and more
companies, print these C/H Robinson like quarters. I think the companies that have historically been
really well-run, the reason they have a long track record of success, you cannot succeed without
using technology well. And so if you have a kind of internal culture of experimentation and innovation,
I think you will do well with AI. I would bet on the best investment banks to be earlier and
better adopters of AI than maybe some of the trailing banks, just sometimes past his prologue.
And I think it's likely to be in this case. One strong opinion I have, all these VCs are sitting
up these holding companies and we're going to use AI to make traditional businesses better.
And the really smart VCs and their great track records, but that's what private equity has been
doing for 50 years. You're just not going to be private equity at their game.
That's what Vista did in the early days, right? Yeah, and I do think this is actually private
equities, maybe had a little bit of a tough run, just multiples have gone up. Now private assets
are more expensive. The cost of financing has gone up. It's tough to take a company public,
because the public valuation is 30% lower than the private valuation. So P's had a tough run.
I actually think these private equity firms are going to be pretty good at systematically applying AI.
We haven't spent much time talking about meta and traffic or open AI. And I'd love your impression
on everything that's going on in this infrastructure side that we talked about. These are three really
important players in this grand game. How does all of this development that we've discussed
so far impact those players specifically? First thing, let me just say about frontier models broadly.
Yeah. In 2023 and 24 hours, fond of quoting Eric Vistria, and Eric Vistria's statement, our friend,
brilliant man, and Eric would always say foundation models of the fastest appreciating assets in history.
And I would say he was 90% right. I modified the statement. I said foundation models without
unique data and internet scale distribution are the fastest appreciating assets in history.
And reasoning fundamentally changed that in a really profound way. So there was a loop, a flywheel
to quote Jeff Bezos, that was at the heart of every great internet company. And it was you made a good
product, you got users, those users using the product generated data that could be fed back into
the product to make it better. And that flywheel has been spinning at Netflix, at Amazon, at Meta, at Google
for over a decade. And that's an incredibly powerful flywheel and it's why those internet businesses
were so tough to compete with. It's why they're increasing returns to scale. Everybody talks about
network effects. They were important for social networks. I don't know to what extent Meta is a
social network anymore like a content distribution, but they just had increasing returns to scale because
of that flywheel. And that dynamic was not present in the pre-reasoning world of AI. You pre-trained the
model. You let it out in the world. And it was what it was. And it was actually pretty hard. They would
do RLHF reinforcement learning with human feedback. And you try and make the bot model better. And
maybe you'd get a sense from Twitter vibes that people didn't like this. And so you tweak it.
They're the little up and down arrows. But it was actually pretty hard to feed that back into the
model. With reasoning, it's early, but that flywheel started to spin. And that is really profound
for these frontier labs. So one, reasoning fundamentally changed the industry dynamics of frontier
labs. And just explain why specifically that is like what is going on. Because if a lot of people
are asking a similar question, they're consistently either liking or not liking the answer,
then you can kind of use that like that has a verifiable reward that's a good outcome. And then
you can kind of feed those good answers back into the model. And we're very early at this flywheel
spinning. Yeah, got it. Like it's hard to do now. But you can see it beginning to spin.
So this is important fact number one for all of those dynamics. Second, I think it's really important
that meta, you know, Mark Zuckerberg at the beginning of this year in January said I'm highly confident
I'm going to get the quote wrong that at some point in 2025, we're going to have the best and most
performant AI. I don't know if he's in the top 100. So he was as wrong as it was possible to be.
And I think that is a really important fact. Because it suggests that what these four companies have
done is really hard to do. Because meta through a lot of money at it. And they failed. Yall McCune had to
leave. They had to have the famous billion dollar for AI researchers. By the way, Microsoft also failed.
They did not make such an unequivocal prediction, but they bought an inflection AI. And there were a lot
of comments from them that we anticipate our internal models quickly getting better and we're going to run
more and more of our AI on our internal models. Nope. Amazon. They bought a company called the Dept AI.
They have their models called Nova. I don't think they're in the top 20. So clearly it's much harder
to do than people thought a year ago. And there's many, many reasons for that. Like it's actually
really hard to keep a big cluster of GPUs coherent. A lot of these companies were used to running their
infrastructure to optimize for cost instead of performance. Complexity and performance. Complexity and
keeping the GPUs running at high utilization rate and a big cluster. It's actually really hard. And
there are wild variations in how well companies run GPUs. If the most anybody because the laws of
physics, you know, maybe you can get two or three hundred thousand black wells, go ahead and we'll see.
But if you have 30% uptime on that cluster and you're competing with somebody who has 90% uptime,
you're not even competing. So one, there's a huge spectrum on how well people run GPUs.
Two, then I think there is, you know, these AI researchers they like to talk about taste. I find it
very funny. You know, why do you make so much money? I have very good taste. You know, what taste
means is you have a good intuitive sense for the experiments to perform. This is why you pay people
a lot of money because it actually turns out that as these models get bigger, you can no longer run
an experiment on a thousand GPU cluster and replicate it on a hundred thousand GPUs. You need to run that
experiment on 50,000 GPUs and maybe it takes, you know, days. And so there's a very high opportunity cost.
You have to have a really good team that can make the right decisions about which experiments to run
on this. And then you need to do all the reinforcement learning during post-training well and test time
and compute well. It's really hard to do and everything's it's easy, but all those things, you know, I used to have the saying, like,
I was retelling on a slow go, pick any vertical in America. If you can just run a thousand stores in 50 states
and have them clean, well lit, stocked with relevant goods at good prices and staffed by friendly employees
who are not stealing from you, you're going to be a 20 billion dollar company, a 30 billion dollar company.
Like 15 companies have been able to do that. It's really hard. And it's the same thing. Doing all of these things well
is really hard. And then reasoning with this flywheel, this is beginning to create barriers to entry
and what's even more important, every one of those labs, X AI, Gemini, Open AI and Anthropic, they have a more advanced
checkpoint internally of the model. Checkpoint is just kind of continuously working on these models
and then you release kind of a checkpoint. And then the reason these models get fast. The one they're
using internally is for sure. And they're using that model to train the next model. And if you do not have
that latest checkpoint, it's getting really hard to catch up. Chinese open source is a gift from God
to matter because you can use Chinese open source. That can be your checkpoint. And you can use that
as a way to kind of bootstrap this. And that's what I'm sure they're trying to do and everybody else.
The big problem and the big giant swing factor, I think China has made a terrible mistake with this
rarest thing. So China, because you know, they have Huawei is in and it's a decent chip versus
something like the deprecated hoppers every science looks okay. And so they're trying to force Chinese
open source to use their Chinese chips. They're domestically designed chips. The problem is Blackwell's
going to come out now in the gap between these American frontier labs in Chinese open source is
going to blow out because of Blackwell. And actually deep seek in their most recent technical paper,
V3.2 said one of the reasons we struggled to compete with the American frontier labs is we don't
have enough compute. That was their very politically correct. Still a little bit risky way of saying
because China said we don't want the Blackwell's right. And they're saying, would you please give
us some light? That might be a big mistake. So if you just kind of play this out,
these four American labs are going to start to widen their gap for Chinese open source.
Which then makes it harder for anyone else to catch up because that gap is growing. So you can't
use Chinese open source to bootstrap. And then geopolitically China thought they had the leverage.
They're going to realize, oh, whoopsy daisy. We do need the Blackwells. And unfortunately,
they'll problem for them. They'll probably realize that in late 26. And at that point, there's
an enormous effort underway DARPA has there's also to really cool DARPA and DOD programs to incentivize
really clever technical solutions for rare earths. And then there's a lot of rare earth deposits
in countries that are very friendly to America that don't mind actually refining it in the
traditional way. So I think rare earths are going to be solved way faster than anyone thinks.
You know, they're obviously not that rare. They're just misnamed. They're rare because they're
really messy to refine. And so geopolitically, I actually think Blackwell's pretty significant.
And it's going to give America a lot of leverage as this gap widens. And then in the context of
all of that, going back to the dynamics between these companies, XAI will be out with first Blackwell
model. And then they'll be the first ones probably using Blackwell for inference at scale. And I think
that's an important moment for them. And by the way, it is funny. Like if you go on OpenRouter,
you can just look. They have dominant share now OpenRouter is whatever it is. It's 1% of API tokens.
But it's an indication they process 1.35 trillion tokens. Google did like 800 or 900 billion. This
is like whatever news last seven days or last month. Anthropic was at 700 billion. Like XAI is doing
really, really well. And the model is fantastic. I highly recommend it. But you'll see XAI come out
with this. OpenAI will come out faster. OpenAI's issue that they're trying to solve a stargate
is because they pay a margin to people for compute. And maybe the people who run their compute are
not the best at running GPUs. They're a high cost producer of tokens. And I think this kind of
explains a lot of their code red recently. Yeah. Look, the $1.4 trillion spending commitments.
And I think that was just like, hey, they know they're going to need to raise a lot of money.
Particularly if Google keeps its current strategy of sucking the economic oxygen out of the room.
And you know, you go from 1.4 trillion rough vibes code red like pretty fast, you know. And the reason
they have code red is because of all these dynamics. So then they'll come out with a model. But they
will not have fixed their per token cost disadvantage yet relative to both XAI and Google and
Anthropic at that point. Anthropic is a good company. You know, they're burning dramatically less cash
than OpenAI and growing faster. So I think you have to give Anthropic a lot of credit. And a lot of
that is their relationship with Google and Amazon for the TPUs and the trainings. And so Anthropic has
been able to benefit from the same dynamics that Google has. I think it's very indicative in this
great game of chess. You can look at Dario and Jensen maybe have taken a few public comments, you know,
that we're between them. Just a little bit of jousting. Well, Anthropic just signed the $5 billion
deal with Nvidia. That is because Dario is a smart man and he understands these dynamics about Black
Wall and Ruben relative to TPU. So Nvidia now goes from having two fighters XAI and OpenAI to three
fighters that helps in this Nvidiaverse Google model. And then if meta can catch up, that's really
important. I am sure Nvidia is doing whatever they can to help meta. You're running those GPUs this
way. Maybe we should twist the screw this way or turn the dial that way. And then it will be also if
Blackwell comes back to China, which it seems like it'll probably happen. That will also be very good
because the Chinese open source will be back. I'm always so curious about the polls of things.
Someone poll would be the other breakthroughs that you have your mind on. Things in the data center
that aren't chips that we've talked about before as one example. I think the most important thing
that's going to happen in the world, in this world, in the next three to four years is data centers
in space. And this has really profound implications for everyone building a power plant or a data center
on planet Earth. Okay. And there is a giant gold rush. I've heard anything about this. Please.
Yeah, you know, it's like everybody thinks like, hey, AI is risky. But you know what? I'm going to build
the data center. I'm going to build a power plant that's going to do a data center. We will need that.
But if you think about it from first principles, data centers should be in space. What are the
fundamental inputs to running a data center? Their power and their cooling. And then there are the chips.
If you think about it from a total cost perspective. Yeah. And just the inputs to making the tokens
come out of the magic machines in space, you can keep a satellite in the sun 24 hours a day.
And the sun is 30% more intense. You can have the satellite always kind of catching light.
The sun is 30% more intense in this results in six times more irradiance and outer space
than on planet Earth. So you're getting a lot of solar energy. Point number two,
because you're in the sun 24 hours a day, you don't need a battery. And this is a giant percentage
of the cost. So the lowest cost energy available in our solar system is solar energy and space.
Second, for cooling, and one of these racks, a majority of the mass in the weight is cooling.
The cooling in these data centers is incredibly complicated. The HVAC, the CDUs,
the liquid cooling, it's amazing to see. And space, cooling is free. You just put a radiator
on the dark side of the satellite. It's fun. It's as close to absolute zero as you can get.
So all that goes away. And that is a vast amount of cost. Okay. Let's think about how these
maybe each satellite is kind of a rock. It's one way to think of it. Maybe some people make bigger
satellites that are through VACs. Well, how are you going to connect those racks? Well, it's
funny in the data center. The racks are over a certain distance connected with fiber optics.
And that just means a laser going through a cable. The only thing faster than a laser going
through a fiber optic cable is a laser going through absolute vacuum. So if you can link these satellites
in space together using lasers, you actually have a faster and more coherent network
than in a data center on Earth. Okay. For training, that's going to take a long time because it's
so big. Yeah, just because it's so big, training will eventually happen. But then for inference,
let's think about the user experience. When I asked rock about you and it gave the nice answer,
here's what happened. A radio wave traveled from my cell phone to a cell tap.
Then it hit the base station, went into a fiber optic cable, went to some sort of metro aggregation
facility in New York, probably within like 10 blocks a year. There's a small little metro router
that's routed those packets to a big XAI data center somewhere. The computation was done and it came
back over the same path. If the satellites can communicate directly with the phone and
Starlink has demonstrated direct to cell capability, you just go boom, it's a much better lower cost
user experience. So in every way data centers in space from a first principles perspective are
superior to data centers on Earth. If we could teleport that into existence, I understand that
that portion, why will that not happen? Is it launch cost? Is it launch ability? Is it capacity?
We need a lot of the starships, like the starships are the only ones that can economically make that
happen. We need a lot of those starships. Maybe China or Russia will be able to land a rocket,
Blue Origin just landed a booster. This is a big idea. And I do think it's an entirely new and
different way to think about SpaceX. And it is interesting that Elon posted your set in an
interview that Tesla SpaceX and XAI are converging and they really are. So XAI will be the intelligence
module for Optimus made by Tesla with Tesla Vision has its perception system. And then SpaceX will have
the data centers in space that will power a lot of the AI, presumably for XAI and Tesla and the
Optimus and a lot of other companies. And it's just interesting the way that they're converging
in each one is kind of creating competitive advantage for the other. If you're XAI, it's really nice.
Do you have this built-in relationship with Optimus? Tesla's a public company, so there's going to be like,
I cannot imagine the level of vetting. They'll go into that. It's your company agreement. And then
you have a big advantage with these data centers in space. And then it's also nice if you're XAI
that you have two companies with a lot of customers who you can use to help build your customer
support agents, your customer sales agents with kind of built-in customers. So they really are all
kind of converging in a neat way. And I do think it's going to be a big moment when that first black
well model comes out from XAI next year. If I go to the other end of the spectrum and I think about
something that seems to have been historically endemic to the human economic experience that shortages
are always followed by gluts in capital cycles. What if in this case the shortages compute like
Mark Chen now is on the record is saying they would consume 10X as much compute if you gave it to
him in like a couple weeks. So like there seems to still be a massive shortage of compute, which is
all the stuff we've talked about today. But there also just seems to be this like iron law of history
that gluts follow shortages. What do you think about that concept as relates to this?
There will eventually be a glut. AI is fundamentally different than the software just in that every time
you use AI takes compute in a way that traditional software just did not. I mean it is true. Like I think
every one of these companies could consume 10X more compute. Like what would happen is just the $200
tier would get a lot better. The free tier would get like the $200 tier. Google has started to monetize
AI mode with ads. And I think that will give everyone else permission to introduce ads into the
free mode. And then that is going to be an important source of ROI. Open AI is tailor made to
absolutely all of them in actions like, you know, hey, here are your three vacations. Would you like
me to book one? And then they're for sure going to collect a commission. Yeah. There's many ways you
can make money. But I think we went into great detail on maybe a prior podcast about how just
inventory dynamics made these inventory cycles inevitable in semis. The iron law of semis is just
customer buffer inventories have to equal lead times. And that's why you got these inventory cycles
historically. We haven't seen a true capacity cycle in semis arguably since the late 90s and that's
because Taiwan semis has been so good at aggregating and smoothing supply. And a big problem in the
world right now is that Taiwan semis not expanding capacity has fastest their customers want. They're
in the process of making mistakes just because you do have intel and with these fabs, they're not
has good and it's really hard to work with their PDK. But now you have this guy, Lee Bo, who's a really
good executive and really understands that business. I mean, by the way, Patrick Helsinger, I think
was also good executive and he put intel on the only strategy that could result in the success. And
I actually think it's shameful that the intel board fired him when they did. But Lee Bo is a good
executive and now he's reaping the benefits of Patrick's strategy. And Intel has all these empty
fabs and eventually given the shortages we have of compute, those fabs are going to be filled.
So I think Taiwan semis in the process of making mistakes, but they're just so paranoid about an
overbuilt and they're so skeptical. They're the guys who met with Sam Altman and laughed and said he's
a podcast bro. He has no idea what he's talking about. You know, they're terrified of an overbuilt.
So it may be that Taiwan semis single handedly that their caution is the governor. I think governors
are good. It's good that powers the governor. It's good that Taiwan semis is a governor.
If Taiwan semi opens up at the same time when data centers in space relieve all power constraints,
but that's like, I don't know, five, six years away. The data centers in space are a majority of
deployed megawatts. Yeah, you can get it overbuilt really fast, but just we have these two really powerful
natural governors. And I think that's good. Smoother and longer is good. We haven't talked about
the power of the than alluding to it through the space thing. Haven't talked about power very much.
Power was like the most uninteresting topic. Nothing really changed for like a really, really long time.
All of a sudden, I'm trying to figure out how to get like gigawatts here and everywhere.
How do you think about are you interested in power?
Oh, very interested. I do feel lucky in a prior life. I was the sector leader for the telecom and
utilities team. So I do have some base level of knowledge. Having watts as a constraint is really good
for the most advanced compute players, because if watts are the constraint, the price you pay for
computers are relevant. The TCO of your compute is absolutely irrelevant, because if you can get three
X or four X or five X more tokens per watt, that is literally three or four X or five X more revenue.
If you're going to build an advanced data center cost 50 billion, a data center with the ASIC
maybe costs 35 billion. If that 50 billion dollar data center pumps out 25 billion of revenue,
and you're the ASIC data center at 35 billion is only pumping out eight billion. Well, like you're
pretty buffed. So I do think it's good for all of the most advanced technologies of the data center,
which is exciting to be as an investor. So as long as power is a governor, the best products
are going to win irrespective of price and have crazy pricing power. That's the first application
that's really important to me. Second, it is in the only solutions to this. We just can't build
nuclear fast enough in America. As much as we would love to build nuclear quickly, we just can't.
It's just too hard. NEPA, all these rules, like it's just, it's too hard. Like a rare ant
that we can move and it could be in a better environment can totally delay the construction
of a nuclear power plant. You know, one ant, you know, that is emergency. That's crazy. Actually,
humans need to come first. We need to have a human-centric view of the world. But the solutions
are natural gas and solar. And the great thing is, the great thing about these AI data centers
is apart from the ones that you're going to do it. For it's on, you can locate them anywhere.
So I think you were going to see, and this is why you're seeing all this activity at Abilene,
because it's in the middle of a big natural gas basin. And we have a lot of natural gas in America,
because of fracking. We're going to have a lot of natural gas for a long time. We can
rip production really fast. So I think this is going to be solved. You know, you're going to have
power plants fed by gas or solar. And I think that's the solution. And you know, you're already
all these turbine-made factors were reluctant to expand capacity, but caterpillar just said,
we're going to increase capacity by 75% over the next few years. So like, this system on the
power side is beginning to respond. One of the reasons that I always still love talking to you is
that you do as much in the top 10 companies in the world as you do looking at brand new companies
with entrepreneurs that are 25 years old trying to do something amazing. And so you have this
very broad sense of what's going on. If I think about that second category of young
enterprising technologists who now are like the first generation of AI native entrepreneurs,
what are you seeing in that group that's notable or surprising or interesting?
These young CEOs, they're just so impressive in all ways. And they get more polished faster.
And I think the reason is, is they're talking to the AI. How should I deal with pitching this investor?
I'm meeting with Patrick O'Shaughnessy. What do you think the best ways I should pitch him are?
And it works. Do a deep research thing. And it's good. Yeah. You know, hey, I have this difficult HR
situation. How would you handle it? And it's good at that. We're struggling to sell our product.
What changes would you make? And it's really good at all of that today. And so in that goes to these
VCs are seeing massive AI productivity in all their companies. It's because their companies are
full of these 23, 24 or even younger AI natives. And they're impressive. I've been so impressed
with like young investment talent. And it's just part of it. Like your podcast is part of that.
They're just very specific knowledge has became so accessible through podcasts in the internet.
Like kids come in and they're just, I feel like they're where I was as an investor. Like in my,
you know, early 30s and they're 22 and I'm like, oh my god, like I have to run so fast to keep up.
These kids who are growing up native in AI, they're just proficient with it in a way that I haven't
tried really hard to become. Can we talk about semis VCs specifically and like what is interesting
in that universe? What do you guys just think is so cool about it and so underappreciated? Is your
average semiconductor venture founder is like 50 years old? Okay. And Jensen, and what's happened
with Nvidia and the market cap of Nvidia has like single handedly ignited semiconductor venture.
But the way it's ignited is ignited in an awesome way. It's like really good for actually
Nvidia and Google and everyone is like, let's just say you're the best DSP architect in the world.
You had made for the last 20 years every two years because that's what you have to do.
Semiconductors, it's like every two years you have to run a race. And if you won the last race,
you start like a foot ahead and over time those compound and make each race easier to win.
But maybe that person and his team, maybe he's ahead of networking at a big public company.
And he's making a lot of money and he has a good life and he's 50 years old. And then because he
sees these outcomes and the size of the markets and the data center, he's like, wow, I don't know,
I just go start my own company. But the reason that's important is that, you know, I forget the number,
but I mean, there's thousands of parts in a black well rack and there's thousands of parts in a TPU
rack. And in the black well rack, maybe Nvidia makes two or three hundred of those parts,
same thing in an AMD rack. And they need all of those other parts to accelerate with them.
So they couldn't go to this one year cadence.
If everything was not keeping up with them. So I think it's the fact that semiconductor
ventures come back with the vengeance, Silicon Valley, stopping Silicon Valley long ago.
My little firm, maybe has done more semiconductor deals in the last seven years than the top 10 VCs
combined. You know, but that's really, really important because now you have an ecosystem of
companies who can keep up. And then that ecosystem of these venture companies is putting pressure
on the public companies that are also need to part of this. And we're going to go to this annual
cadence, which is just so hard. But it's one reason I'm really skeptical of these ASICs
that don't already have some degree of success. So I do think that's a super, super important
dynamic. And one that's absolutely foundational and necessary for all of this to happen.
Because not even Nvidia can do it alone. AMD can't do it alone. Google can't do it alone.
You need the people who make the transceivers, you need the people who make the wires,
who make the backplanes, who make the lasers. They all have to accelerate with you.
And one thing that I think is very cool about AI as an investor is it's just it's the first time
where every level of the stack that I look at at least the most important competitors are public
and private Nvidia. They're very important private competitors, Broadcom important
private competitors, Marvel, important private competitors. You know, Lumiddon,
coherent, all these companies. There's even like a wave of innovation and memory,
which is really exciting to see because memory is such a gaining factor. But always something
that could slow all this down and be a natural governor is if we get our first true DRAM cycle
since the late 90s. So to say more of that means if the price of DRAM, if like DRAM wafer is valued at
like a five-karat diamond in the 90s when he had these true capacity cycles before Taiwan,
so we kind of smoothed everything out and DRAM became more of an oligopoly. You would have these
crazy shortages where the price would just go 10x, unimaginable relative to the last 25 years,
where like a giant DRAM cycle, a good DRAM cycle is the price starts to start,
stops going down. An epic cycle is maybe it goes up, you know, whatever it is, 30, 40,
50%. But I mean, if it starts to go up by X's instead of percentages, that's a whole different game.
By the way, we should talk about SAS. Yeah, let's talk about it. What do you think is going to happen?
Well, I think that applications, SAS companies are making the exact same mistake that brick-and-mortar
retailers did with e-commerce. So brick-and-mortar retailers, particularly after the telecom bubble crashed,
you know, they looked at Amazon and they said, "Oh, it's losing money." e-commerce is going to be a low
margin business. How can it ever be more efficient as a business? Right now, our customers
pay to transport themselves to the store and then they pay to transport the goods out. How can it
ever be more efficient if we're sending shipments out to individual customers? Amazon's vision,
of course, will eventually we're just going to do it out of street and drop off a package at every
house. And so they did not invest in e-commerce. They clearly saw customer demand for it,
but they did not like the margin structure of e-commerce. That is the fundamental reason
that essentially every brick-and-mortar retailer was really slow to invest in e-commerce. And now,
here we are, and, you know, Amazon has higher margins, and they're North American retail business,
then a lot of retailers that are mass market retailers. So margins can change. And if there's a
fundamental transformative kind of new technology that customers are demanding, and it's always a
mistake not to embrace it. And that's exactly what the SaaS companies are doing. They have their 70,
80, 90% gross margins, and they are reluctant to accept AI gross margins. The very nature of AI is,
you know, software you write at once, and it's written very efficiently, and then you can distribute
it broadly at very low cost. And that's why it was a great business. AI is the exact opposite,
where you have to recompute the answer every time. And so a good AI company might have gross margins
of 40%. The crazy thing is, because of those efficiency gains, they're generating cash way earlier
than SaaS companies did historically, but they're generating cash earlier, not because they have
high gross margins, but because they have very few human employees. And it's just tragic to watch
all of these companies. Like, you want to have an agent? It's never going to succeed if you're not
willing to run it at a sub 35% gross margin, because that's what the AI natives are ready to get at.
Maybe they're ready in a 40. So if you're trying to preserve an 80% gross margin structure,
you are guaranteed that you will not succeed in AI. Absolute guarantee. And this is so crazy to me,
because one, we have an existence proof for software investors being willing to tolerate gross margin
pressure, as long as gross profit dollars are okay. It's called the cloud. People don't remember,
but when Adobe converted from on-premise to a SaaS model, not only did their margins implode,
they're actually revenues imploded too, because you went from charging up front to charging over
a period of years. Microsoft, it was less dramatic, but Microsoft was a tough stock in the early days
of the cloud transition, because investors were like, "Oh my God, you're an 80% gross margin
business." And the cloud is the 50s, and they're like, "Well, it's going to be gross profit dollar
creative, and probably will improve those margins over time." Microsoft, they bought GitHub,
and they used GitHub as a distribution channel for Copilot for Coding. Let's become a giant business,
a giant business. Now for sure, it runs at much lower gross margins, but there are so many SaaS
companies. I can't think of a single application SaaS company that could not be running a successful
agent strategy. And they have a giant advantage over these AI natives, and that they have a cash
alternative business. And I think there is room for someone to be a new kind of activist or
constructivist and just go to SaaS companies and say, "Stop being so dumb. All you have to do is say,
"Here are my AI revenues, and here are my AI gross margins." And you know it's real AI, because it's
low gross margins. I'm going to show you that. And here's a venture competitor over here that's
losing a lot of money. So maybe I should take my gross margins to zero for a while, but I have this
business that the venture funded company doesn't have. And this is just such a like obvious
playbook that you can run Salesforce, ServiceNow, HubSpot, GitLab, Alacian, all of them could run this.
And the way that those companies could or should think about the way to use agents is just to
ask the question, "Okay, what are the core functions we do for the customer now? Like, how can we
further automate that with agents effectively?" If you're in CRM, well, what our customers do,
they talk to their customers. We're customer relationship management software, and we do some
customer support too. So make an agent that can do that. And sell that at 10 to 20% to let that
agent access all the data you have. Because what's happening right now, another agent made by
someone else is accessing your systems to do this job, pulling the data into their system, and then
you will eventually be turned off. It's just crazy. And it's just because, oh, wow, but we want to
preserve our 80% gross margins. This is a life for death decision. And essentially, everyone except
Microsoft is failing it to quote that bimbo from that Nokia guy long ago, like their platforms are
burning platform. Yeah, there's a really nice platform right over there, and you can just hop to it,
and then you can put out the fire in your platform that's on fire. And now you got two platforms,
and it's great. You know, your data centers and space thing makes me wonder if there are other kind
of less discussed off the wall, things that you're thinking about in the markets in general,
that we haven't talked about. It does feel like since 2020 kicked off in 2022, punctured this.
Kind of a series of rolling bubbles. So in 2020, there is a bubble in like EV startup EVs that
were not Tesla. And that's for sure a bubble, and they all went down, you know, 99% and there was kind
of a bubble in more speculative stocks than we had the meme stocks, you know, game stop. And now it
feels like the rolling bubble is in nuclear and quantum. And these are fusion and SMR. It would be a
transformative technology. It's amazing. But sadly, from my perspective, none of the public ways you
can invest in this are really good expressions of this theme are likely to succeed or have any real
fundamental support. And same thing with quantum. I've been looking at quantum for 10 years. We have
a really good understanding of quantum. And the public quantum companies again are not the leaders.
From my perspective, the leaders in quantum would be Google, IBM, and then the Honeywell quantum.
So the public ways you can invest in this theme, which probably is exciting, are not the best. So
you have two really clear bubbles. I also think quantum supremacy is very misunderstood. People hear
it. And they think that it means that quantum computers are going to be better than classical computers
at everything with quantum. You can do some calculations that classical computers cannot do.
That's it. That's going to be really exciting and awesome. But it doesn't mean that quantum
takes over the world. I think the thought that I have had this is maybe less related to markets
than just AI. I have just been fascinated that for the last two years, whatever AI needs to keep
growing and advancing, it gets. Have you ever seen public opinion change so fast in the United States
on any issue has nuclear power? Just happen like that. Like that. And like, why did that happen
right when AI needed it to happen? Now we're running up on boundaries of power and earth. All of a
sudden, data centers in space, it's just a little strange to me that whenever there is something,
a bottleneck that might slow it down, everything accelerates. Ruben is going to be such an easy
seamless transition relative to Blackwell and Ruben's a great chip. And then AMD getting into the
game with the MI450, whatever AI needs, it gets. You're a deep reader of sci-fi. So yeah, exactly.
You're making me think of Kevin Kelly's great book, What Technology Wants. He calls it the
Technium, like the like the overall mass of technology that just like is supplied by humans.
Absolutely more powerful. Yes, it's just more powerful. And now we're going into an in-state.
You founded a trade ease in the pre-Chatchy PT and pre-COVID era. The world has changed dramatically.
How has a trade ease changed the most in that same period?
Before a trade ease, I had never had a big position where it wasn't my position. I was doing the fundamental
work. I was the analyst. I was really a one-man show. It was such a crazy feeling. The first time
I made money on an idea that was not my own. And that was CrowdStrike in 2020. And then it never happened
to me. Ever. And I'd run funds for 15 years at that point. I think that was a big evolution for me.
I've been thinking a lot about actually the NFL in the context of investing. I think it is so
interesting. Sam Darnold, Baker Mayfield, Daniel Jones, left for dead, utter failures,
embarrassment of draft picks. Okay. And now there's some of the best quarterbacks in the league.
It just turned out they needed a different system and a different coaching mindset. I have really been
trying to work on it. How I can make sure that if there is someone who is clearly talented
and working really hard, how do I make sure that if I have that person, they succeed. And I just think
investing is all about finding the right balance between the courage of your convictions and the
flexibility to admit when you're wrong to quote Michael Steinhardt. It's all about finding the right
balance between the arrogance to believe that you have a varied view that's going to be right
in a game that is being played of tens of millions of people worldwide with the humility to recognize
that at any moment, there might be a new data point that is outside of your expected probability
space that invalidates that varied view and to be really, really open to that. And I think a lot
of that comes down to finding a investment philosophy and process that fits your own emotional
makeup so you can be rational when you're wrong and you could strike that balance between conviction
and flexibility. What every investment firm has to find is the right balance between incentivizing
risk-taking and accountability from mistakes. If you don't have enough accountability, people take
way too much risk. On the other hand, if you have too much accountability, people don't take risk.
What I have tried to do and really, really institute is make it really safe for people to change their
mind. I want to make it safe for people to say, here's some new data points that have invalidated
my investment hypothesis. I'm the child of two attorneys. I've never taken an argument personally.
I love to argue. Okay, it's like I love it. Like it's just like plus board and I can do it all day.
But I do believe at some level kind of investing is the search for truth. And if you find truth
first and you write about it big of truth, that's how you generate alpha. And it has to be a truth
that other people have not yet seen. You're searching for hidden truths. I think that the best way to
arrive at truth is through discourse. And so I try to really incentivize people to tell me that I'm
wrong, celebrate it when people tell me I'm wrong, having what we call bull bear lunches, where the
aside analyst presents their case, and then the skeptic presents their case, and there's a very robust
discussion that I generally try not to participate in. I would say decision that takes an
amidst amount of work, but decision qualities high after that. I want people to feel safe,
taking risk. I want people to feel safe changing their mind. And then I want people to feel extremely
safe, telling me that I'm wrong and telling each other that they're wrong. Because the more we have
that, like, definitionally, if you were not wrong, you're not learning. If I'm not wrong about
three things in a day, I didn't learn anything. So I want to be told I'm wrong as much as possible.
And so I try to incentivize all of that. I have a selfish closing question speaking of young people.
So my kids who are 12 and 10, but especially my son who's older, is developing an interest in what
I do, which I think is quite natural. And I'm going to try to start asking my friends who are the
most passionate about entrepreneurship and investing. Why they are so passionate about it. And what about
it is so interesting and life-giving to them? How would you pitch what you've done the career you
built this part of the world to a young person that's interested in this? The earliest thing I can
remember is being interested in history, looking at books with pictures of the Phoenicians and the
Egyptians and the Greeks and the Romans. I loved history. I vividly remember, I think in the
second grade, as my dad drove me to school every day, we went through the whole history of World War
2 in one year. And I loved that. And then that translated into a real interest in current events
very early. So like has a pretty young person, you know, I don't know if it was eighth grade or seventh
grade or ninth grade. I was reading The New York Times and The Washington Post and I would get so
excited when the mail came because it made that maybe there was an economist or newsweek or a time
or a US news. And I was really into current events. You know, because current events is kind of like
applied history and watching history happen. Thinking about what might happen next. I didn't know
anything about investing. My parents were both attorneys. Like I was anytime I wanted argument,
I was super rewarded. Like, you know, if I could make a reasonable argument why I should stay
up late. My parents would be so proud and they'd let me simply, but I had to beat them. You know,
that was the way I grew up. I was just kind of going through life and I really loved to ski and I
loved rock climbing. And I go to college and rock climbing is by far the most important thing in my
life. I dedicate myself to them completely. I climbed. I did all my homework at the gym. I got to
the rock climbing gym at 7 a.m. Would skip a lot of classes to stay at the gym. I do my homework on
like a big bouldering mat every weekend. I went and climbed somewhere with the Dartmouth Mount
Nearing Club. It was super important to me. I'm not a good athlete. So I was never a very good
climber, but I did dedicate myself entirely to rock climbing. And as part of that, like on
climbing trips, the movie routers came out while I was in college. So we started playing poker.
I like to play chess. I mean, I was never that good at chess or poker. You know, never really
dedicated myself to either. And my plan after two or three years of college was I was going to leave.
I was a ski bum at Alton College. I was a housekeeper. I cleaned a lot of toilets. It was shocking to
me how people treated me. And it is like permanently impacted how I treat other people. You know,
you'd be cleaning somebody's room and they'd be in it. And they'd be reading the same book as you.
And like, you know, you'd say, Oh, that's a great book. You know, I'm about where you are.
And they look at you like you're basically like you speak. And did they get even more shocked?
You read. You know, so it like had a big impact on how I've like just treated everyone since then.
Like being nice is free. But anyways, I was going to be a ski bum in the winter. I was going to work
on a river in the summers. And that was how I was going to support myself. And then I was going to
climb in the shoulder seasons. I was going to try and be a wildlife photographer and write the next
great American. Wow. I can't believe I knew this. That was my plan. This was like my plan of record.
I was really lucky. My parents were very supportive of everything I wanted to do. My parents had very
strict parents. So of course, they're extremely permissive with me. So, you know, I'll probably end up
being a strict parent. You know, just the cycle continues. And my parents were lawyers. You know,
they've done reasonably well. They both grew up and I would say very economically disadvantaged
circumstances. Like my dad talks about like he remembers every person who bought him a beer.
Are you going to afford a beer? He worked the whole way through college. He was there on a scholarship.
He had one pair of shoes all through high school. And so they were super on board with this plan.
And I'd been very lucky. They sent me to college. And I didn't have to pay for college. So they said,
you know, Gavin, we think this plan of being a ski bum, riverafety guide, wildlife photographer,
climbing the shoulder seasons, tried to write it all, but we think it sounds like a great plan.
But you know, we've never asked you for anything. We haven't encouraged you to study anything. We've
supported you and everything you've wanted to do. Will you please get one professional internship?
Just one. And we don't care what it is. The only internship I could get, this was at the end of
myself our summer at Dartmouth. It was an internship with Donaldson, Loughkin, Engineer at DLJ.
My job was to, every time DLJ published a research report, it was in like the private wealth
management division. And I worked for the guy who ran the office. And my job was whenever they
produced a piece of research, I would go through and look at which of his clients owned that stock,
then I would mail it to the clients. So this day, we wrote on General Electric. So I need to mail
the GE report to these 30 people, mail the Cisco report to these 20 people. And then I started
like reading the reports and I was like, oh my god, this is like the most interesting thing imaginable.
So investing, I kind of conceptualized it. It's a game of skill and chance kind of like poker.
And you know, there's obviously chance in investing. Like if you're an investor in a company and a
meteor hits their headquarters, that's bad luck. But like you own that outcome. So there is chance
that is irreducible, but there's skill too. But so that really appealed to me. And the way you got an
edge in this greatest game of skill and chance imaginable was you had the most thorough knowledge
possible of history. And you intersected that with the most accurate understanding of current
events in the world to form a differential opinion on what was going to happen next in this game
of skill and chance, which stock is mispriced in the period mutual system. That is the stock market.
And that was like day three. I went to the bookstore and I bought like the books that they had,
which were penal inches books. I read those books in like two days. I'm a very fast reader.
Then I read all these books about Warren Buffett. Then I read Market Wizards. Then I read Warren
Buffett's letters to shareholders. This is like during my internship. Then I read Warren Buffett's
letters to shareholders again. Then I taught myself accounting. There's this great book,
why stocks go up and down. Then I went back to school. I changed my majors from English and history
to history and economics. And I never looked back in it consumed. Like I continued to really focus
on climbing, but like instead of like I would be in the gym and I would print out everything that
the people on the Motley Fool wrote. They had these fools. They were early to talking about return
on invested capital incremental ROIC is like a really important indicator. And I would just read it
and I would underline it. And I read books. And then I'd read the Wall Street Journal. And then
eventually there was a computer term finally set up near the gym. And I'd go to that gym and just
read news about stocks. And it was the most important thing in my life. And like I barely kept my
grades up. And yeah, that's how I got into it. History, current events, skill and chance.
And I am a competitive person. And I've actually never been good at anything else. Okay. I got
picked last for every sports team. Like I love to ski. I've literally spent a small fortune.
I'm private ski lessons. I'm not that good of a skier. I like to play ping pong. All my friends
could beat me. I tried to get really good at chess. And this was before we actually had to play the
games. And my goal was to beat one of the people. I'm sure there's a park somewhere. Right there.
Most famous one is right there. Okay. Well, there's one in Cambridge. And I wanted to beat one of them.
Never beat one of them. I've never been good at anything. I thought I would be good at this.
And the idea of being good at something other than taking a test that was competitive was very
appealing to me. That's why I think that's been a really important thing too. And to this day,
this is the only thing I'm good at. I'd love to be good at something else. I'm just not.
I think I'm going to start asking this question of everybody. The ongoing education of Pearson Mave,
amazing place to close. I love talking about everything. Thank you so much. This is great, man. Thank you, thank you, thank you.
If you enjoyed this episode, visit JoinColossus.com where you'll find every episode of this podcast
complete with hand out of the transcripts. You can also subscribe to Colossus Review,
our quarterly print digital and private audio publication featuring in-depth profiles
of the founders, investors, and companies that we admire most. Learn more at JoinColossus.com/Subscribe.
[Music]
Key Points:
Bramps AI automates 85% of expense reviews, allowing finance teams to focus on strategic thinking.
Ridgeline is praised for redefining asset management technology and helping firms scale faster.
Alpha Sense offers AI-powered channel checks for real-time expert-driven perspectives on public companies.
Invest Like The Best podcast explores markets, ideas, and investing strategies.
Gavin Baker discusses topics like technology, AI, and investing on podcasts.
Gemini 3, Blackwell chip, and scaling laws in AI are discussed in detail.
Google's TPU and Nvidia's GPU are compared in the context of AI advancements.
Summary:
The transcription covers various topics related to technology, finance, and AI. Bramps AI streamlines expense reviews, allowing finance teams to focus on strategic tasks. Ridgeline is commended for its innovative asset management technology, aiding firms in scaling efficiently. Alpha Sense provides AI-powered channel checks for early insights into public companies. The "Invest Like The Best" podcast delves into market strategies and ideas. Gavin Baker discusses technology and investing on podcasts. Detailed discussions on Gemini 3, Blackwell chip, and scaling laws in AI are presented. A comparison between Google's TPU and Nvidia's GPU in the context of AI advancements is also highlighted.
FAQs
Bramps AI aims to automate 85% of expense reviews with 99% accuracy, allowing finance teams to shift from processing tasks to strategic thinking.
Ridgeline redefines asset management by helping firms scale faster, operate smarter, and stay ahead of the curve, acting as a technology partner rather than just a software vendor.
Alpha Sense offers AI-powered channel checks that provide real-time expert-driven perspectives on public companies, giving investors an edge weeks before traditional sources.
The podcast explores markets, ideas, stories, and strategies to help listeners invest their time and money better, with an invitation to delve deeper through the Colossus Review publication.
Gavin Baker discusses the progress in AI models, emphasizing the importance of scaling laws for pre-training and post-training, showcasing the ongoing advancements and challenges in the industry.
Chat with AI
Ask up to 3 questions based on this transcript.
No messages yet. Ask your first question about the episode.