Episode Transcript
Available transcripts are automatically generated. Complete accuracy is not guaranteed.
Josh:
If I had a crown in my hands, I would place it on the heads of Google because they have done it again. (00:00):
undefined
Josh:
They have the world's best AI model ever in history by a shockingly large margin. (00:04):
undefined
Josh:
Gemini 3.0 just got released. (00:09):
undefined
Josh:
It's available now to anybody in the world to go use it. And the benchmarks (00:11):
undefined
Josh:
are kind of blowing everyone's expectations out of the water, myself included. (00:15):
undefined
Josh:
And most importantly, it places another data point on the chart that shows we (00:18):
undefined
Josh:
are continuing to ascend up this exponential curve towards AGI. (00:22):
undefined
Josh:
And the roadmap is still intact and we are very quickly moving through it. (00:26):
undefined
Josh:
This, EJs, I was just, I was going through the benchmarks before recording this (00:30):
undefined
Josh:
and I, it's, it's shocking because we live in this world and yet somehow I'm (00:32):
undefined
Josh:
still continually blown away by the progress that's made by these models. So let's get into it. (00:37):
undefined
Josh:
Please walk everyone through, tell me, what did Gemini and the Google team just (00:41):
undefined
Josh:
release with this 3.0 update? (00:45):
undefined
Ejaaz:
People probably think we say the world's best model every week, (00:47):
undefined
Ejaaz:
but this time we really, really mean it. (00:51):
undefined
Ejaaz:
Like they have blown every single other model provider out the water. (00:53):
undefined
Ejaaz:
The things that this thing can do. Well, how about how about I just show you? (00:58):
undefined
Ejaaz:
How about I show you, Josh? (01:01):
undefined
Josh:
Please, let's see some examples. (01:03):
undefined
Ejaaz:
We have a thread here. And Sundar basically says, you can give Gemini 3 anything, (01:04):
undefined
Ejaaz:
images, PDFs, scribbles on a napkin, and it'll create whatever you like. (01:09):
undefined
Ejaaz:
For example, an image becomes a board game. (01:13):
undefined
Ejaaz:
A napkin sketch transforms into a full website. (01:15):
undefined
Ejaaz:
And a diagram could turn into an interactive lesson, right? So there's two examples (01:18):
undefined
Ejaaz:
I want to show you, Josh. I want to get your opinion on this. (01:22):
undefined
Ejaaz:
So number one, there's a short video of someone playing pickleball here and (01:25):
undefined
Ejaaz:
he or she rather uploads it into Gemini and says, hey, can you tell me how well (01:30):
undefined
Ejaaz:
I've done here and how I can improve my game? (01:36):
undefined
Ejaaz:
And it analyzes the entire video. (01:38):
undefined
Ejaaz:
It knows that she's wearing a knee brace. It analyzes her positions, (01:40):
undefined
Ejaaz:
telling her where she can move to better position herself to score the point. (01:44):
undefined
Ejaaz:
That's pretty nuts. But before I get your reaction to that, because Josh, (01:48):
undefined
Ejaaz:
I know you're an athlete. I know you're very competitive when it comes to these (01:50):
undefined
Ejaaz:
things. So this is a tool you could definitely use. (01:53):
undefined
Ejaaz:
The second thing is probably applicable to a lot of listeners on this show. (01:54):
undefined
Ejaaz:
They've embedded Gemini 3 into Google search and into new generative UI experiences. (01:59):
undefined
Ejaaz:
The way I would summarize this is it basically is very intuitive, Josh. (02:05):
undefined
Ejaaz:
It understands what you're asking for without you needing to really kind of explain yourself. (02:10):
undefined
Ejaaz:
The example they're showing on the video here is, can you explain the three body problem to me? (02:14):
undefined
Ejaaz:
And rather than just give you kind of like this simplistic text, (02:20):
undefined
Ejaaz:
which explains the concept, it decides to create a video diagram from scratch (02:22):
undefined
Ejaaz:
to show you a visual depiction of how this works. (02:27):
undefined
Ejaaz:
Right. Give me your reaction in order from one to two. So starting with the sports. (02:30):
undefined
Josh:
Okay, let's go. The first example. (02:35):
undefined
Ejaaz:
Yes, sir. (02:36):
undefined
Josh:
So this is really cool, the napkin example, where you can scribble something (02:37):
undefined
Josh:
down on a piece of paper, it'll generate it in the real world. (02:40):
undefined
Josh:
What all these examples are kind of showing me is what we always talk about (02:43):
undefined
Josh:
with Google, where it has this awareness of physics, reality, (02:46):
undefined
Josh:
and visuals and understanding what it's seeing. (02:51):
undefined
Josh:
And all three of these examples are leaning into that. So it leads me to believe (02:53):
undefined
Josh:
Gemini really is a multimodal first model, where it's meant to ingest, (02:56):
undefined
Josh:
meant to understand the world around us. This example of the chessboard and (03:00):
undefined
Josh:
the napkin is amazing because a lot of people oftentimes have sketches. (03:04):
undefined
Josh:
You just draw it down on paper and it intuitively understands it. (03:08):
undefined
Josh:
But the one that was most surprising to me is the video example. (03:10):
undefined
Josh:
Because as far as I'm concerned. (03:13):
undefined
Josh:
As far as I'm aware, there has never been a model that can ingest video and (03:16):
undefined
Josh:
understand the video that it sees. And if it does exist, I've never tried it before. (03:20):
undefined
Josh:
So the idea that you can, I mean, I play baseball growing up. (03:23):
undefined
Josh:
I, if I could take a video of myself swinging and get a corrective coach to (03:26):
undefined
Josh:
walk me through exactly what was wrong. (03:30):
undefined
Josh:
A lot of people play golf. I'm sure who are watching this. (03:32):
undefined
Josh:
If you could have a phone recording of you playing golf and it can actually (03:34):
undefined
Josh:
critique it and then critique me as if you are a tiger woods, (03:38):
undefined
Josh:
critique me as if you are whoever else is good that plays golf. (03:40):
undefined
Josh:
I don't know, Rory McIlroy, whatever they are, but like critique Critique me (03:43):
undefined
Josh:
as if you are an expert who is really good at golf and can give me some feedback (03:45):
undefined
Josh:
on how I can better my swing. (03:49):
undefined
Josh:
And what this offers in this one just narrow case example is now you have this (03:51):
undefined
Josh:
personal tutor that can do anything. If you're dancing, if you're doing anything (03:54):
undefined
Josh:
physical, if you're whatever it is, it can evaluate things for you. (03:58):
undefined
Josh:
Even if you have a video, a podcast, EJS, we uploaded to Gemini 3.0, (04:01):
undefined
Josh:
it could critique us. What did we do well? What did we not? (04:05):
undefined
Josh:
What did the visuals look like? How can we improve them? And that awareness (04:07):
undefined
Josh:
of video is like really cool. (04:10):
undefined
Ejaaz:
Yeah, I just want to say, I think the closest we got to this was with GPT, (04:13):
undefined
Ejaaz:
where you can upload an image, like what's under my car bonnet and say, (04:18):
undefined
Ejaaz:
hey, what's wrong? My car stopped working. (04:22):
undefined
Ejaaz:
And it can kind of like identify the point that you need to kind of like change, (04:24):
undefined
Ejaaz:
change the oil, blah, blah, blah. (04:28):
undefined
Ejaaz:
But that's just a static image. To go from that to live video and for it to (04:29):
undefined
Ejaaz:
analyze all the frames in that video and then give you a response on that is (04:35):
undefined
Ejaaz:
a massive leap upwards. We just haven't seen that anyway. So yeah, you're right. (04:38):
undefined
Josh:
It's amazing. And every example we go through, it kind of breaks the mold of (04:42):
undefined
Josh:
what I believe should be possible. (04:45):
undefined
Josh:
And I find that it's going to be difficult to use Gemini 3.0 because there are (04:47):
undefined
Josh:
so many possibilities now that have not existed previously. You kind of need (04:50):
undefined
Josh:
to relearn how to engage with AI because it's so capable. (04:54):
undefined
Josh:
And there's a fourth example here that I just want to touch on briefly, (04:57):
undefined
Josh:
which was also cool, is that it works just as well for the other things. (05:00):
undefined
Josh:
The example is a trip planning one where it starts to plan a trip and a vacation. (05:03):
undefined
Josh:
And it shows you a full list that is fully interactive of all the places broken up day by day. (05:07):
undefined
Josh:
And there's an option that you could just choose visual layout. (05:12):
undefined
Josh:
And you see on the screen here, it'll take every single day of your trip, (05:15):
undefined
Josh:
break it into images, section it out into this really nice visual grid. (05:18):
undefined
Josh:
So what I'm seeing the themes here are, okay, real world understanding, (05:22):
undefined
Josh:
video first, and really nice presentation, which I think a lot of models sometimes struggle on. (05:25):
undefined
Josh:
Demo's out of control. I'm excited to use it. Everyone else can use it now. (05:31):
undefined
Josh:
It's live. Now I want to get to benchmarks, EJS, because this is where things (05:35):
undefined
Josh:
get kind of crazy, where we could actually compare one to another and see exactly (05:39):
undefined
Josh:
how impressive this is relative to everybody else. So please, we have the card here. (05:43):
undefined
Josh:
Walk us through what we're seeing in this model card and all the specs that we need to know. (05:47):
undefined
Ejaaz:
As you guys probably know by now, benchmarks is typically how we evaluate any (05:50):
undefined
Ejaaz:
typical AI model against each other. (05:54):
undefined
Ejaaz:
And they're measured against a range of different benchmarks. (05:57):
undefined
Ejaaz:
A benchmark can be considered as sort of like a test. (05:59):
undefined
Ejaaz:
Now, right at the top, you've got humanity's last exam. This is by default, (06:01):
undefined
Ejaaz:
the hardest exam that an AI model is tested against. (06:06):
undefined
Ejaaz:
And it's kind of like an academic reasoning test with no tools accessible to it. (06:10):
undefined
Ejaaz:
It scored a very impressive 37.5%, which is more than I think is about a 15% (06:14):
undefined
Ejaaz:
increase from its previous model. (06:21):
undefined
Ejaaz:
Very, very impressive. But what really blew my mind was the second stat listed (06:23):
undefined
Ejaaz:
here, which is the ARK AGI 2 benchmark. (06:28):
undefined
Ejaaz:
Josh, when I say this 2x'd the previous state-of-the-art model, I absolutely mean it. (06:31):
undefined
Ejaaz:
In fact, let me just show you this chart here. (06:38):
undefined
Ejaaz:
Now, you may notice a couple of bustly specs here. (06:41):
undefined
Ejaaz:
GPT-5 Pro, Grok 4 Thinking. And then can you see that outlier right at the top (06:45):
undefined
Ejaaz:
right. Do you see that, Josh? (06:51):
undefined
Josh:
That's insane. The two outliers. (06:53):
undefined
Ejaaz:
The two outliers. So these are the Gemini 3 Pro and the Gemini 3 deep thinking model. (06:55):
undefined
Ejaaz:
Deep thinking being like, you know, a large number that can like kind of give (07:02):
undefined
Ejaaz:
you a more research response. They are a... (07:04):
undefined
Ejaaz:
Stand out from every single other model. And the reason why this is so crazy, (07:07):
undefined
Ejaaz:
well, there's a few reasons. (07:12):
undefined
Ejaaz:
Number one, all the other model progressions, as you can see over time, (07:13):
undefined
Ejaaz:
has been kind of impressive, but kind of small. (07:18):
undefined
Ejaaz:
Like they've been a good jump. It's been impressive, but it hasn't been as impressive (07:23):
undefined
Ejaaz:
to be like, oh, you know, another model provider couldn't catch up. (07:28):
undefined
Ejaaz:
These results from Google literally put it miles ahead of every other model. (07:31):
undefined
Ejaaz:
So when I sat at this chart, I think, wow, Google probably has the lead for (07:36):
undefined
Ejaaz:
another six months and in six months time, they're going to have a more impressive model by then. (07:40):
undefined
Ejaaz:
So at this point, I'm kind of thinking, can anyone catch up to Google? (07:46):
undefined
Ejaaz:
Josh, do you have any reactions to this benchmark? (07:49):
undefined
Josh:
This is the chart that like the first thing I said to myself when I saw this (07:51):
undefined
Josh:
is like, oh, my God, there is no wall there. (07:55):
undefined
Josh:
We are not going to stop scaling. The scaling will still apply because these (07:57):
undefined
Josh:
two new data points that we have blow everything else out of the water. (08:01):
undefined
Josh:
And this is how exponential growth happens. (08:03):
undefined
Josh:
It seems like a really small cluster down there in the bottom, (08:05):
undefined
Josh:
but the reality is that was the top just a couple hours ago. (08:08):
undefined
Josh:
And Gemini kind of refactored this entire chart to make it seem like it's so (08:12):
undefined
Josh:
small because the progress is so high. (08:17):
undefined
Josh:
And although Gemini 3.0 thinking is seemingly the most impressive, (08:18):
undefined
Josh:
the really anomaly chart is the Gemini 3.0 Pro, which is basically a vertical (08:23):
undefined
Josh:
line up from these other models, where the score is higher, but the cost is (08:27):
undefined
Josh:
actually slightly lower. (08:32):
undefined
Josh:
And if you connect a dot between these averages, you start to see literal vertical (08:33):
undefined
Josh:
line in terms of improvement and acceleration in these models. (08:37):
undefined
Josh:
And that to me shows that there is no scaling wall that we're hitting. (08:40):
undefined
Josh:
Like we can continue to scale resources, energy, compute, and we could continue (08:45):
undefined
Josh:
along this path towards AGI in a world where some people were saying, (08:48):
undefined
Josh:
we don't know if it continues. The answer to me is very clearly, it continues. (08:51):
undefined
Josh:
This is a step much closer to AGI. (08:55):
undefined
Josh:
And again, that real world understanding makes it feel much closer to AGI than it ever has before. (08:58):
undefined
Josh:
Because now it really like intuitively understands the world through video, (09:04):
undefined
Josh:
through photo, through audio, through basically every sensory input we have (09:08):
undefined
Josh:
outside of what taste and feel. (09:11):
undefined
Josh:
So this to me, I saw this chart. I was like, oh, my God, Gemini, (09:13):
undefined
Josh:
you really outdid yourselves. (09:17):
undefined
Ejaaz:
I'm just going to be honest. I think over the last couple of months, (09:18):
undefined
Ejaaz:
I've been getting a little bored with the models that have been released by other model providers. (09:22):
undefined
Ejaaz:
And it led me to think that we're not going to make many breakthroughs until (09:27):
undefined
Ejaaz:
some model provider figures out a new, unique way to train their model. (09:32):
undefined
Ejaaz:
Gemini or Google has convinced me otherwise with this release. (09:37):
undefined
Ejaaz:
But I know you guys are probably like fed up with listening to us hop on about benchmarks. (09:41):
undefined
Ejaaz:
So how about I materialize that for you in a much more easy to understand way, right? (09:45):
undefined
Ejaaz:
So here are the four big takeaways that you need to learn about Gemini 3. (09:50):
undefined
Ejaaz:
Number one, for its intelligence, for the intelligence that you're getting, (09:55):
undefined
Ejaaz:
it is not that super expensive. (10:00):
undefined
Ejaaz:
Google trained this from scratch, as this tweet says, using their own GPU infrastructure. (10:03):
undefined
Ejaaz:
And it used this kind of like layout called a mixture of experts, (10:07):
undefined
Ejaaz:
which basically means that whenever you prompt the model, it's not going to use the entire model. (10:11):
undefined
Ejaaz:
So it actually ends up being cheaper than what it could eventually become. (10:16):
undefined
Ejaaz:
One million token context input, 64K token output. We'll get to the costs in (10:19):
undefined
Ejaaz:
a bit of a second, But the point that I'm making here is that it's not as expensive (10:24):
undefined
Ejaaz:
as you would expect for the intelligence that you're getting. (10:27):
undefined
Ejaaz:
Now, if you compared Gemini 3 to GPT 5.1 from OpenAI, on a relative basis, (10:29):
undefined
Ejaaz:
it is more expensive. But for the jump in intelligence that you're getting, (10:36):
undefined
Ejaaz:
it's way better. So it's, in my opinion, worth it. (10:39):
undefined
Ejaaz:
Number two, when it comes to computer use, so that means letting the AI model (10:42):
undefined
Ejaaz:
control your computer and do tasks for you whilst you go do something else. (10:47):
undefined
Ejaaz:
It is state of the art. It is the best here. They measured it against a benchmark (10:52):
undefined
Ejaaz:
called ScreenSpot Pro, which is a benchmark which kind of like analyzes its (10:56):
undefined
Ejaaz:
ability to understand images and visuals on a desktop. It just absolutely crushes it. (11:00):
undefined
Ejaaz:
Number three, it is the best AI for math by far. (11:06):
undefined
Ejaaz:
So again, the point I'm making here or the theme that we're seeing here is it's (11:09):
undefined
Ejaaz:
not just good at one thing, it's good at many things, which makes it the best (11:13):
undefined
Ejaaz:
generalist AI model in the world right now, by far. (11:17):
undefined
Ejaaz:
And the final thing, Josh, and this is where it might slip up. (11:21):
undefined
Ejaaz:
I'm curious to get your take on this. (11:24):
undefined
Ejaaz:
It is insanely good at coding, but we don't quite know if it is the best at coding yet. (11:26):
undefined
Ejaaz:
What I mean by that is it completely crushed everyone else on one coding benchmark, (11:32):
undefined
Ejaaz:
but the coding benchmark that matters, which is the software engineering, (11:37):
undefined
Ejaaz:
SWE, it didn't do as well as its competitor, Claude 4.5 from Anthropic. (11:40):
undefined
Ejaaz:
So those are the four main takeaways. (11:45):
undefined
Josh:
I would much prefer a model that understands the world than understands how to code. (11:47):
undefined
Josh:
And I think we're starting to see these subset niches where if Anthropic has (11:51):
undefined
Josh:
the best coding model, that's great. (11:55):
undefined
Josh:
Let them focus on code. Let them narrowly make that the best model. (11:56):
undefined
Josh:
Let Google handle everything else. And I think that's what Gemini is focusing (12:00):
undefined
Josh:
on. So the code thing doesn't really bother me because I don't care to use Gemini for code. (12:03):
undefined
Josh:
I'm happy to be in Claude Camp for code and then use Gemini for everything else. (12:07):
undefined
Josh:
And then one of the points earlier that you mentioned on the pricing is. (12:11):
undefined
Josh:
I find it a little interesting because it's a little bit more than just a little bit more expensive. (12:14):
undefined
Josh:
The pricing, I was looking through it and it's for over 200,000 tokens. (12:18):
undefined
Josh:
They're charging $4 for inputs and $18 for outputs. (12:22):
undefined
Josh:
Now, relative to GPT 5.1, which just got released, they're charging for a million (12:27):
undefined
Josh:
tokens, $1.25 in, $10 out. (12:32):
undefined
Josh:
So you're talking about, what is that? that's about $20 versus $1.25 on inputs. (12:35):
undefined
Josh:
And that is a fairly significant margin that you're paying for this quality. (12:42):
undefined
Josh:
So we're starting to see the trade-offs happening on that Pareto curve that (12:45):
undefined
Josh:
we talked about in a few episodes earlier, where there are trade-offs coming (12:48):
undefined
Josh:
in terms of cost and quality. (12:52):
undefined
Josh:
And it's clear that while OpenAI may have optimized for cost, (12:53):
undefined
Josh:
Google is kind of optimizing for a little further up the cost curve in exchange for super high quality. (12:57):
undefined
Josh:
And it seems like this is kind of a balanced data point for now because unless (13:02):
undefined
Josh:
you are using this via API and you're requiring a ton of tokens, (13:06):
undefined
Josh:
a $20 a month Google membership will get you all of the use that you need. And that is just fine. (13:09):
undefined
Josh:
So in terms of a usability perspective, I think that's okay. (13:14):
undefined
Josh:
But it's just an interesting thing to know is that this is a better model. (13:18):
undefined
Josh:
It is also more expensive. (13:21):
undefined
Josh:
And that is a trade-off that was made. And in the case that OpenAI decides to (13:22):
undefined
Josh:
make this trade-off with GPT-6 or Grok decides to make this with Grok-5, (13:25):
undefined
Josh:
or Grok 6. I'm losing track of all these models now. I think we're going to (13:30):
undefined
Josh:
start to see the dynamic shift in terms of that Pareto curve and what model (13:34):
undefined
Josh:
architects decide to remove and add. (13:37):
undefined
Josh:
And in this case, it looks like Google added quality, but they also did add (13:39):
undefined
Josh:
quite a significant cost increase. (13:42):
undefined
Ejaaz:
I personally don't think it matters. I think it's a nothing burger. (13:44):
undefined
Ejaaz:
I think that if Google wanted to make it affordable for everyone, (13:46):
undefined
Ejaaz:
including the developers that want to get API access, that think it might be (13:50):
undefined
Ejaaz:
too expensive, they could subsidize it. (13:52):
undefined
Ejaaz:
They are a cash flow giant. They have enough money to do that. (13:54):
undefined
Ejaaz:
OpenAI has been doing that for so long now that it doesn't even matter. (13:57):
undefined
Ejaaz:
I don't see any reason why Google couldn't do that. (14:00):
undefined
Ejaaz:
The other reason is Google just released their latest TPU, which is their GPU (14:02):
undefined
Ejaaz:
that they use to train their models and inference their models. (14:08):
undefined
Ejaaz:
And typically with every generation, we get a much cheaper cost of inference. (14:11):
undefined
Ejaaz:
I think by the time that they release their next generation model, (14:15):
undefined
Ejaaz:
which might be, you know, Gemini 3.1, we're going to see a considerable reduction (14:18):
undefined
Ejaaz:
in the cost for using Gemini 3 Pro and Gemini Pro deep research. (14:24):
undefined
Ejaaz:
So I'm not too worried about that. I think it's kind of like a short-term problem (14:29):
undefined
Ejaaz:
and not a long-term problem. (14:32):
undefined
Ejaaz:
But kind of speaking of TPUs, I just want to take a moment to really kind of (14:34):
undefined
Ejaaz:
belay the point that using their own TPUs to train a state-of-the-art model (14:39):
undefined
Ejaaz:
that is 2x better than the previous state-of-the-art model and probably puts (14:44):
undefined
Ejaaz:
them in a six-month lead after Google started off on the back foot, (14:48):
undefined
Ejaaz:
creating probably the worst model I've ever seen and (14:53):
undefined
Ejaaz:
changing that all around in what's it under two years is (14:56):
undefined
Ejaaz:
nothing short of insanity tpus is (14:59):
undefined
Ejaaz:
uh google's kind of version of the gpu gpu is (15:02):
undefined
Ejaaz:
kind of like what nvidia controls the monopoly over this is the hardware that (15:05):
undefined
Ejaaz:
you use to train your ai and inference your ai the uh unique part here is that (15:09):
undefined
Ejaaz:
google's never used an nvidia gpu uh or in any considerable way to train their (15:13):
undefined
Ejaaz:
models they've always trained it in-house and that's such a difficult and tricky (15:19):
undefined
Ejaaz:
thing to do because designing and building these (15:23):
undefined
Ejaaz:
TPUs at scale, these GPUs at scale is a super hard and complex thing. (15:25):
undefined
Ejaaz:
You need so much talent, you need so much expertise and insight to be able to do that. (15:29):
undefined
Ejaaz:
The unique thing about Google's TPUs, well, there's two main takeaways. (15:33):
undefined
Ejaaz:
Number one, it's cheaper to train the same amount of intelligence that an NVIDIA (15:36):
undefined
Ejaaz:
GPU is. So it's more cost efficient. (15:41):
undefined
Ejaaz:
And the second way is, and this is their secret sauce, you can stack those TPUs (15:43):
undefined
Ejaaz:
on top of each other in a really scalable way that you can start training really, really large models. (15:47):
undefined
Ejaaz:
If you wanted to train the same size model with nvidia (15:53):
undefined
Ejaaz:
gpus it would cost way more and it would take way (15:57):
undefined
Ejaaz:
longer so google made a really risky and big bet about a decade ago saying we're (16:00):
undefined
Ejaaz:
going to build our infrastructure in-house and we're not going to rely on nvidia (16:05):
undefined
Ejaaz:
and we're going to benefit from the full stack experience and this model is (16:07):
undefined
Ejaaz:
a prime example of that bet paying off so i just want to call them out like (16:11):
undefined
Ejaaz:
it's not like google has gotten lucky here they've been planning it for a while now The (16:15):
undefined
Josh:
Interesting thing to me is that this is the first number one model in the world (16:19):
undefined
Josh:
built on something other than an NVIDIA GPU. (16:24):
undefined
Josh:
And that's fairly significant because every company in the world is trying, (16:27):
undefined
Josh:
but this is proof that it's actually possible. (16:31):
undefined
Josh:
And I think when we talk about Tesla and AI5 and the XAI team, (16:32):
undefined
Josh:
when we talk about OpenAI working with whoever they're working with to build (16:37):
undefined
Josh:
their own in-house GPUs, I think this sets a precedent that it is possible. (16:41):
undefined
Josh:
And I suspect that will result in more companies putting their foot on the gas. (16:45):
undefined
Josh:
When it comes to kind of destructing part of NVIDIA's monopoly that it holds (16:49):
undefined
Josh:
over GPUs. So that to me is the interesting takeaway of this. (16:53):
undefined
Josh:
And hearing that it was fully done, trained on these TPUs, that's very high (16:56):
undefined
Josh:
signal to me saying, okay, there is an architecture chips happening. (17:00):
undefined
Josh:
There is a real benefit to vertical integration if you could figure out manufacturing (17:03):
undefined
Josh:
these compute units at scale. (17:06):
undefined
Josh:
And now the race is on for everyone to do this. (17:08):
undefined
Josh:
Because again, using the Apple example, the M series chips, unbelievable, (17:12):
undefined
Josh:
and they unlocked the best computers in the world. (17:15):
undefined
Josh:
And if companies can really start to refine this vertical integration of their (17:16):
undefined
Josh:
own chips, you're going to see that exponential curve go vertical times 10. (17:20):
undefined
Josh:
Like it is going to, I suspect that is very... (17:24):
undefined
Josh:
Obviously now how we reach AGI faster than people previously thought, (17:28):
undefined
Josh:
because the efficiency improvements from those vertical integrations, (17:32):
undefined
Josh:
once they're able to manufacture these at scale, are going to be unbelievable. (17:35):
undefined
Josh:
And I'm so excited for that to happen in the near future. Google has a big head (17:39):
undefined
Josh:
start, but let me tell you, the other companies are not far behind. (17:43):
undefined
Ejaaz:
Well, let me introduce you to another big advantage of being the big dog, Google. (17:46):
undefined
Ejaaz:
You thought you were going to come on to this episode and listen to us hopping (17:52):
undefined
Ejaaz:
on about a generalized model? No. (17:55):
undefined
Ejaaz:
You're forgetting Google has many other products in their arsenal, (17:57):
undefined
Ejaaz:
and you're forgetting that they can plug in their new state-of-the-art model into all of them. (18:00):
undefined
Ejaaz:
So Google not only today announced Gemini 3, but they also announced a different product. (18:05):
undefined
Ejaaz:
It's called Google Anti-Gravity, which is basically a new software environment (18:10):
undefined
Ejaaz:
for you to code up AI agents, except this time these AI agents are going to (18:15):
undefined
Ejaaz:
be super, super smart because they get plugged in with Gemini 3. (18:20):
undefined
Ejaaz:
Now, if you remember earlier, I mentioned that one of the cool benchmarks that (18:24):
undefined
Ejaaz:
this new model sets is in computer use, which means that it can control your (18:27):
undefined
Ejaaz:
computer, it can do things autonomously for you. (18:31):
undefined
Ejaaz:
Now, typically, the reason why we haven't really spoken about that on the show (18:34):
undefined
Ejaaz:
is that they've been kind of lame. (18:36):
undefined
Ejaaz:
Like they can book you a dinner reservation and do different kinds of stuff. (18:38):
undefined
Ejaaz:
With this model, it's way more intuitive. It can do way more intelligent tasks (18:42):
undefined
Ejaaz:
and it can take a lot more complex work off of your hands such that the value (18:46):
undefined
Ejaaz:
that it produces to you over like the eight hours that you take to sleep overnight (18:50):
undefined
Ejaaz:
would be considerable for you to actually be serious to use in your enterprise, (18:54):
undefined
Ejaaz:
in your business, or just at your at-home lifestyle, right? (18:58):
undefined
Ejaaz:
So the point I want to make around here is Google's moat is not just its intelligence (19:01):
undefined
Ejaaz:
or ability to create new models. It's not its TPUs. (19:07):
undefined
Ejaaz:
It's its distribution. It's the entire product suite that it has that regular (19:10):
undefined
Ejaaz:
users like you and I that use Gmail, that use Google Suite can now kind of benefit (19:15):
undefined
Ejaaz:
from simply by plugging in that model. (19:19):
undefined
Ejaaz:
And I think like products like this, anti-gravity, I bet you, (19:22):
undefined
Ejaaz:
Josh, we're going to see a slew of new Google product releases over the next (19:25):
undefined
Ejaaz:
couple of weeks simply because they created this model. (19:28):
undefined
Josh:
I hope so. I guess the contrarian take is like, okay, how many people are actually (19:32):
undefined
Josh:
going to want to use them? (19:36):
undefined
Josh:
We just spoke about how Claude is the superior code model. Everyone loves Cursor. (19:37):
undefined
Josh:
No one really uses the mobile applications of these. A lot of people are engaging (19:42):
undefined
Josh:
with AI on their phone. So maybe it works for the right type of person. (19:46):
undefined
Josh:
But Google still does have that product problem where they kind of have a tough time. (19:49):
undefined
Josh:
They have the amazing intelligence. They just have a tough time using it. (19:54):
undefined
Josh:
I mean, I don't have the Gemini app on my phone. I mostly use Grok and ChatGPT. (19:56):
undefined
Josh:
And there is this bar that they (20:00):
undefined
Josh:
still need to cross that I think they're trying with Google AI Studio. (20:01):
undefined
Josh:
And we had Logan Kilpatrick on, who was the head of that, to talk about it when Nano Banana came out. (20:05):
undefined
Josh:
But there is still a bit of a long shot for them to get good at products to actually develop this. (20:08):
undefined
Josh:
But what we saw this week is that there was a resounding, overwhelming amount (20:14):
undefined
Josh:
of support, to your point, you guys, where the market just believes in Google. (20:17):
undefined
Josh:
And in a week where all of the stocks, all of the Mag7 was down, (20:20):
undefined
Josh:
Google was the one anomaly. Google was up this week. (20:24):
undefined
Josh:
And I think it's because the market is starting to realize, one, (20:26):
undefined
Josh:
vertical integration through these TPUs is a huge deal. (20:29):
undefined
Josh:
Two, Google has an existing business that is not reliant on AI. (20:33):
undefined
Josh:
And sure, AI places a huge hand on that scale, but it is not everything. (20:36):
undefined
Josh:
And they are cashflow positive in the absence of AI. So all of this innovation (20:41):
undefined
Josh:
that they're doing is really just applying later fluid on top of an already (20:45):
undefined
Josh:
great business. And the market's starting to evaluate that properly. (20:48):
undefined
Josh:
So Google is positioned very strongly. (20:51):
undefined
Josh:
They have very high intelligence. Gemini 3 rocks. (20:53):
undefined
Josh:
And I mean, again, we continue on the bull train for Google. (20:56):
undefined
Josh:
I am a believer. I am a supporter. (20:59):
undefined
Josh:
I am stoked that they have the crown. I assumed it was only a matter of time. (21:01):
undefined
Josh:
And now the question is, who's next? (21:04):
undefined
Josh:
Who is the next competitor? Who's going to set the next plot on that chart and (21:06):
undefined
Josh:
set the vertical trajectory on the exponential curve we're on? (21:10):
undefined
Josh:
Do you have any guesses who you think it's going to be? (21:14):
undefined
Ejaaz:
Yeah, well, I don't because I don't think it's going to be anyone for a while. (21:16):
undefined
Ejaaz:
I said this earlier in the show and I'm going to say it again. I think there's going (21:20):
undefined
Ejaaz:
to be a six month period now where either the other model providers don't release (21:23):
undefined
Ejaaz:
the model because it's not as good as Google's or they just kind of release (21:29):
undefined
Ejaaz:
these kind of mediocre kind of like consumer products that kind of maybe benefit (21:33):
undefined
Ejaaz:
certain consumers in one way or another, (21:38):
undefined
Ejaaz:
but doesn't really kind of break the generalized model (21:41):
undefined
Ejaaz:
standard that Google has just set. (21:44):
undefined
Ejaaz:
Just a last point on the kind of Google bull case thesis, they may not just (21:47):
undefined
Ejaaz:
play in the same ring as cursor does. (21:51):
undefined
Ejaaz:
Like I was critiquing Microsoft on another episode, Josh, do you remember? (21:53):
undefined
Ejaaz:
And then I got off that episode and I was just like, Microsoft like dominates (21:57):
undefined
Ejaaz:
the enterprise environment. (22:02):
undefined
Ejaaz:
All the boomer companies and institutions love Microsoft and they have all their data and memory. (22:04):
undefined
Ejaaz:
And just because you and I don't use it or just I'll speak for myself, (22:09):
undefined
Ejaaz:
just because I don't use it and I think it's doesn't mean that they're not absolutely crushing. (22:12):
undefined
Ejaaz:
Google just came off a hundred billion dollar quarter of revenue. (22:15):
undefined
Ejaaz:
That's like the highest they've ever had. So I don't want to be too hasty to (22:19):
undefined
Ejaaz:
say that like Google's not going to make it because they can't make a sick consumer (22:23):
undefined
Ejaaz:
product like OpenAI can maybe. (22:26):
undefined
Ejaaz:
I just think they're maybe playing in different fields. (22:28):
undefined
Ejaaz:
But to the point around like I don't think anyone else is going to catch up. (22:31):
undefined
Ejaaz:
Look at these comments. I want to show you two comments all right one is from sam altman (22:34):
undefined
Ejaaz:
He goes, congrats to Google and Gemini 3. This looks like a great model. (22:38):
undefined
Ejaaz:
The other is from the almighty being, Elon Musk, saying, I can't wait to try this out. (22:42):
undefined
Ejaaz:
And this is just one of a series of tweets that he's been putting out this week (22:47):
undefined
Ejaaz:
saying, can you guys just drop Gemini 3? Because I need to see how good this thing is. (22:49):
undefined
Ejaaz:
And the reason why I bring up these two people is both Sam Altman and Elon Musk (22:53):
undefined
Ejaaz:
have released new versions of their models, GPT and Grok respectively, (22:57):
undefined
Ejaaz:
but it's been the 0.1 upgrade. (23:02):
undefined
Ejaaz:
It's GPT 5.1. It is Grok 4.1. and they are almost identical updates. (23:04):
undefined
Ejaaz:
You want to know what the biggest and coolest thing about their model updates were? (23:10):
undefined
Ejaaz:
Personality traits, which don't get me wrong, is cool. Like I would like my (23:13):
undefined
Ejaaz:
model to kind of respond in a very intuitive manner and get me, (23:16):
undefined
Ejaaz:
but it's nowhere near the state of the art standard that we've just seen broken by Gemini. (23:19):
undefined
Ejaaz:
So the point I'm making is, I think these two companies might have run out of fuel for the near term. (23:23):
undefined
Josh:
Grok is going to be next. You think? They're the next one. By the end of Q1, (23:28):
undefined
Josh:
Grok will have the crown. (23:32):
undefined
Josh:
Why? And I assume by a fairly large margin. But I assume it will be a different type of crown. (23:33):
undefined
Josh:
And this is where I'm really excited to see how these models progress. (23:39):
undefined
Josh:
We spoke a little bit earlier about how Cursor is kind of the coding model. (23:42):
undefined
Josh:
Google has a very deep understanding of the real world and physics and video (23:46):
undefined
Josh:
and understanding how that works. (23:50):
undefined
Josh:
Grok and the XAI team are very focused on the pursuit of truth and information. (23:51):
undefined
Josh:
And I think that's kind of the alley that we'll see them going down. (23:56):
undefined
Josh:
So they have the real-time data with X. They have the pursuit of truth. (23:59):
undefined
Josh:
And where Google and OpenAI and all these other companies are trained on an (24:03):
undefined
Josh:
existing data set, the XAI team and the Grok team are developing an entirely (24:06):
undefined
Josh:
new synthetic data set that is maximally truth-seeking. (24:10):
undefined
Josh:
And we saw that early version with Grokopedia that should provide the most accurate (24:15):
undefined
Josh:
and I guess thoughtful information. It should be the best of thinking because (24:19):
undefined
Josh:
it's the closest to source truth. So while I think, (24:23):
undefined
Josh:
Gemini will probably be better at physics and video and understanding the real (24:26):
undefined
Josh:
world for quite some time. (24:30):
undefined
Josh:
I suspect Grok will be really good at just communicating via text. (24:31):
undefined
Josh:
If text is a modality in which we interface from, Grok should be really good. (24:36):
undefined
Josh:
And again, the rate of acceleration, Grok has been around for the least amount (24:40):
undefined
Josh:
of time. They're accelerating the fastest. (24:42):
undefined
Josh:
And I'm very, very, very excited for a Grok 6, Grok 5, whatever we're at announcement, (24:44):
undefined
Josh:
hopefully early next year. So that's the predictions. (24:49):
undefined
Josh:
That's the episode. That's Gemini 3.1. It is an unbelievable new model. (24:52):
undefined
Josh:
Everyone could try it out. So here's how you try it out. (24:56):
undefined
Josh:
I believe you need to be a Google premium plus subscriber, whatever it's called. It's $20 a month. (24:58):
undefined
Josh:
You can go on the Gemini website and it's just a text box and you can play around (25:04):
undefined
Josh:
with it. They also have a mobile application. (25:07):
undefined
Josh:
It's very easy to download on your phone, play around with it. (25:09):
undefined
Josh:
I'd love to see examples of cool things because I think one of the problems (25:12):
undefined
Josh:
for me and one of the things I'd love help with from anyone who's listening (25:15):
undefined
Josh:
is how do you use this thing to test it? (25:17):
undefined
Josh:
What do I ask it? And how are you interfacing with it to get the maximum amount of results from it? (25:20):
undefined
Josh:
Because intuitively, I would never think to record myself and ask for feedback, (25:25):
undefined
Josh:
but that's a new possibility. (25:28):
undefined
Josh:
So I guess the challenge to anyone who's listening is figuring out how to get (25:29):
undefined
Josh:
the most out of these new models as these new features get released. (25:33):
undefined
Josh:
And Gemini 3 has just opened up the gates to a gazillion new use cases. (25:36):
undefined
Ejaaz:
Yeah, I mean, this is a super cool release for Google. (25:41):
undefined
Ejaaz:
And weirdly enough, it's not the only release over the last week. (25:46):
undefined
Ejaaz:
I mean, I've got a list pulled up here. They've released new Android iOS updates. (25:49):
undefined
Ejaaz:
They've got a new search AI mode. They've released anti-gravity that we mentioned earlier. (25:53):
undefined
Ejaaz:
We've got CIMA2 research, which we demoed on a previous episode. (25:57):
undefined
Ejaaz:
You should definitely go check that out. (26:00):
undefined
Ejaaz:
I mean, they are just not stopping and they're a force to reckon with. (26:02):
undefined
Ejaaz:
And kind of similar to them, Josh, just to kind of round this episode out and (26:07):
undefined
Ejaaz:
thank you guys for listening. (26:11):
undefined
Ejaaz:
We are here in Argentina, in (26:13):
undefined
Ejaaz:
Buenos Aires. We are kind of meeting some of the fans that are out here. (26:15):
undefined
Ejaaz:
And we spoke to one just this afternoon, Josh. And you know what he said to me? Have a guess. (26:18):
undefined
Josh:
What's that? (26:23):
undefined
Ejaaz:
He said, your podcast Limitless is like the state-of-the-art AI podcast. (26:24):
undefined
Ejaaz:
In fact, it is 2X better than any other AI podcast that I've ever heard. And you know what? (26:30):
undefined
Ejaaz:
That sounds very similar to (26:35):
undefined
Ejaaz:
the Gemini 3. So you could potentially call us the Gemini 3 of AI podcast. (26:36):
undefined
Ejaaz:
And so if you're a listener to this, if you are a non-subscriber on our YouTube, (26:41):
undefined
Ejaaz:
you should probably click that subscriber button. (26:46):
undefined
Ejaaz:
You should probably click that notification button because guess what? (26:48):
undefined
Ejaaz:
We've got more episodes coming this week. And guess what? the five star ratings help us out massively. (26:51):
undefined
Ejaaz:
So if you enjoyed this episode and if you want to hear more episodes of this (26:56):
undefined
Ejaaz:
nature and of cutting edge news in AI, you should give us a follow and we will see (26:59):
undefined