Elon's Recipe for Winning the AI Race: Grok5 and Colossus
Josh:
It's been a big couple of weeks for elon we had a few pretty hit
Josh:
episodes last week talking about starlink talking about the ai5 chip
Josh:
and this week it's just another big breakthrough ejaz this week we're coming
Josh:
out with a lot of new grok and xai news which is pretty exciting i mean one
Josh:
of the leading headlines he said i now think xai has a chance of reaching agi
Josh:
with grok 5 never thought that before and now there's two things that kind of
Josh:
spawn this one which we'll get into a little bit later, which is the Grok Fast model.
Josh:
It is remarkable. It is a full order of magnitude better than anything else for its size.
Josh:
And it is really, really impressive. But the thing we're going to start with,
Josh:
Ejaz, is this chart that we're showing on screen right here,
Josh:
which is the single thing that convinced Elon, wait a second,
Josh:
maybe, maybe just maybe, Grok 5 could actually lead to AGI.
Josh:
And it's because we're seeing this crazy anomaly on the chart where Grok 4 was
Josh:
kind of ahead, but somehow without any new major release, Grok 4 is now way ahead.
Josh:
So Ejaz, can you explain to us like what's going on in this chart?
Josh:
How did they get so good so fast without a major new model release?
Josh:
I mean, this didn't even come from XAI, did it?
Ejaaz:
It's a good question. And no, it didn't come directly from XAI.
Ejaaz:
It actually came from two random AI researchers, one called Jeremy Berman and
Ejaaz:
the other one called Eric Pang, who tweaked Grok 4's model,
Ejaaz:
also known as fine tuning, to basically make it a hell of a lot smarter.
Ejaaz:
And so they put it to the ultimate test, Josh.
Ejaaz:
It's this thing called the Arc AGI benchmark.
Ejaaz:
And for those of you who have not been spending all your time researching benchmarks,
Ejaaz:
the Arc AGI benchmark tests how good your AI model is at being successful.
Ejaaz:
Intelligently human. What I mean by that is it presents the AI model with puzzles
Ejaaz:
that it's never seen before, that it can't possibly have been trained to solve
Ejaaz:
and sees how good it does. Now, Josh, let me ask you this question.
Ejaaz:
Before Grok 4 itself was released, what do you think the highest score was on this benchmark?
Josh:
Lower, but I'm not sure how much lower. I don't know this particular numbers,
Josh:
but maybe I'll guess five to 10% lower than what the best is now,
Josh:
kind of like an incremental improvement.
Ejaaz:
Nope nope nope it was way way
Ejaaz:
lower in fact it only scored between five to eight
Ejaaz:
percent from the top models from open ai google and all those kinds of things
Ejaaz:
and then grok 4 came yeah grok 4 came along and it broke that frontier and scored
Ejaaz:
22 percent guess how much these two models that these two random ai researchers um scored on wait
Josh:
So you're I'm looking at the screen. I'm seeing 79.6%. Is that right?
Josh:
Is this a 4X multiple on base Grokfor?
Ejaaz:
80%. And this had nothing to do with the XAI team at all. I want you to focus
Ejaaz:
on this chart that I'm showing you right now.
Ejaaz:
And look at my cursor circling around these two orange dots that are off into the distance.
Ejaaz:
You see Grokfor thinking over here, which was basically the heaviest,
Ejaaz:
most expensive model that Elon and the XAI team released.
Ejaaz:
Um when they launched croc 4 and they were just completely beaten by these two
Ejaaz:
models but i'm sure you're probably thinking josh how the hell did these two
Ejaaz:
researchers do that and um you know why aren't they being hired by elon immediately they
Josh:
Don't have the resources of a giant lab.
Ejaaz:
Like they're
Josh:
Competing against i mean if you remember these people are getting billion dollar
Josh:
offers to come work for a single employer and there's a collection of these
Josh:
so how is it that one individual it's being a collection of these people.
Ejaaz:
So these two researchers introduced two novel
Ejaaz:
ways of training their models one is called
Ejaaz:
open source program synthesis and the
Ejaaz:
other is called test time adaptations before
Ejaaz:
i get into an explanation as to how these work i
Ejaaz:
want to remind the audience that what really makes a model really intelligent
Ejaaz:
um is largely part in due to the data that it's trained on people spend so much
Ejaaz:
money i'm talking hundreds of millions to billions of to acquire the best data to train their models.
Ejaaz:
And the reason why this is so important is the model, when it's trying to answer
Ejaaz:
a question, draws on the data that it's been trained on, right? So it's hoping that
Ejaaz:
it can look back on the data that it's been trained on and find the right answer
Ejaaz:
somewhere in all of these tokens and characters, right, Josh?
Ejaaz:
These researchers decided to flip that completely on its head.
Ejaaz:
It's this thing called open source program synthesis where the model designs
Ejaaz:
its own bespoke solutions in real time.
Ejaaz:
So it doesn't even look at the data that it was trained on.
Ejaaz:
It just looks at the puzzle that it's presented with and it tries to break it
Ejaaz:
down into smaller components. So let's say the puzzle has 10 different steps
Ejaaz:
to reach to the end goal, the correct answer.
Ejaaz:
It'll break it down into 10 different little steps, whereas normally a model
Ejaaz:
would just look at the complete set of 10 steps and think, hmm,
Ejaaz:
how do I get from step one to step 10?
Ejaaz:
It just solves each step one at a time.
Ejaaz:
And that was like the massive breakthrough that they made.
Ejaaz:
And if this sounds familiar, you're probably thinking of this technique known
Ejaaz:
as reinforcement learning,
Ejaaz:
which basically has like the model like repeatedly go at a problem over and
Ejaaz:
over again this is pretty similar but it's the next step up in that field
Josh:
Okay got it yeah this this news kind of really annoyed me because of how seemingly
Josh:
simple it was i mean jeremy burman in the case of this i got some examples of
Josh:
specifically how he did it and he he was originally writing in python code but
Josh:
then he switched to just writing instructions in plain english and i think this
Josh:
is such an important thing that a lot of people forget.
Josh:
I mean, myself included, I'm speaking for myself here, that a lot of this really
Josh:
challenging, difficult work with engaging with LLMs is really just done in plain English.
Josh:
You're just writing sentences to a model in hopes that it produces a better
Josh:
output for you. It's not this crazy complex code base, although that exists deep down.
Josh:
But the way that they achieve this is actually just by writing plain English.
Josh:
So I did a little bit of digging. I have a few notes on how it works.
Josh:
And his system, it basically starts by having Grok 4. He chose Grok
Josh:
for is this model of choice it produces 30 english descriptions
Josh:
of rules to transform inputs into outputs so it
Josh:
takes that and then it tests these descriptions on training examples by
Josh:
pretending each is a test and scoring how well
Josh:
they match the correct outputs and then the top five descriptions get
Josh:
revised individually with feedback on mistakes like highlighting the
Josh:
wrong cells and stuff like that and then it combines the elements into
Josh:
the top one to create these pool descriptions so it kind of has this iterative loop
Josh:
where it tests itself it creates more examples it
Josh:
gets better data it confirms that it's the right output and that's
Josh:
generally why you see the actual outputs of this model are
Josh:
a little more expensive but the quality of it is amazing because it
Josh:
just continues to do this like self-iterative loop on itself and
Josh:
get better and better and better again all in plain english so if you are listening
Josh:
to this podcast in english you are fully capable of doing this because you speak
Josh:
the language and this isn't anything crazy it's just very refined um prompts
Josh:
that you're feeding to a model that result in these unbelievable outputs that
Josh:
are now best in the world.
Josh:
That's the coolest part to me, EJS. I don't know about EJS.
Ejaaz:
No, no, I agree. And it reminds me of Andrew Carpathy's hit tweet three months
Ejaaz:
ago where he goes, the new number one programming language
Ejaaz:
Turned out to be English. It's English. Right?
Ejaaz:
And kind of like to emphasize, again, how important this is.
Ejaaz:
This isn't just another frontier breakthrough of another benchmark.
Ejaaz:
I'm talking about the hardest benchmark that has just been 3x'd by two random researchers.
Ejaaz:
Right? This is, again, puzzles that are problem sets that an AI model has never
Ejaaz:
seen before. Typically, when you put an AI model up against a benchmark,
Ejaaz:
it has some kind of context.
Ejaaz:
Kind of think of yourself taking an exam at school or at university.
Ejaaz:
You can look at past papers. You can look at books. You kind of know what topics
Ejaaz:
they're going to talk about.
Ejaaz:
This is completely foreign to an AI model. And therefore, it is the hardest test.
Ejaaz:
So to have something achieve this almost feels like, and Josh,
Ejaaz:
I hate to say it, but I have to say it, like AGI.
Ejaaz:
And I think the fact that none other than Elon himself was taken completely aback by this.
Ejaaz:
I mean, again, to reiterate the tweet, I now think XAI has a chance of achieving
Ejaaz:
AGI with Grok5, never thought that before.
Ejaaz:
And the fact that he is now saying, hey, by the way, Grok5 starts to train in
Ejaaz:
a few weeks and you know what?
Ejaaz:
I think it's going to be out by the end of this year.
Ejaaz:
I think just speaks to the importance of this development.
Josh:
Yeah, I think one of the things that was really startling for me was the realization
Josh:
of how little resources it takes to get so good. And then I was wondering, well, why?
Josh:
Clearly, this isn't anything super novel, although they did do some unique training frameworks.
Josh:
And I think the reason that I, the conclusion that I came to was just scale.
Josh:
I mean, the cost per query, the cost per token of these new super high end models
Josh:
that just came out is very high.
Josh:
And you can't really scale that to a lot of people because the companies are
Josh:
just resource constrained.
Josh:
So it leads me to believe and leads me to think, well, what happens when a company
Josh:
with a lot of resources dedicates all of their brainpower to this specific type
Josh:
of reinforcement learning, like we're going to see with Grok5,
Josh:
and they do so in a way that's compressed enough, that's efficient enough to
Josh:
actually run it at scale on the servers without melting everything down without
Josh:
charging $1,000 a month per membership.
Josh:
And I think that's probably what we see with Grok5 is this new juiced up reinforcement
Josh:
learning, but efficient and actually built for scale.
Josh:
I mean, even if it just launches at the specs of these two individual researchers,
Josh:
that's a huge win because that's an incredible model.
Ejaaz:
Yeah. And it's open source. It's open source and available for everyone.
Josh:
It's pretty remarkable. Yeah. So I think very interesting things coming.
Josh:
If I was a betting man, I would be betting big on Grok5. I think they very much
Josh:
see a solution that people really want.
Ejaaz:
I was just thinking about why
Ejaaz:
both of us are finding this development both amazing, but really annoying.
Ejaaz:
And I think it's because to some degree
Ejaaz:
we both believe that in order for AI models today
Ejaaz:
to get to AGI we would need to completely re-architect
Ejaaz:
how they're designed you know transformers was
Ejaaz:
the big breakthrough that's why models that we know and use today
Ejaaz:
are so smart but it's not as
Ejaaz:
smart as we expected it and there was this kind of like lag of improvement and
Ejaaz:
now we suddenly see a 3x improvement where this model is kind of breaking this
Ejaaz:
leading benchmark and so now I I think I'm starting to believe that maybe if
Ejaaz:
we invest hundreds of billions of dollars in the post-training part,
Ejaaz:
where typically we've been investing in the pre-training, in the compute,
Ejaaz:
but if we invest it in the post-training, we may clearly reach AGI before redesigning
Ejaaz:
the entire thing up front.
Ejaaz:
Does that resonate with you, Joshua? Or am I just, do I sound crazy?
Josh:
No, it does. It does. It's funny because we frequently record the show and you
Josh:
expect to be surprised and then something happens.
Josh:
You're like, oh my, I wasn't expected to be surprised in that way. and this
Josh:
is one of those things where i mean i wasn't expecting to see a new leader in
Josh:
between major model releases from an independent researcher so the fact that
Josh:
this is even possible really just blows the doors off of a lot of expectations
Josh:
i had and this isn't even the only interesting news this week from the xai team
Josh:
because they released new model alert grok for,
Josh:
fast let me tell you each as when i saw how this model worked i was like whoa
Josh:
this is um again blown away super.
Ejaaz:
Impressed can you run us through some of the highlights please
Josh:
This spec sheet yeah so first of all the leading headline
Josh:
two million token context window is outrageous i think the current leader is
Josh:
google with the gemini model they have 2.5 pro and flash both of them i believe
Josh:
have a million tokens uh this is two million tokens of context for those that
Josh:
aren't aware context is the basically active memory of a language model it's the more context you can,
Josh:
collect, the more clarity it has into the actual data that it's talking about
Josh:
and conversing with, you want that number to be bigger.
Josh:
This is the biggest by far, by a doubling.
Josh:
So that's a really important headliner. The second one, probably even more outrageous,
Josh:
47 times cheaper than ROC4.
Josh:
Which is crazy because when you look at it on
Josh:
the scale below if you can scroll down just a little
Josh:
bit grok 4 is right in line with every other great
Josh:
model it is a grok 4 fast is just beneath
Josh:
o3 it's above deep seek it's above cloud 4 sonnet it's a above cloud 4 opus
Josh:
it's just like this remarkable model that is better than a lot of the leading
Josh:
models but 47 cheaper than the base model and i think that's going to be a pretty
Josh:
interesting thing when we get into like scaling these models and using them for code.
Josh:
EJ, as we talked last week about how good the Grok model is for coding because
Josh:
it was so cheap and so effective.
Josh:
This is another case of that. And the way they did that, I was so interested
Josh:
in how they were able to come up with like the secret sauce to do it.
Josh:
And basically what they did is they taught to model to spend its brainpower
Josh:
only on tools when it helps.
Josh:
So they use this like large scale reinforcement learning to train Grok for fast
Josh:
to choose when to think in depth and when to answer questions quickly.
Josh:
So what that resulted in was about 40% what we're seeing here on the screen.
Josh:
40% fewer thinking tokens on average than we've gotten from the previous model,
Josh:
which is a significant difference. Oh, and by the way, it's number one on Elmarina.
Josh:
So this was crazy. EJs, what were your reactions when you saw the team drop this?
Ejaaz:
I already thought these tokens were cheap. I thought these models were cheap enough.
Ejaaz:
Do you remember when OpenAir released GPT-5? They kept flexing GPT-5 mini saying,
Ejaaz:
hey, you now have the power of our previous best model, but actually it's more intelligent.
Ejaaz:
And it's like, I think it was something like five times cheaper.
Ejaaz:
And I was like, holy shit, holy crap. I was like, that is like crazy magnitude.
Ejaaz:
And now it's like, now we've got 47X cheaper than Grok 4.
Ejaaz:
Grok 4, by the way, was already cheap compared to some of the Frontier bottles.
Ejaaz:
So I don't know how far this can go, but kind of zooming out,
Ejaaz:
I have never been more confident that,
Ejaaz:
than now that cutting edge super intelligence will be available for anyone and everyone.
Ejaaz:
This isn't going to be some kind of closeted technology where only the rich
Ejaaz:
can buy devices and run it.
Ejaaz:
I think anyone and everyone will have fair game access to this.
Ejaaz:
And think about the dynamics that that changes up, Josh.
Ejaaz:
Like you can have someone in the complete middle of nowhere with a cell phone
Ejaaz:
attached to Elon Musk's new 5G Starlink satellite that's beaming down to him.
Ejaaz:
And he could kind of produce something that the world ends up using because
Ejaaz:
he has access to this cheap model that is actually super intelligent and can
Ejaaz:
be used to create whatever crazy invention that he has or she has that dreams
Ejaaz:
up. I just think this is insane.
Josh:
Yeah, the efficiency improvements are the thing that's always most exciting
Josh:
to me because, I mean, as we get more cheaper tokens and as the tokens become
Josh:
more portable and lightweight, I mean, you could have the world of knowledge
Josh:
on your phone even without necessarily an internet connection because these
Josh:
models are getting so lightweight, so condensed.
Josh:
Um so effective it's like it's really it's
Josh:
unbelievably impressive and what i was really interested in
Josh:
is comparing this to the other models because i know google was
Josh:
kind of doing a similar thing they were leading along the frontier and oh yeah
Josh:
here's this post from gavin baker that i love because it shows how google has
Josh:
kind of dominated this thing called the pareto frontier and on the chart you
Josh:
can very clearly see how there's this kind of arc that hugs the outer bounds
Josh:
of all of the models and it shows that like gemini pro has been really good on a few
Josh:
So I briefly want to just talk about the Pareto Frontier concept because it's
Josh:
really interesting and it will explain to you exactly why Grok4Fast is way out there.
Josh:
It totally shattered what it is. So, I mean, basically, it's funny.
Josh:
I was doing a little bit of research on this and the Pareto Frontier is done
Josh:
by an Italian economist named Velfredo Pareto.
Josh:
So I just thought that was a fun fact because great name. Basically,
Josh:
it comes from the economist and decision theory.
Josh:
And it's a way to decide optimal trade-offs when you have multiple objectives
Josh:
you're trying to achieve all at the same time. So imagine you're trying to optimize
Josh:
two things that might conflict a little bit, like you want to make a product
Josh:
as powerful as possible, but also inexpensive as possible, like these models.
Josh:
So in this scenario, there's a set of best possible solutions where you can't
Josh:
improve one aspect, like the power, without making the other aspect, like the cost worse.
Josh:
And what we're seeing in this chart here is Google has made a series of those
Josh:
decisions, those tradeoffs that
Josh:
have led to the absolute Pareto optimal outcome along this outer band.
Josh:
What grok has done is they actually made a new trade-off
Josh:
that isn't necessarily a trade-off it's more of an innovation that allows
Josh:
them to unlock this perceived frontier this limiting factor
Josh:
that was on the outer band and just shatter it and create a new pareto
Josh:
optimal trade-off using these best things and they did that by doing a lot of
Josh:
magic but basically what they have now is they have a really smart model that
Josh:
actually sits above gemini 2.5 flash and not too far below the pro model but
Josh:
it is a order of magnitude cheaper and i think that's where that outlier that
Josh:
cost effectiveness is really unbelievable when it comes to,
Josh:
um distributing these tokens widely so now if you're writing code if you're
Josh:
creating an application if you're just if you're paying for tokens this is very
Josh:
clearly the model you want to use.
Ejaaz:
What you just described is elon and
Ejaaz:
xai literally charting a new path which
Ejaaz:
is kind of like um very behavioral of elon in general um and another thing that
Ejaaz:
i thought was really cool about this is the reinforcement learning infrastructure
Ejaaz:
team was kind of key behind getting this model as fast and as cheap and as efficient
Ejaaz:
as we're describing it, right, Josh?
Ejaaz:
They used this kind of like agent framework, which was extremely compatible
Ejaaz:
with the infrastructure that they used to train and iterate on this model in the first place.
Ejaaz:
And what I wanted to point out here is there's a theme between the two topics
Ejaaz:
that we've discussed so far on this episode, Josh.
Ejaaz:
Number one, when we described the two models that the researchers created that
Ejaaz:
broke the ArcGIS benchmark, they specifically used a technique which used reinforcement
Ejaaz:
learning, a new reinforcement learning technique.
Ejaaz:
And the reason, if you remember, why Jeremy Berman picked Grok4 specifically
Ejaaz:
was he said it was the best reasoning model because in the way that had been
Ejaaz:
trained via reinforcement learning.
Ejaaz:
And now we're seeing yet again, this GrokFast model achieving what it can because
Ejaaz:
of reinforcement learning.
Ejaaz:
So I'm seeing a theme or noticing a theme here where XAI and Elon are basically
Ejaaz:
the leaders in reinforcement learning,
Ejaaz:
which i think is going to probably play in their favor maybe it's a hint that
Ejaaz:
the models that are going to be closest to agi that are the quickest that are
Ejaaz:
the cheapest are embedded in reinforcement learning techniques that are just
Ejaaz:
completely breakthrough
Josh:
Yeah it seems like they the team really reasons
Josh:
i mean this is a core elon uh notion but
Josh:
like they really do reason from first principles and what's important and what matters and
Josh:
you're seeing that throughout the entire product as they advance and
Josh:
i think what's really exciting what i'm most stoked about um
Josh:
for this show in general is is to compare this
Josh:
next round of models like will gemini 3
Josh:
and grok 5 like how are they
Josh:
going to compete with each other because those are both going to be remarkable
Josh:
models and it seems to me like those are like those are currently the top dogs
Josh:
i mean as far as gpt5 was kind of a little bit of a miss anthropic's been a
Josh:
little bit quiet uh gemini and xai are on fire but this also there was there
Josh:
was one last thing of news before we sign off today well i.
Ejaaz:
Was I was gonna say, like I'm highlighting this sentence here for those who are just listening.
Ejaaz:
And it says, you know, we built this reinforcement learning infrastructure team
Ejaaz:
with a new agent framework to help train Grok4 fast.
Ejaaz:
But specifically so that we can harness the power of Colossus 2.
Ejaaz:
And if I remember correctly, Josh, there was some breaking news around Colossus
Ejaaz:
2. Elon was getting into some fights. Can you walk us through it?
Josh:
Yeah, it's funny. There was this whole report from SemiAnalysis,
Josh:
which does a really great job. I would highly recommend checking them out.
Josh:
And they released this report on the XAI data center build out.
Josh:
And it was so funny to see, because a lot of times you just see satellite pictures
Josh:
or you read headlines and you're not really sure what's going on.
Josh:
The sole purpose of SemiAnalysis is to actually have boots on the ground,
Josh:
check the satellite images, and look at it with a scientific engineering point
Josh:
of view where they actually understand what is going on.
Josh:
And they shared their findings in one of these articles. And I found one of
Josh:
these stories was so funny because it's such a testament to how the XAI team
Josh:
works, where they were having problems with their energy generation in Memphis,
Josh:
Tennessee, because people were complaining and they were having a tough time getting permits.
Josh:
And the core crux of every large AI data center is energy.
Josh:
So they were like, this is unacceptable. well, we need energy immediately.
Josh:
So what do they do? Well, they jumped over the state lines.
Josh:
They went over to Mississippi a couple of miles down the road and they built
Josh:
these new generators right down the road across the state line.
Josh:
They got the permits they needed.
Josh:
They said, we don't, you don't want us, Tennessee. We'll just go right over to Memphis.
Josh:
You could see here, they took the power lines, they ran them back into Tennessee
Josh:
and now they're powering the data center.
Josh:
And part of the article was, was this funny story, but part of the article also is,
Josh:
is colossus 2 um being built
Josh:
in the sheer scale that colossus 2 is going to be and it's going
Josh:
to be over a gigawatt of energy um which
Josh:
is i don't know how many hundreds of thousands of homes is
Josh:
going to power but this is like a remarkable amount of power and a tremendous
Josh:
amount of gpus and they're planning to make these all coherent and they're using
Josh:
them exclusively i believe to train this new grok 5 model so as this new training
Josh:
center comes online they will be using this new cutting-edge world's largest
Josh:
supercomputer to train the world's perceivedly best model.
Josh:
But I found this funny because the day that this article came out,
Josh:
there was another post from another CEO of a very prominent company saying,
Josh:
hey, wait a second, we have something a little bit bigger.
Josh:
Than Colossus 1 currently. And that was from Microsoft's CEO, Satya Nadella.
Josh:
And he had this post where he said they just added over two gigawatts of new energy capacity.
Josh:
So Ejaz, this is just a really crazy brawl between these people who are building
Josh:
larger and larger AI data centers.
Josh:
And it eventually leads to the big news that just dropped a little earlier today.
Josh:
But do you have any commentary before we get to the huge number?
Ejaaz:
Yeah, there's actually one thing I wanted to point out, which is when Elon first
Ejaaz:
announced that he was building out this Colossus 2 data center,
Ejaaz:
it made headlines that it cost $20 billion.
Ejaaz:
And everyone thought it was crazy. Everyone was yelling, this is an AI CapEx bubble.
Ejaaz:
There is no products that prove that all this investment makes sense.
Ejaaz:
And now you have Satya Nadella, CEO of Microsoft, announcing that he's probably
Ejaaz:
going to be investing twice as much of that to build two gigawatts of new capacity.
Ejaaz:
Again, validating that there is a need for energy and compute to train these new models.
Ejaaz:
Don't forget that Microsoft last week acquired a random European data center for,
Ejaaz:
I think it was about, what, $10 billion, which caused its stock price to 3x
Ejaaz:
because it itself wasn't worth that much at the time of the reporting happening.
Ejaaz:
And then it leads us to the even bigger announcement, which released this morning,
Ejaaz:
which is NVIDIA will be investing not one,
Ejaaz:
not 10, not 20, but $100 billion in OpenAI over the next couple of years.
Ejaaz:
And you might be asking why?
Ejaaz:
Well, it's because OpenAI is going to be investing in so many data centers that
Ejaaz:
is going to produce so much power. I don't know how many gigawatts.
Ejaaz:
I think it's actually 10 gigawatts which is 10x colossus 2 5x um uh fair water
Ejaaz:
which is satya nadella's thing for all my mathematician mathematician fans out
Ejaaz:
there it is just crazy josh are we in a bubble or is there a need for all of this
Josh:
So here's the thing as i keep going back
Josh:
and forth about the bubble conversation because a hundred billion
Josh:
dollars is such an outrageous amount of money to spend on making
Josh:
what is already a remarkable language model even more
Josh:
remarkable um like the product is great and
Josh:
at least me personally as a user of these
Josh:
products i'm definitely getting closer to a wall of things that i use them for
Josh:
where if a model is marginally smarter my experience doesn't get that much better
Josh:
um but i was so that's like one school of thought and then the other is thinking
Josh:
well this is probably the only thing we'll ever need to spend money on going forward ever.
Josh:
So it makes sense to throw all of it at it now.
Josh:
Because in the case that you do solve AGI, you get hyperintelligence,
Josh:
it solves all of your problems. And it gives you the better questions to ask
Josh:
in order to solve better problems. So it really...
Josh:
It would appear, assuming that we continue on this trajectory of improvement,
Josh:
that it makes sense to take every disposable dollar you can to get better and better compute.
Josh:
And this will probably just extend forever. As we are able to harness more energy
Josh:
from the sun, from nuclear energy, a lot of that new energy and compute will
Josh:
just go to making better AI, which will then serve better downstream effects for how society works.
Josh:
So is it a bubble on the long term? I think absolutely not on the short term.
Josh:
I don't know. Where do you get the revenue from?
Josh:
I don't know. I mean, it's a ton of money, but you know what?
Ejaaz:
I think the reason why you and I feel this disassociation between the amount,
Ejaaz:
how large these numbers are in investing in infrastructure versus what we're
Ejaaz:
actually seeing is we're not going to be seeing AGI before some other fields
Ejaaz:
or some other professions see it first, right?
Ejaaz:
The clear example is coding. Coding has just
Ejaaz:
been on an absolutely exponential improvement rate
Ejaaz:
that has beaten out any other ai feature ever you now
Ejaaz:
have ai models that can code as well
Ejaaz:
as a senior staff engineer which is getting paid like
Ejaaz:
300 to 500 a year right um so my guess is this investment is worth it um my
Ejaaz:
guess is the investment is going to come to fruition in professions in use cases
Ejaaz:
in jobs that we won't see but we'll maybe talk about or see the kind of like effects.
Ejaaz:
Maybe it's in science where we create a new drug that cures cancer or whatever that might be, right?
Ejaaz:
I think different types of professionals will see AGI and reap the rewards of
Ejaaz:
this investment before average consumers see it.
Ejaaz:
And then I think the other thing that I want to mention, Josh,
Ejaaz:
is this isn't specific to US or Western spending.
Ejaaz:
In fact, our foes over the seas in China or in Asia have been working on this
Ejaaz:
for like the last five years.
Ejaaz:
They've been building out massive data centers, which I think has like build
Ejaaz:
up an aggregate of like 300 gigawatts over the next five years, at least.
Ejaaz:
And they've been investing in this so heavily. So it's not just a Western thing.
Ejaaz:
It's an Asian thing as well. China is investing so heavily in
Ejaaz:
this if this is a bubble if we are completely
Ejaaz:
wrong this will be the uh biggest most highest profile l that the world has
Ejaaz:
taken it's not just going to be a us thing it's not just going to be a map something
Ejaaz:
it's not just going to be a sam altman thing it's going to be an everyone's
Ejaaz:
involved type of thing kind of like world ending event yeah
Josh:
Too big to fail so i i do love
Josh:
this incentive structure where everyone is incentivized to make it work because
Josh:
everyone's equally at risk in terms of their exposure to the technology so that
Josh:
i think i could be happy to sleep at night where at least u.s and china are
Josh:
aligned in one thing in which they want to achieve agi they want the smartest
Josh:
models they're going to make their money pay off the best they can so,
Josh:
All the power to them. But I think, is that a wrap for us today,
Josh:
EJ? We got anything else?
Ejaaz:
That is a wrap,
Josh:
Josh. That's it. That's a wrap on our little like XAI mini episode.
Josh:
There was one fact that I wanted to just do a little like fun fact check, which is a gigawatt.
Josh:
And according to Grok, it powers approximately 750,000 to 850,000 average US homes per one gigawatt.
Josh:
So the scale we're talking is like a tremendous amount of gigawatts.
Josh:
I mean, this NVIDIA project is 10 of those, which means that's about,
Josh:
I mean, on the high end, eight and a half million U.S.
Josh:
Homes can be powered by a singular data center. So we're going to hope this works out.
Josh:
I think right now it seems like, I mean, Grok is cooking. The XAI team is on
Josh:
fire and they are in between models.
Josh:
I cannot wait until they get this new Colossus training cluster up or even Microsoft's.
Josh:
I mean, Microsoft's got a huge cluster.
Josh:
What are you doing with it, dog? Like, let's see. Let's see your stats.
Josh:
Let's see your numbers. Put a number up on the ArcGIS leaderboard. um
Josh:
but yeah i think that's that's a wrap on all the fun exciting new
Josh:
things about xai the comment section is by energy stocks yeah
Josh:
by energy stocks um i we read
Josh:
all the comments i read every single comment i try to reply to them too so i
Josh:
would love for you to share either what you think about the show or or who
Josh:
you think is winning this ai race currently do you like are
Josh:
we just kind of like do we have elon derangement syndrome are
Josh:
we just kind of like obsessed with everything he builds or is this it feels
Josh:
like it's pretty grounded i feel like we have some good examples about how well
Josh:
they're doing so i'd love to hear if you agree or disagree that
Josh:
would be a fun little thing for the comments but anyway that's a wrap on today's episode
Josh:
we have a couple more exciting ones coming this week so buckle up uh
Josh:
the next one the next one coming i think ejaz myself
Josh:
and we might even have a guest for that episode we'll be probably in an all-out
Josh:
brawl it's good that we're recording remotely because we might like the blood
Josh:
could possibly be drawing so yeah buckle up for for that one there's a lot to
Josh:
look forward to this week but that's it for this week this episode so thank
Josh:
you so much for watching as always please don't forget to subscribe like comment,
Josh:
all the fun things, share it with your friend.
Music:
Music
