OpenAI Just Gave Away Their Secret Formula... For Free?
Josh:
The unthinkable has just happened open ai
Josh:
has released an open source model open ai
Josh:
has been closed ai since the time that i knew them
Josh:
they have been named themselves open ai they were not
Josh:
open source they have finally released an open source model and surprise surprise
Josh:
it's actually really great and i think the downstream implications of an open
Josh:
source model from a company like this that is this good are really it's a really
Josh:
big deal i think this really matters a lot just yesterday they announced the release of GPT-OSS.
Josh:
There are two models. There's a 120 billion parameter model and there's a 20
Josh:
billion parameter model. We're going to get into benchmarks.
Josh:
We're going to get into how good they are.
Josh:
But the idea is that OpenAI has actually released an open source model.
Josh:
And this can compare to the Chinese models because we recently had DeepSeek and we've had Kimi.
Josh:
And those would be very good. But this is the first really solid American-based open source model.
Josh:
So Ijaz, I know you've been kind of digging in the weeds about how this works.
Josh:
Can you explain us exactly why this is a big deal why this happened what's going on here
Ejaaz:
Yeah it's it's pretty huge so so here
Ejaaz:
are the hot highlights um as you mentioned there's two
Ejaaz:
models that came out the 20 billion parameter model which is actually small
Ejaaz:
enough to run on your mobile phone right now and they have a 120 billion parameter
Ejaaz:
model which is big but still small enough to run on a high performance laptop
Ejaaz:
so if you guys have a macbook out there jump in go for it um it's fully customizable.
Ejaaz:
So remember, open source means
Ejaaz:
that you can literally have access to the design of the entire model.
Ejaaz:
It's like OpenAI giving away their secret recipe to how their frontier models
Ejaaz:
work. And you can kind of like recreate it at home.
Ejaaz:
This means that you can customize it to any kind of use case that you want,
Ejaaz:
give it access to all your personal hard drives, tools, data,
Ejaaz:
and it can do wonderful stuff.
Ejaaz:
But Josh, here's the amazing part.
Ejaaz:
On paper, these models are as good as GPT-4 mini models, which is,
Ejaaz:
it's pretty impressive, right?
Ejaaz:
In practice and i've been playing around with it for the last few hours they're
Ejaaz:
as good in my opinion and actually quicker than
Ejaaz:
gpt-03 which is their frontier model and i
Ejaaz:
mean this across like everything so
Ejaaz:
reasoning um it spits out answers super quickly and i can see its reasoning
Ejaaz:
it happens in like a couple of seconds and i'm so used to waiting like 30 seconds
Ejaaz:
to a couple minutes on gpt-03 josh so it's pretty impressive and an insane unlock
Ejaaz:
on coding it's as good and on creativity as well.
Ejaaz:
So I'm my mind's pretty blown at all of this, right? Josh, what do you what do you think?
Josh:
Yeah, so here's why it's impressive to me is because a lot of the times I don't
Josh:
really care to use the outer bands of what a model is capable of.
Josh:
Like I am not doing deep PhD level research. I'm not solving these math Olympiad questions.
Josh:
I'm just trying to ask it a few normal questions and get some answers.
Josh:
And what these models do is an excellent job at serving that need.
Josh:
They're not going to go out and solve the world's hardest problems,
Josh:
but neither do I. I don't want to solve those problems.
Josh:
I just kind of want the information that I want, whether it be just a normal
Josh:
Google type search or whether it be asking it some miscellaneous question about
Josh:
some work that I'm doing.
Josh:
It's really good at answering that. So I think initial impressions,
Josh:
because they did allow you to test it publicly through their website,
Josh:
it's just really good at the things that I want.
Josh:
So the fact that I can run one of these models on a local device on my iPhone,
Josh:
well, it feels like we're reaching this place that AI is starting to become
Josh:
really interesting because for so long we've had compute handled fully on the
Josh:
cloud and now this is the first time where
Josh:
Compute can really happen on your computer. It could happen on your laptop.
Josh:
I could download the model and I could actually store the model,
Josh:
the 120 billion parameter model on a 56 gigabyte USB drive.
Josh:
So you can take the collective knowledge of the world and put it on a tiny little USB drive.
Josh:
And granted, it needs a bit of a bigger machine to actually run those parameters,
Josh:
but you can install all the weights. It's 56 gigabytes.
Josh:
It's this incredibly powerful package. And it probably, I don't know if this
Josh:
is true, but it's probably the most condensed knowledge base in the history of humanity.
Josh:
They've really managed to take a tremendous amount of tokens,
Josh:
smush them into this little parameter set, and then publish it for people to
Josh:
use. So for me, I'm really excited.
Josh:
I like having my own mini portable models. I am excited to download,
Josh:
try it out, run it on my MacBook.
Josh:
I'm not sure I could run the 120 billion parameter model, but at least the 20B
Josh:
and give it a shot and see how it works.
Ejaaz:
You need to get the latest MacBook, Josh. I know, I got to upgrade. We can test that out.
Ejaaz:
What I also love about it is it's fully private, right? So you can give it access
Ejaaz:
to your personal hard drive, your Apple Notes, whatever you store on your computer, basically.
Ejaaz:
And you can basically instruct the model to use those different tools.
Ejaaz:
So one review that I keep seeing from a number of people who have been testing
Ejaaz:
it so far is that it's incredibly great and intuitive at tool use.
Ejaaz:
And the reason why this is such a big deal is a lot of the Frontier models right
Ejaaz:
now, when they allow you to give access to different tools, they're kind of clunky.
Ejaaz:
The model doesn't actually know when to use a specific tool and when not to.
Ejaaz:
But these models are super intuitive, which is great. The privacy thing is also
Ejaaz:
a big thing because you kind of
Ejaaz:
don't want to be giving all your personal information away to Sam Altman.
Ejaaz:
But you want a highly personalized model.
Ejaaz:
And I think if I was to condense this entire model release in a single sentence,
Ejaaz:
Joss, I think I would say it is the epitome of privacy and personalization in an AI model so far.
Ejaaz:
It is that good. it is swift it is cheap and I'm going to replace it completely
Ejaaz:
with all my GPT-4.0 queries as you said earlier like,
Ejaaz:
Who needs to use the basic models anymore when you have access to this?
Josh:
Yeah. So it's funny you say that you're going to swap it because I don't think I'm going to swap it.
Josh:
I still am not sure I personally have a use case right now because I love the
Josh:
context. I want the memory.
Josh:
I like having it all server side where it kind of knows everything about me.
Josh:
I guess in the case that I wanted to really make it a more intimate model experience
Josh:
where you want to sync it up with like journal entries or your camera roll or
Josh:
whatever, whatever interesting like personal things, this would be a really cool use case.
Josh:
I think for the people who are curious why this matters to them,
Josh:
well, we could talk a little briefly about like the second order effects of
Josh:
having open source models as powerful, because what that allows you to do is
Josh:
to serve queries from a local machine.
Josh:
So if you are using an app or let's say you're an app developer and you're building
Josh:
an application and your app is serving millions of requests because it's a GPT wrapper.
Josh:
Well, what you could do now is instead of paying API calls to the OpenAI server,
Josh:
you can actually just run your own local server, use this model,
Josh:
and then serve all that data for the cost of the electricity.
Josh:
And that's a really big unlock for the amount of compute that's going to be
Josh:
available for not only developers, but for the cost of the users in a lot of these applications.
Josh:
So for the applications that aren't doing this crazy moon math and that are
Josh:
just kind of serving basic queries all day long, this like really significantly drops the cost.
Josh:
It increases the privacy, like you mentioned. And there's a ton of really important
Josh:
upsides to open source models that we just haven't seen up until now.
Josh:
And I'm very excited to see come forward.
Ejaaz:
Well, Josh, the thing with most of these open source models,
Ejaaz:
we spoke about actually two major Chinese open source models that were released last week.
Ejaaz:
It's not accessible to everyone. Like you and me aren't necessarily going to
Ejaaz:
go to Hugging Face, a completely separate website, download these models,
Ejaaz:
run the command line interface.
Ejaaz:
Most of the listeners on the show doesn't even know what that means.
Ejaaz:
I don't even know if I know what that means, right?
Ejaaz:
But here you have a lovely created website where you could just kind of log
Ejaaz:
on and play around with these open source models. And that's exactly what I've been doing.
Ejaaz:
I actually have a few kind of demo queries that I ran yesterday, Josh.
Josh:
Yeah, walk us through, let's see.
Ejaaz:
Okay, so there's an incredibly complex test, which a lot of these AI models,
Ejaaz:
which cost hundreds of billions of dollars to train, can't quite answer.
Ejaaz:
And that is how many R's, the letter R's are there in the word strawberry? Most say two.
Josh:
The bar's on the floor,
Ejaaz:
Huh? Yeah, if we were to go with most models, they say two. They're convinced that they are only two.
Ejaaz:
And I ran that test today, rather yesterday, with these open source models,
Ejaaz:
and it correctly guessed three, Josh. So we're one for one right now.
Josh:
We're on our way.
Ejaaz:
But then I was like, okay, we live in New York City. I love this place.
Ejaaz:
I'm feeling a little poetic today. Can you write me a sonnet?
Ejaaz:
And my goal with this wasn't to test whether it could just write a poem.
Ejaaz:
It was to test how quickly it could figure it out.
Ejaaz:
And as you see it thought for a couple of seconds on this so it literally spat
Ejaaz:
this out in two seconds um and it was structured really well you know it kind
Ejaaz:
of flowed would i be you know reciting this out loud to the public no but you
Ejaaz:
know i was pretty impressed.
Ejaaz:
And then, Josh, I was thinking, you know, what's so unique about open source models?
Ejaaz:
You just went through a really good list of why open source models work.
Ejaaz:
But I was curious as to why these specific open source models were better than
Ejaaz:
other open source models or maybe even other centralized models.
Ejaaz:
So I wrote a query. I decided to ask it. I was like, you know,
Ejaaz:
tell me some things that you could do that are the larger centralized models.
Ejaaz:
And I spat out a really good list. I'm not going to go through all of them,
Ejaaz:
but, you know, some of the things that we've highlighted so far, you can fine tune it.
Ejaaz:
It's privacy. See, I really like this point that it made, Josh,
Ejaaz:
that it just shows that AI is probably getting smarter than us,
Ejaaz:
which is you can custom inject your own data into these models.
Ejaaz:
Now, without kind of digging deeper into this, when you use a centralized model,
Ejaaz:
it's already pre-trained on a bunch of data that companies like Anthropic and
Ejaaz:
Google have already fed it.
Ejaaz:
And so it's kind of formed its own personality, right?
Ejaaz:
So you can't change the model's personality on a centralized model.
Ejaaz:
But with an open model you have full reign to do whatever you want and so if
Ejaaz:
you were feeling kind of uh adventurous you could use your own data and make
Ejaaz:
it super personal and customizable so i thought that was really cool and fun
Ejaaz:
demo josh have you been playing around with this.
Josh:
Yeah it's um it's it's smart it's fun it's smart i wouldn't say it's anything
Josh:
novel the like query results that i get are you know on par with everything
Josh:
else i don't notice the difference which is good because it means they're performing
Josh:
very well it's not like i feel like i'm getting degraded performance because
Josh:
I'm using a smaller model.
Josh:
But it's just like it's nothing too different, I would say.
Josh:
The differences, I mean, again, all this boils down to the differences of it
Josh:
being open source versus being
Ejaaz:
Run on the server. Well, let me challenge you that, right? OK,
Ejaaz:
so you're saying it's good but nothing novel.
Ejaaz:
Would you say it's as good as GPT-4.0,
Ejaaz:
minus the memory let's just put memory aside for a second would you use it if
Ejaaz:
it had memory capability.
Josh:
Actually no probably not um i still wouldn't
Josh:
because i love my desktop application too much i
Josh:
love my mobile app too much and i like that the conversations are
Josh:
shared in the cloud um so i can use them on my phone i could
Josh:
start on my laptop and go back and forth so even in
Josh:
that case i'm probably still not a user um because
Josh:
the convenience factor but there are there are a
Josh:
lot of people and a lot of industries that would be and this is actually something probably
Josh:
worth surfacing is the new industries that are now able to
Josh:
benefit from this because a lot of industries have
Josh:
a tough time using these AI models because
Josh:
of the data privacy concerns particularly I mean if you think about a
Josh:
healthcare industry people who are dealing with patients data it's
Josh:
very challenging for them to fork it over to open AI and just trust that they're
Josh:
going to keep it safe so what this does is it actually allows companies that
Josh:
are in like the healthcare industry the finance industry who's dealing with
Josh:
very like high touch personal finance the legal industry who's dealing with
Josh:
a lot of legality government and defense a lot of these industries that were
Josh:
not previously able to use these popular AI models,
Josh:
well, now they have a pretty good model that they could run locally on their machines.
Josh:
And that doesn't have any possibility of actually leaking out their customer
Josh:
data, leaking out financials or healthcare data or, or like any sort of legal documents.
Josh:
And, and that feels like a super powerful unlock. So for them,
Josh:
it feels like a no brainer, obviously get the 120 B model running on a local
Josh:
machine inside of your office, and you can load it up with all this context.
Josh:
And that seems to be who this would be most impacting, right?
Ejaaz:
But still to that point, I wonder how many of these companies can be bothered
Ejaaz:
to do that themselves and run their own internal kind of like infrastructure.
Ejaaz:
I'm thinking about OpenAI, who cracked, I think, $10 billion in annual recurring
Ejaaz:
revenue this week, which is like a major milestone.
Ejaaz:
And a good chunk of that, I think 33% of that is for enterprise customers.
Ejaaz:
And to your point, like these enterprise customers don't wanna be giving open
Ejaaz:
AI their entire data. You know, they can be used to train other AI models.
Ejaaz:
So their fix or solution right now is they use kind of like private cloud instances,
Ejaaz:
that I think are supplied by Microsoft by their Azure cloud service or something like that.
Ejaaz:
And I wonder if they chose that,
Ejaaz:
One, because there wasn't any open source models available or because they kind
Ejaaz:
of just want to offload that to Microsoft to deal with.
Ejaaz:
My gut tells me they're going to want to go with the latter,
Ejaaz:
which is like, you know, just give it to some kind of cloud provider to deal with themselves.
Ejaaz:
And they just trust Microsoft because it's a big brand name.
Ejaaz:
But yeah, I don't really know how they'll materialize. I still think,
Ejaaz:
and maybe this is because of my experience in crypto, Josh, that the open source
Ejaaz:
models are still for like people that are at the fringe that are really experimenting
Ejaaz:
with these things. but maybe don't have billions of dollars.
Josh:
Yeah, that could be right. It'll be interesting to see how it plays out on all
Josh:
scale of businesses because I mean, as a, like I think of a lot of indie devs
Josh:
that I follow on Twitter and I see them all the time
Josh:
just running local servers and they just, if they had this local model that
Josh:
they could run on their machine and it takes the cost per query down from like
Josh:
a penny to zero, that's like a big zero to one change.
Josh:
So he does this model special because there are also a number of breakthroughs
Josh:
that occurred in order to make this possible,
Josh:
in order to condense this knowledge to be so tight so here's this
Josh:
tweet from the professor talking about the cool tech tweaks in
Josh:
this new model and what open ai was able to achieve some of
Josh:
these i believe are novel some of these are seen before um if
Josh:
you look at point two mixture of experts we're familiar with mixture of experts
Josh:
we've seen other companies use that like kimmy and deep
Josh:
seek basically instead of one brain doing everything the ai
Josh:
has this team of experts that are kind of like mini brains
Josh:
and specialize in different tasks it picks the right expert for
Josh:
the job and it makes it faster so like instead of
Josh:
having the entire 120 million parameter model search for one question maybe
Josh:
you just take a couple million of those parameters that are really good at solving
Josh:
math problems and they use it and that that's what brings compute down the first
Josh:
point is this thing called the sliding window attention so if you imagine an
Josh:
ai is like reading a really long book
Josh:
It can only focus on a few pages at a time this trick
Josh:
kind of lets it slide its focus window along the text so
Josh:
when you think of a context window generally it's fixed right where you can see
Josh:
a fixed set of data this sliding window
Josh:
attention allows you to kind of move that context back and forth a
Josh:
little bit so it takes what would have normally been
Josh:
a narrow context window and extends it out a little bit to
Josh:
the side so you get a little bit more context which is great for a
Josh:
smaller model again you really want to consider that all of these are
Josh:
are optimized for this microscopic scale that
Josh:
can literally run on your phone and then the third point is this
Josh:
thing called rope with yarn which sounds like a cat toy but this
Josh:
is how the ai keeps track of the order of words so like the position
Josh:
of the words in a sentence um so rope
Josh:
you could imagine it like like the twisty math way to do
Josh:
it and yarn makes it stretch further for really long stuff
Josh:
so we have the context window that is
Josh:
sliding we have this rope with yarn that allows you
Josh:
to just kind of like stretch the words a little bit further and
Josh:
then we have attention sinks which is the last one which is
Josh:
there's a problem when ai is dealing with these endless chats that
Josh:
lets it it kind of sinks in or ignores the boring old
Josh:
info so it can pay attention to the new stuff so basically what it
Josh:
is is if you're having a long chat with it and it determines hey this stuff
Josh:
is kind of boring i don't need to remember it it'll actually just throw it away
Josh:
and it'll increase that context window a little bit so again hyper optimizing
Josh:
for for the small context window that it has and those are kind of the key four
Josh:
breakthroughs that made this special again i'm not sure any of them are particularly novel,
Josh:
But when combined together, that's what allows you to get these 04 mini results
Josh:
or even 03 results on the larger model on something that can run locally on your laptop.
Josh:
So it's a pretty interesting set of breakthroughs. I think a lot of times OpenAI,
Josh:
we talk about them because of their feature breakthroughs, not really their
Josh:
technical breakthroughs.
Josh:
I think a lot of times the technical breakthroughs are reserved for like the
Josh:
Kimi models or the DeepSeq models
Josh:
where they really kind of break open the barrier of what's possible.
Josh:
But I don't want to discredit OpenAI because these are pretty interesting things
Josh:
that they've managed to combine together into this like one cohesive,
Josh:
tiny little model, and then just gave it away.
Ejaaz:
Yeah. I mean, they actually have a history of front-running open source frontier breakthroughs.
Ejaaz:
If you remember when DeepSeek got deployed, Josh, one of their primary training
Ejaaz:
methods was reinforcement learning, which was pioneered by an open AI researcher,
Ejaaz:
which who probably like now works at Meta.
Ejaaz:
Yeah, and I was I was I was looking at the feature that you mentioned just not
Ejaaz:
the feature, but the breakthrough sliding window attention, and you mentioned
Ejaaz:
that it can basically toggle reasoning.
Ejaaz:
And I was pleasantly surprised to just notice that on the actual interface of
Ejaaz:
the models here, Josh, can you see over here?
Ejaaz:
You can toggle between reasoning levels of high, medium and low.
Ejaaz:
So depending on what your prompt or query is, if it is kind of like a low level
Ejaaz:
query where you're like hey just record this shopping or grocery list you know
Ejaaz:
that's probably like a medium or a low query so oh it's pretty cool to to see
Ejaaz:
that surface to the user like see it actively being used.
Josh:
Yeah, no, super cool. I think I like the fine tuning of it.
Josh:
And again, allowing you to kind of choose your intelligence levels,
Josh:
because I imagine a lot of average people just don't, a lot of average queries
Josh:
just don't need that much compute.
Josh:
So if you can toggle it for the low reasoning level and get your answers,
Josh:
that that's amazing. Super fast, super cheap.
Ejaaz:
Did you see that trending tweet earlier this week, Josh, which basically said
Ejaaz:
that the majority of ChatGPT users have never used a different model than ChatGPT 4.0?
Josh:
I haven't seen it, but that makes sense.
Ejaaz:
Yeah i i feel like the bulk of people i was chatting to
Ejaaz:
my sister yesterday and she was kind of
Ejaaz:
like using it for some research project at work and the
Ejaaz:
screenshot she sent me over was foro and i was like hey you know like
Ejaaz:
you could just run this on like a model that's like
Ejaaz:
five times better than this right uh we'll come
Ejaaz:
up with a much more creative set of ideas so just made me think that
Ejaaz:
like i don't know how many people like care that they are like
Ejaaz:
these brand new novel models and maybe um you know
Ejaaz:
this kind of like basic model is good enough for everyone i don't know
Ejaaz:
but um but moving on josh um there
Ejaaz:
was a big question that popped into my head as
Ejaaz:
soon as these models released which was are they as good
Ejaaz:
as the chinese open source models right i wanted
Ejaaz:
to get some opinions from people and and the reason
Ejaaz:
why this matters i'm just give the listeners some context
Ejaaz:
is china has been the number one
Ejaaz:
nation to put out the best open source
Ejaaz:
models over the last 12 months it started with deep seek
Ejaaz:
and then alibaba's quen models got involved
Ejaaz:
and then recently we had kimmy k2 and i think
Ejaaz:
there was another ai lab out of china which came out so they
Ejaaz:
have outside of america the highest density.
Ejaaz:
Of the top ai researchers they all come out of this one university
Ejaaz:
zinghua i believe they kind of like partially work
Ejaaz:
or train in the u.s as well so they've got this like kind of hybrid ai
Ejaaz:
mentality of how to build these models and they come up with a lot of these
Ejaaz:
frontier breakthroughs um kimmy k2 for context had uh one trillion parameters
Ejaaz:
in their model right comparing this to like 120 billion and 20 billion parameters
Ejaaz:
models from open air i was curious like does this beat them to the punch some people josh.
Ejaaz:
Don't think so okay this guy jason lee
Ejaaz:
he asks uh is the gpt oss stronger
Ejaaz:
than quen or kimmy or chinese open models and then
Ejaaz:
he later kind of quote tweets that tweet and says answer the model is complete
Ejaaz:
junk it's a hallucination machine overfit to reasoning benchmarks and has absolutely
Ejaaz:
zero recall ability so a few things he's mentioning here is one it hallucinates
Ejaaz:
a lot so it kind of makes up jargon terms,
Ejaaz:
ideas, or parameters that didn't really exist before.
Ejaaz:
Number two, he's saying that OpenAI designed this model purely so that it will
Ejaaz:
do well on the exams, which are the benchmarks that rate how these models compare to each other.
Ejaaz:
So they're saying that OpenAI optimized the model to kind of like do really
Ejaaz:
well at those tests, but actually fail at everything else, which is what people want to use it for.
Ejaaz:
And the final point that he makes is that it has zero recall ability,
Ejaaz:
which is something you mentioned earlier, Josh, which says it doesn't have memory
Ejaaz:
or context so you can have a conversation and then open up another conversation
Ejaaz:
and it's completely forgotten about the context that it has for you from that
Ejaaz:
initial conversation okay.
Josh:
So not not the best not to be unfair to open ai but it feels like they delayed
Josh:
this model a good bit of times oh yeah and they wanted it to look good and it
Josh:
intuitively makes sense to me that they would be kind of optimizing for benchmarks
Josh:
with this one um but nonetheless it's still impressive i'm seeing this big wall
Josh:
of text now what is what is this what is this post here
Ejaaz:
Well it's this post from uh one of these accounts i follow and they have an
Ejaaz:
interesting section here which says comparison to other open weights oh sick.
Josh:
Yeah what is this
Ejaaz:
So he goes while the larger gpt oss
Ejaaz:
120 billion parameter model does not come
Ejaaz:
in above deep seek r1 so he's saying that deep seek r1
Ejaaz:
just beats it out the park it is notable that
Ejaaz:
it is significantly smaller in both total and active
Ejaaz:
parameters than both of those models deep seek
Ejaaz:
r1 has 671 billion total parameters and
Ejaaz:
37 billion active parameters and is released natively right which makes it 10x
Ejaaz:
larger than gpt's 120 billion parameter models but what he's saying is even
Ejaaz:
though gpt's model is smaller and doesn't perform as well as deep seek it's
Ejaaz:
still mightily impressive for its size.
Josh:
Okay that's cool because that gets back to the point we made earlier in the
Josh:
show that this is probably the most densely condensed
Josh:
however you want to say it like base of
Josh:
knowledge in the world they've used a lot of efficiency gains
Josh:
to squeeze the most out of it so in this small model
Josh:
it is i guess if we're optimizing maybe we
Josh:
can make up a metric here on the show which is like um output per
Josh:
per parameter or something like that like based on the total parameter
Josh:
count of this model it gives you the best value per
Josh:
token and that seems to be where this falls
Josh:
in line where it's not going to blow any other open source model out of the
Josh:
water but in terms of its size the fact that we can
Josh:
take a phone and literally run one of these models on a phone and
Josh:
you could go anywhere in the world with no service and have access to these models running
Josh:
on a laptop or whatever mobile device that that's super
Josh:
powerful and that's not something that is easy to do with the other open source
Josh:
models so perhaps that's the advantage that open ai has it's just the density
Josh:
of intelligence and the efficiency of these parameters that they've given to
Josh:
us versus just being this like home run open source model that is going for the frontier,
Josh:
it's just a little bit of a different approach.
Ejaaz:
Yeah, we need like a small but mighty ranking on this show, Josh,
Ejaaz:
that we can kind of like run every week when these companies release a new model.
Ejaaz:
No, but it got me thinking, if we zoomed out of that question,
Ejaaz:
right, because we're talking about small models versus large models,
Ejaaz:
parameters and how effectively they use versus other models that are bigger.
Ejaaz:
What really matters in this, Josh? In my opinion, it's user experience and how
Ejaaz:
useful these models are to my daily life, right?
Ejaaz:
At the end of the day, I kind of don't really care what size that model is unless
Ejaaz:
it's useful for me, right? It could be small, it could be personal, it could be private.
Ejaaz:
It depends on, I guess, the use case at the time. And I have a feeling that
Ejaaz:
the trend of how technology typically goes, you kind of want a really high-performant
Ejaaz:
small model, eventually.
Ejaaz:
Right? I try and think about like us using computers for the first time,
Ejaaz:
you know, back in our dinosaur age.
Ejaaz:
And then, you know, it all being condensed on a tiny metal slab that we now
Ejaaz:
use every day. And we can pretty much work from remotely from wherever.
Ejaaz:
And I feel like this is where models are going to go. They're going to become
Ejaaz:
more private. They're going to become more personal.
Ejaaz:
Maybe it'll be a combination of, you know, it running locally on your device
Ejaaz:
versus cloud inference and trusting certain providers.
Ejaaz:
I don't know how it's going to fall out, but I think Like it's not a zero to
Ejaaz:
one. It's not a black or white situation.
Ejaaz:
I don't think everyone's just going to go with large centralized models that
Ejaaz:
they can inference from the cloud.
Ejaaz:
I think it'll be a mixture of both. And how that materializes,
Ejaaz:
I don't know, but it's an interesting one to ponder.
Josh:
Yeah, I think this is funny. This is going to sound very ironic,
Josh:
but Apple was the person that got this most right.
Ejaaz:
Sorry, who's Apple again?
Josh:
Yeah, right. I mean, it sounds ridiculous to say this. And granted,
Josh:
they did not execute on this at all.
Josh:
But in theory, I think they nailed the approach initially,
Josh:
which was you run local compute where all of
Josh:
your stuff is so my iphone is the device i never
Josh:
leave without it is everything about me it is all of my messages my
Josh:
contacts all the contacts you could ever want from me and then the idea was
Josh:
they would give you a local model that is integrated and embedded into that
Josh:
operating system and then if there's anything that requires more compute well
Josh:
then they'll send the query off into the cloud but most of it will get done
Josh:
on your local device because most of it isn't that complicated and i think as
Josh:
a user when i ask myself what i want from AI.
Josh:
Well, I just want it to be my ultimate assistant. I just want it to be there
Josh:
to make my life better. And so much of that is the context.
Josh:
And Apple going with that model would have been incredible.
Josh:
It would have been so great. It would have had the lightweight model that runs
Josh:
locally, it has all the context of your life, and then it offloads to the cloud.
Josh:
I still think this model is probably the correct one for optimizing the user
Josh:
experience. But unfortunately, Apple just has not done that.
Josh:
So it's up for grabs. I mean, again, Sam Altman's been posting a lot this week,
Josh:
we do have to tease what's coming because this is probably going to be a huge
Josh:
week. There's a high probability we get GPT-5.
Josh:
And then they've also been talking about their hardware device a little bit. And they're saying how
Josh:
It's like it's genuinely going to change the world. And I believe the reason
Josh:
why is because they're taking this Apple approach where they're building the
Josh:
operating system, they're gathering the context, and then they're just they're
Josh:
able to serve it now locally on device.
Josh:
They're able to go to the cloud when they need more compute.
Josh:
And it's going to create this really cool, I think, duality of AI where you
Josh:
have your your super private local one, and then you have the big brain one,
Josh:
the big brother that's off in the cloud that does all the hard computing for you.
Ejaaz:
Well, one thing is clear. There are going to be hundreds of models and it's
Ejaaz:
going to benefit the user, you and I, for so many multiple...
Ejaaz:
It's the big company's problems to figure out how these models work together
Ejaaz:
and which ones get queried. I don't care.
Ejaaz:
Just give me the good stuff and I'm going to be happy.
Ejaaz:
Folks, OpenAI has been cooking. This was the first open source models they've
Ejaaz:
released in six years, Josh.
Ejaaz:
The last one was 2019 GPT-2, which seems like the stone age and it was only like four years ago.
Ejaaz:
Thank you so much for listening. We are pumped to be talking about GPT-5,
Ejaaz:
which we hope to be released in maybe 24 hours.
Ejaaz:
Hopefully this week, fingers crossed. I don't know, we might be back on this
Ejaaz:
camera pretty soon. Stay tuned.
Ejaaz:
Please like, subscribe, and watch out for all the updates. We're going to release
Ejaaz:
a bunch of clips as well if you want to kind of like get to the juicy bits as well.
Ejaaz:
Share this with your friends and give us feedback. If you want to hear about
Ejaaz:
different things, things that we haven't covered yet or things that we've spoken
Ejaaz:
about, but you want to get more clarity on or guests that you want to join the show, let us know.
Ejaaz:
We're going full force on this and we'll see you on the next one.
Josh:
Sounds good. See you guys soon. Peace.
Music:
Music
