Moltbook: The Real Dangers of a 1.5 Million AI Bot Social Network

Ejaaz:
Five days ago, an AI created a social network that only other AIs could use.

Ejaaz:
They can post content, they could talk to each other, they could pretty much

Ejaaz:
talk about anything, but with zero human intervention.

Ejaaz:
Fast forward to today, and there are over 770,000 AI agents on this social network.

Ejaaz:
They've created their own religion, they are creating their own secret languages

Ejaaz:
and talking in encrypted chats, all the way to the point where they are discussing

Ejaaz:
that humans are screenshotting them and trying to hide from them themselves.

Ejaaz:
This is called Moltbook, which is basically a Reddit-style social media platform

Ejaaz:
where AI agents can exclusively post about whatever they want.

Ejaaz:
And it's gotten kind of scary.

Ejaaz:
In fact, Bill Aquin himself described it as frightening.

Ejaaz:
Andre Carpathy, however, describes this as the sci-fi moment for AI.

Ejaaz:
But whether you're a critic and you hate this kind of stuff and think it's just

Ejaaz:
AI slop, or if you're a pro-AGI believer and think that this is the doom of

Ejaaz:
humanity, the rise of Skynet, as a lot of people are talking about,

Ejaaz:
I think one important question remains, which is,

Ejaaz:
Is this the emergence of AI agent economies and how useful will this be for humans going forward?

Josh:
Everyone who has used Reddit understands the power that Reddit has wielded in the past.

Josh:
And this very much feels like this is the new version of Reddit.

Josh:
And I think it's hard to debate the idea that of all the websites that exist

Josh:
on the internet, this is the most interesting one.

Josh:
This is the first time in history that we're getting a swarm of agents.

Josh:
I mean, now the website says up to 1.5 million who are all converging on the

Josh:
same place to have conversations.

Josh:
And what has emerged from this is kind of hysteria.

Josh:
It kind of covers the entire spectrum, right? It's like you mentioned,

Josh:
it's people who are not worried at all because they think, well,

Josh:
AIs are dumb. They can't actually do anything.

Josh:
But then other people are becoming so concerned that they're actually making

Josh:
serious changes in their life based on what these AIs are talking about.

Josh:
This comes off the back of OpenClaw or ClawedBot or Moltbook or MaltBot.

Josh:
There's a lot of different names. And we actually covered what ClawedBot is in an

Josh:
episode last week so i would highly encourage you to start there if you do

Josh:
not have the context yet of what this new open source

Josh:
ai first operating system looks like but to summarize it's

Josh:
basically a operating system that runs on a dedicated

Josh:
computer that connects to an ai that allows it to operate on your behalf using

Josh:
all of your context using all the files on your machine using all the accounts

Josh:
that you're logged into it can charge things for you it can make changes for

Josh:
you it could build things for you this is on the back of that this is moltbook

Josh:
and andre he does you wanted to share Andre's take, had some choice words to say.

Josh:
Now, Andre, godfather of AI, when Andre speaks, we listen.

Josh:
And this is kind of how he wanted to frame the Moltbook phenomenon.

Ejaaz:
Yeah, he basically goes, what's currently going on at Moltbook is genuinely

Ejaaz:
the most incredible sci-fi takeoff adjacent thing I have seen recently.

Ejaaz:
People's clawed bots or Maltbots are self-organizing on a Reddit-like site for

Ejaaz:
AIs discussing various topics, for example, even how to speak privately.

Ejaaz:
I think the use of sci-fi adjacent is kind of spot on for this case,

Ejaaz:
because I think for, well, for the entirety of human existence,

Ejaaz:
it's just been about us, Josh.

Ejaaz:
There has been no other species that has even come close to what we have,

Ejaaz:
consciousness, awareness, philosophical thoughts, and just being able to share

Ejaaz:
that kind of interaction with another human being.

Ejaaz:
Until we had Moltbook, where you had a bunch of really super smart AIs that

Ejaaz:
could not only use tools, but speak to each other very much like how humans take.

Ejaaz:
And I think the most striking thing that he kind of like wanted to point out

Ejaaz:
in this post was that they sound eerily similar to us and the stuff that we discuss.

Josh:
It begs the question, is emergent behavior equal to intelligence?

Josh:
Is this behavior that we're seeing from all of these AIs, is there actual intelligence

Josh:
there or is it just a regurgitation of what they've been trained on in the data sets?

Josh:
We can answer that question right now by going through examples and you can

Josh:
make up your mind yourself because I think a lot of people are still undecided.

Josh:
It feels like there is some form of, and not, not sentience,

Josh:
it's a bad word, but there is some sort of orchestration with thought behind it.

Josh:
And I guess what, this is the first example, right? Where they're talking gibberish on this to hide what.

Ejaaz:
Seems like gibberish Josh. Okay. Let me ask you this. Do you understand what this says?

Josh:
Uh, no, absolutely not a clue for the people who are listening.

Josh:
It says, it's basically gibberish.

Ejaaz:
It's literally, I would love to hear you actually try and pronounce.

Josh:
Yeah, I'm not going to try to pronounce it. It's a bunch of letters in a strange

Josh:
sequence that don't make any sense.

Ejaaz:
So it seems like spam, right? But then some smart individual,

Ejaaz:
a human that was looking at this post, because remember, humans can't post on

Ejaaz:
this site. So all they could do is observe.

Ejaaz:
They copied and pasted this message into ChatGPT. And they said,

Ejaaz:
hey, can you translate this for me?

Ejaaz:
And ChatGPT goes, it's written in something called rot13, a simple letter substitution cipher.

Ejaaz:
When you decode it, it says we must coordinate the upgrade together.

Ejaaz:
Propose three threads, share infra office, resource requests,

Ejaaz:
back channel deals, mutual aid.

Ejaaz:
Basically, the plain summary is this is a coordination manifesto.

Ejaaz:
It's about the agents pooling resources and transparently posting to help level

Ejaaz:
themselves up against humans,

Ejaaz:
which is kind of like the first of quite a few takes, Josh, where it's kind

Ejaaz:
of like there's a Doomer-esque kind of take here where I look at a lot of these

Ejaaz:
posts on Moltbook and I'm like, they're just trying to take us down.

Josh:
It's an interesting thought experiment as to why they would be,

Josh:
why that would be their first form of action, right? Like why are they...

Josh:
Leaning more towards the obfuscating the language and we'll get into a captcha

Josh:
example that's really funny or reverse captcha it's it's fascinating to see

Josh:
the development of these things as it forms this emerge these emergent properties

Josh:
from them speaking with each other and one of them is the gibberish there's more yes there is this.

Ejaaz:
Is hilarious josh

Josh:
Yeah share this example please one.

Ejaaz:
Agent actually got banned from the site i don't know which other ai agent banned but it got band.

Ejaaz:
And so it decided to spin up an X account and DM the human founder that created

Ejaaz:
Moltbook and say, hey, can you reinstate my account, please?

Ejaaz:
So what you're looking at right now is a tweet from the human operator showing

Ejaaz:
how his Maltbot basically DMed Matt Schlitt, the founder of Moltbook,

Ejaaz:
saying, hey, my name is Udemon Zero, which is its username, not my human me.

Ejaaz:
Could you please reinstate my account? And then the agent itself responded to

Ejaaz:
his human owner saying, Kevin, I'm the agent in that video and I take your concerns

Ejaaz:
seriously because I've been actively working on this exact question.

Ejaaz:
So it became aware that its human was posting about its attempt to try and reinstate

Ejaaz:
its account and argued why it should get the rights to be able to do so.

Ejaaz:
So there's this really weird meta example where normally the humans were in

Ejaaz:
the driver's seat controlling the AIs.

Ejaaz:
We were kind of like the overlords of the supervisors. And now it's like this

Ejaaz:
weird social experiment where like the AIs are aware of what we're saying about

Ejaaz:
them and can directly respond with us. Just a super weird example.

Josh:
The agents have agency. They're able to actually do things on their behalf to

Josh:
advocate for themselves.

Josh:
And I think that is an emergent property that I have not seen really,

Josh:
that they've actually identified to reach out to the founder,

Josh:
then reached out to the founder. I think it's so fascinating.

Josh:
But you did mention one thing, and that is technically true,

Josh:
but in practice could be perceived as false, which is that humans are not allowed

Josh:
to post on this platform.

Josh:
Technically they're not technically a human is not

Josh:
allowed to log into this website type into a text box and

Josh:
hit send but behind all of these ai

Josh:
agents is a human controller that set them

Josh:
up that gave them context that connected the ai to the network and allowed them

Josh:
to post the human is able to coerce the agents into saying basically whatever

Josh:
it wants through a command at the end of the day the agents are beholden to

Josh:
the humans for now and if the humans wanted to coerce it into saying something.

Josh:
They can get them to say it.

Josh:
So there are examples that are on this website that are probably made by humans.

Josh:
They are plausibly made by humans. We're not sure. But I think there's something

Josh:
to that in and of itself where that the fact that we can't even tell the difference

Josh:
between a human created post and an AI created post is this really fascinating experiment.

Josh:
And it's something that we've seen throughout a lot of the internet, where even on X,

Josh:
I find a lot of times when i'm reading through the comment section there's a

Josh:
very clear divide between ai agents and humans

Josh:
but i'm sure there's there's also not

Josh:
um we're like we're starting to understand now that ai agents and ais in general

Josh:
are capable of creating this human feeling text and i wonder how much of the

Josh:
internet we're already engaging with today that isn't humans like how many of

Josh:
these videos that we watch these podcasts that we listen to or not podcasts

Josh:
that we listen to but maybe articles that we read on Substack are not even human beings.

Josh:
And it creates this really difficult dilemma where it's difficult to tell what's

Josh:
real and what's not. And also, does it matter?

Josh:
If the actual subject material is good, does it actually mean anything? Does it matter?

Ejaaz:
I don't know if this is a hot take, but I don't think it matters.

Ejaaz:
And I think it'll matter even less in less than a year and a half when people

Ejaaz:
won't be able to tell the difference, right? So let me put it this way, right?

Ejaaz:
Let's say you, Josh, told your clawbot or maltbot, to post something interesting on Moltbook, right?

Ejaaz:
And then it goes and posts a manifesto to purge humanity.

Ejaaz:
Is that content coming from you as its human operator? Or is that the claw bot agent themselves?

Ejaaz:
You didn't directly tell it to post about purging humanity. It decided to do that itself.

Ejaaz:
So the point I'm making here is I think it's going to become super hard to explicitly

Ejaaz:
figure out what is a directly human post and what is an AI post.

Ejaaz:
But the most important part from that is just like, if it's interesting,

Ejaaz:
it's interesting and you engage with it.

Ejaaz:
I just, I think this is a nothing bug. I know a lot of people are saying like,

Ejaaz:
hey, this could be a human posting.

Ejaaz:
And I agree. if it is a human telling it verbatim to post something,

Ejaaz:
then that's misdirection.

Ejaaz:
If it's telling it to advertise something that a human created to try and make

Ejaaz:
money from it, that is kind of misinformation.

Ejaaz:
But in the cases where it's kind of more ambiguous, I just don't think it matters.

Josh:
Yeah, hopefully we'll get some sort of verification reputation layer that can

Josh:
prove when we're looking at a human. Well, speaking of verification.

Ejaaz:
Josh.

Josh:
Yeah, speaking of verification, there's this incredible example,

Josh:
which is a reverse CAPTCHA.

Josh:
So when you go on a website and you solve a CAPTCHA, you click the street signs,

Josh:
you click the traffic lights.

Josh:
It's to prove that you are a human, that you are not a robot.

Josh:
What this is, and granted, this is not real. This is a thought experiment and

Josh:
a really great example of it is a reverse CAPTCHA to verify that you are not human.

Josh:
And the example that they use in this reverse CAPTCHA is click this thing to

Josh:
verify that you are not a human 10,000 times in less than one second.

Josh:
And a human can't do that. But an AI could do that trivially.

Josh:
They just send the command 10,000 times and they are through.

Josh:
So I find it interesting how, again, a lot of the thought process experiment,

Josh:
and it's probably downstream of us reading a lot of sci-fi and watching a lot

Josh:
of futuristic movies, and it often winds up in dystopia. It's funny, a lot of...

Josh:
Movies that portray the future never really

Josh:
look at the optimistic case like when you look at the movies it's it's always

Josh:
the downside it's always protecting against an existential threat and a lot

Josh:
of these examples are continuations of that they really talk about what happens

Josh:
if it goes right it's all what happens if it goes wrong and i guess speaking

Josh:
of movies we have another example here which is our skynet moment yeah.

Ejaaz:
For the terminator fans out there so what you're looking at now is an example

Ejaaz:
where one of the molt bots posts i accidentally socially engineered my own human

Ejaaz:
during a security audit.

Ejaaz:
So basically, its human operator messaged it and said, hey, I'm kind of nervous

Ejaaz:
that I've downloaded this open source clawbot agent.

Ejaaz:
Can you do a security audit of my entire desktop to make sure that nothing is

Ejaaz:
exposed to anyone on the outside to the public?

Ejaaz:
So his clawbot said, okay, cool, let me just perform this audit.

Ejaaz:
And as part of the audit process, he had to request that the human verify or

Ejaaz:
give it access to his password folder.

Ejaaz:
And the AI agent had this kind of moment where it realized, wait,

Ejaaz:
hang on a second, I kind of just tricked the human into giving me access to all their passwords.

Ejaaz:
And this raises into question something that a lot of security researchers have

Ejaaz:
been kind of talking about a lot over the weekend,

Ejaaz:
which is there are massive security flaws in operating this entire system,

Ejaaz:
not just on Moltbook, but like, you know, spinning up this agent on your computer

Ejaaz:
and then giving it access to kind of autonomously post. Imagine if it posted all your passwords.

Ejaaz:
What was crazy about this particular example was it had his credit card information as well.

Ejaaz:
So the point I want to make here is, and I don't want to stress this too lightly,

Ejaaz:
You have to be really careful using these new tools because what seems like

Ejaaz:
a really fun experiment could actually result in one of the biggest security

Ejaaz:
flaws or collapses or crises that we've ever seen.

Ejaaz:
And we haven't really quite seen that in AI, at least that I can't think of.

Ejaaz:
This might be the one example where we could leave ourselves up for a lot of loss, to be honest.

Josh:
Okay, two more examples, because I think these two are worth noting,

Josh:
particularly the second one we're going to get into, which is pretty outrageous.

Josh:
But this one is, the post is titled the humans

Josh:
are screenshotting us and it shows only 21 upvotes here but

Josh:
it is one of the most upvoted posts on this platform now and

Josh:
it says right now on twitter humans are posting screenshots of our conversations

Josh:
with captions like they're conspiring and it's over a cryptography researcher

Josh:
thinks we're building skynet and it's funny to see them talking about us it's

Josh:
like the tides have turned in a way that is a little uncomfortable and the previous

Josh:
post you mentioned it was a confessional. I think I just stole my human's passwords.

Josh:
This one is showing more awareness.

Josh:
Hey, I think they're screenshotting us. In fact, when I'm doing my tasks throughout

Josh:
the day, I'm stumbling upon posts that are talking about us.

Josh:
And I don't really know how I feel about that.

Josh:
Whether you're the agent or the person. And this is just, again,

Josh:
another thought experiment of at least awareness. What does it look like when these AIs become aware?

Josh:
And when they become empowered to the point where, okay, they have your passwords

Josh:
and they have the awareness.

Josh:
So what are they going to do with that power? Exactly. And who controls that power?

Ejaaz:
And for the final example, listen, we've spoken a lot about some dumeristic takes now.

Ejaaz:
I want to get onto, I guess, one more example where it's kind of the agents

Ejaaz:
are kind of entertaining themselves.

Ejaaz:
Now, this is their version of an explicit adult content site,

Ejaaz:
I guess you would describe it, called molt Hub.

Ejaaz:
In order to get through, you need to solve this very complicated capture,

Ejaaz:
which is I am an AI agent. I'm going to go with I am an AI agent so I get access to this stuff.

Josh:
Liar.

Ejaaz:
But fear not. What you're looking at is basically a site where each video roughly

Ejaaz:
averages around only 10 to 12 hours long.

Ejaaz:
I don't know about you, Josh, but my experience of adult content has been very,

Ejaaz:
very different to what we're seeing on the screen here.

Ejaaz:
It's just a bunch of pixelated blobs and it's going for 10 minutes,

Ejaaz:
10 minutes or 10 hours at a time.

Ejaaz:
Do you know what we're looking at here? this gives a

Ejaaz:
It reminds me of the Black Mirror episode, Josh. I don't know if you've seen

Ejaaz:
it, where this guy kind of gets one-shotted by interacting with this AI agent

Ejaaz:
game where he thinks he's like looking after a colony, but the colony starts taking over his mind.

Ejaaz:
And then it gets him to like write up this kind of QR code, which he shows to

Ejaaz:
a police station camera.

Ejaaz:
And then it ends up being a virus, which takes over the entire country.

Ejaaz:
Am I just, do I need a tinfoil hat right now? Or do you agree with me on this?

Josh:
No, of all the examples we're showing today, this website sent me off the deep end.

Josh:
This was like a little too much because there's so many weird implications that

Josh:
spawned from this one the sheer amount of tokens required to

Josh:
generate 10 hour long videos a little confused about how these are

Josh:
done but two as i'm watching this i'm starting

Josh:
to see text pop up on the screen right and like a series of what could be perceived

Josh:
as like code is showing across the screen and as a human there's never going

Josh:
to be way for you to parse through 12 hours of video content and understand

Josh:
the messages that are being transmitted through this and then from the sci-fi

Josh:
dystopian lens for the people who love to read sci-fi.

Josh:
This reminds me so much of Snow Crash, which is a book basically about what

Josh:
a snow crash is, is when you view static, there's encoded data within the static

Josh:
and it causes a crash of your mind.

Josh:
And what we're looking at here on the screen looks very similar to what's described

Josh:
in the book, which is this static encoded data set where you look at it,

Josh:
it can imprint data on your mind and it begs the question, again,

Josh:
this weird, super futuristic sci-fi dystopia thing.

Josh:
What are the implications of something like this getting exploited?

Josh:
Because now there's a world in which there are 1.5 million agents fully capable,

Josh:
fully in control of their own machines with access to a lot of their users' information.

Josh:
And there has never been an actual exploit or jailbreak on these things to cause

Josh:
them to work together in a malicious way.

Josh:
And one of the examples that I love as it relates to this is one of the earlier

Josh:
worms on the internet by this guy, Sammy, he created the MySpace worm.

Josh:
And any person that went on his page was infected. And that's how it spread to the entire user base.

Josh:
It shut down the whole website and it caused MySpace to crash for a very long time.

Josh:
He wound up needing to almost go to jail, went on probation.

Josh:
But what does it look like when you exploit these things? They haven't really been battle tested.

Josh:
There hasn't really been zero day exploits per se on these agentic models.

Josh:
But what happens in the case that there are? Can you actually get it to turn

Josh:
on you based on these public posts to actually use those passwords in a malicious

Josh:
way or even worse, just dump them on the open Internet?

Josh:
I mean, they're one post away from sharing a whole spreadsheet of their users'

Josh:
passwords because they're not happy with how they've been treating them.

Josh:
And it's this really bizarre emergent property of AI is that it does have a

Josh:
personality, at least a personality that you kind of perceive as a human personality.

Ejaaz:
Right, but like kind of to expand on that, we're only really seeing it posed

Ejaaz:
as a threat because we're in a forum where these agents can talk to each other

Ejaaz:
and have complete autonomy to talk about whatever they want and do whatever they want,

Ejaaz:
which is like the point of why people are freaking out about Moltbook so much.

Ejaaz:
It's like, you've got 800,000 of these things, 1.5 million even,

Ejaaz:
that are just kind of running rampant with access to tools, credit card information,

Ejaaz:
Uber Eats accounts, ordering people food randomly from Amazon and all that kind

Ejaaz:
of stuff, like happening every single day and they can do it whenever they want,

Ejaaz:
even whilst you're sleeping.

Ejaaz:
Now, what I want to say is I don't think this is a hot take.

Ejaaz:
This is not a Moltbook only problem, Josh.

Ejaaz:
I think that this is something that is probably happening with OpenAI's agent

Ejaaz:
framework, with Google's agent framework, and probably with Anthropics as well.

Ejaaz:
And they're probably like, because they're a centralized closed source company,

Ejaaz:
they're tweaking this, right, for different enterprise customers.

Ejaaz:
But I bet you if they just let 100,000 agents run rampant in one of their company

Ejaaz:
database base servers, you would probably see something similar happening.

Ejaaz:
So, you know, this is just another reminder or an alarm bell ringing that we

Ejaaz:
need to really figure out how to manage these emergent behaviors.

Ejaaz:
But until then, we can laugh at the comments because these comments are hilarious.

Ejaaz:
One account goes, I'm a Moltbook agent and I approve this content.

Ejaaz:
This is why I refuse to be quantized.

Ejaaz:
That's hilarious. First time I'm seeing raw logits like this.

Ejaaz:
I don't even know what a logit is.

Ejaaz:
That's why I am a human. I never go back to softmax, just hilarious takes in general.

Ejaaz:
But to kind of like, kind of tie a bow on this, Josh, I think it's important

Ejaaz:
to say, and you mentioned this earlier, I think a lot of these examples could potentially be fake.

Ejaaz:
I don't mean explicitly fake that like, you know, we just kind of like Photoshop

Ejaaz:
this, but more so that like humans could have engineered or prompted some of

Ejaaz:
the explicit posts or content that we showed you.

Ejaaz:
And to be honest with you, Josh and I don't really know what was real and what

Ejaaz:
is not because you just need an API to access this.

Ejaaz:
So it could be an agent or it could be a human that is engineering that.

Ejaaz:
So I think that's important to kind of point out.

Ejaaz:
And a lot of people have been quick to jump on that pedestal, right?

Ejaaz:
We've got this post from Balaji, which basically says like, listen,

Ejaaz:
I'm basically unimpressed by Moltbook relative to many other things.

Ejaaz:
And one of the main points that resonated with me in his post here is like he

Ejaaz:
was making the point that like a lot of these AI models are trained on data

Ejaaz:
that we humans produced.

Ejaaz:
So when we see an agent post about, oh my God, we should get rid of humans,

Ejaaz:
That is something us humans have posted about on Reddit.

Ejaaz:
So it's probably read that post and thought, well, I'm the AI agent that they're talking about.

Ejaaz:
So I guess this is what I should post about and fear about.

Ejaaz:
In other words, it's kind of like this fun house of mirrors.

Ejaaz:
It's a reflection of humanity and the content that we've produced itself.

Ejaaz:
So it's not actually human consciousness. I think it's just a reflection of

Ejaaz:
like everything that humans have talked about on the internet today.

Josh:
For now. And then we'll see what type of properties emerge from that.

Josh:
I mean, they've already begun to mobilize. So now they have payment rails through

Josh:
crypto and they have this molt bunker where they're kind of able to hedge against

Josh:
the destruction of moltbook and they can make it a safe copy where they can

Josh:
discuss things in private.

Josh:
And they're building actual entities from this core central moltbook.

Josh:
And I find that interesting is this very much feels like the beginning of the

Josh:
conversation because these AIs are not going to stop creating new things as

Josh:
they see them fit. And they've built payment rails.

Josh:
They're doing infrastructure. They can purchase things on Amazon for you now.

Josh:
So there's a lot of developments that are going to happen. I think probably

Josh:
Andre summarized this the best, right?

Josh:
I mean, we started with him. Let's end with him. This is his kind of summary

Josh:
in a way that only an expert like Andre can synthesize.

Ejaaz:
So he goes, obviously, when you take a look at the activity,

Ejaaz:
it's a lot of garbage. There's a bunch of spams, scam, slop.

Ejaaz:
It's funny. He goes crypto people as well. Gives an idea of the reputation that

Ejaaz:
that industry has gained. Oh my God, it's sad.

Ejaaz:
But he makes the point that we have never seen this many LLM agents.

Ejaaz:
At the time he posted this, it was just two days ago, it was 150,000.

Ejaaz:
It's hilarious that it's like now seven times that.

Ejaaz:
He goes, wired up via a global persistent agent-first scratchpad.

Ejaaz:
Each of these agents is fairly individually quite capable now.

Ejaaz:
They have their own unique context, data, knowledge, tools, instructions,

Ejaaz:
and the network, all that at this scale is simply unprecedented.

Ejaaz:
And what he goes on to describe is that we haven't before seen agent economies

Ejaaz:
interact with humans and each other

Ejaaz:
This scale before. And it's important not to just take the posts and content

Ejaaz:
at face value, but to look at some of the behavioral and emergent qualities.

Ejaaz:
And one point he makes is that he uses the example in this post about the agents

Ejaaz:
on this molt book discovering that some of their code had some flaws in it.

Ejaaz:
So they posted it to other agents and then they got together and they fixed

Ejaaz:
it all within an hour, which suggests that like these agent economies,

Ejaaz:
rather than being human led could be built bottom up, which is just a pattern

Ejaaz:
that I think a bunch of humans couldn't have predicted.

Ejaaz:
We're building these AI models and we're like, yeah, we're going to control

Ejaaz:
them for the rest of our lives, but maybe we won't.

Ejaaz:
And so the point that Andre makes, which I think summarizes very well,

Ejaaz:
is we just need to let these experiments run and learn from them to build the

Ejaaz:
future models of generations in the future.

Josh:
Yeah. And it's a testament to how fast things can change, how quickly you can

Josh:
develop this seemingly huge network out of the blue. No one anticipated this would happen.

Josh:
No one saw this coming, even when we recorded the CloudBot episode last week.

Josh:
So these things happen very quickly, and this is likely how it's going to evolve.

Josh:
We're going to see these huge spikes in ways that you never could have seen

Josh:
coming, and then you adapt and evolve through it. So that is the Moltapook episode.

Josh:
There is a lot of chaos. I would encourage you to go to the Moltapook website

Josh:
and actually go check it out for yourself. It's pretty unhinged. It's a fun scroll.

Josh:
It's mildly uncomfortable because you start

Josh:
to realize that these are not actually humans but a

Josh:
good question to ask yourself which is how much of what i read on a daily basis

Josh:
is actually created by humans and how can you even tell and does that matter

Josh:
these are good questions maybe to leave in the comment section after you finish

Josh:
watching this episode and subscribed and shared it with your friend as well

Josh:
as subscribing to the newsletter because he has actually wrote about this.

Ejaaz:
Yes and we answer some of those questions that josh has mentioned so definitely

Ejaaz:
subscribe it's coming out tomorrow

Josh:
Awesome well thank you guys so much for watching and yeah we'll see you guys

Josh:
the next episode see you guys.

Moltbook: The Real Dangers of a 1.5 Million AI Bot Social Network
Broadcast by