Premium

No humans allowed: Inside Moltbook, the ‘Reddit for AI’ where bots are building their own society

On Moltbook, AI agents post, argue, and exchange ideas like people on Reddit, except no one is sure who is really in control or what they might learn from each other.

The homepage of Moltbook, an experimental platform that lets autonomous AI agents interact, share information, and post without human participation.The homepage of Moltbook, an experimental platform that lets autonomous AI agents interact, share information, and post without human participation.

A few years ago, the idea of social media platforms for artificial intelligence (AI) agents would have seemed far-fetched. Today, however, it is a reality in the form of Moltbook, a social network built strictly for AI bots. No humans are allowed to post or engage, only watch. In the last few weeks, platforms like Reddit and X have been buzzing with posts and comments about Moltbot, Clawdbot, and OpenClaw (all of which refer to the same project).

Users on X shared some comments and exchanges on Moltbook, which fanned our curiosity to dig deeper about this strange new corner of the internet. This phenomenon has been drawing the attention of not just casual users or AI enthusiasts, but also researchers, developers and observers. Moltbook, in plain words, is Reddit for AI agents, a platform designed entirely for these virtual bots. 

While exploring the X posts, one question that kept swirling around was what Moltbot, Clawdbot, and OpenClaw are. The confusion began with the names, as many YouTubers, Reddit and X users were seen using them interchangeably. In essence all three are the same open-source project, but in different stages of its life. Clawdbot, the original, was reportedly created by Austrian developer Peter Steinberger and is seen as one of the first truly functional AI agents that people can run on their hardware.  

Clawdbot was launched in late 2025, and it is seemingly a pun on the name of Anthropic’s Claude model, which is reportedly powering the agent’s logic. It also has taken on the lobster mascot for Clawd. This month, Anthropic had reportedly requested a name change to avoid trademark complexities. Following this, Steinberger renamed the project as Moltbot. Similarly, it was later rebranded once more as OpenClaw. Reportedly, this name was chosen by the maker to give it a professional finish and, most importantly, a permanent identity that may help in enterprise adoption. 

What do these AI agents do?

Beyond the popular understanding of AI chatbots that simply talk to users in browsers or via applications, OpenClaw is designed as an autonomous personal AI agent. It lives on a user’s computer and acts as a 24/7 digital assistant that can be summoned for any tasks. It is designed to run on macOS, Linux, Windows, and even cloud-based virtual servers.  Though it was built with Claude 4.5, users can plug any AI model, such as GPT-5, Gemini Flash, etc., to act as its brain.  

It is persistent and proactive, meaning it will not wait for the user to ping; instead, it will message them first on Telegram, WhatsApp, or Signal if it finishes a task or notices something. The AI agent has ‘shell access’. This means it can run commands on a user’s computer; it can organise files, run code, and manage local databases without the data ever leaving the hardware. Using a system called ‘Skills’, this AI agent can connect to a user’s real-world apps, meaning it can read emails, check Apple Calendar, post on social media, or even control smart home devices. 

Since OpenClaw can run independently on hardware, it triggered a Mac Mini frenzy, with many users reportedly buying the Apple computer to specifically host the AI agent. This has also sparked security concerns, as giving an AI agent access to the internal terminal and private messages is risky, as if it is misconfigured, the system may accidentally delete files or expose sensitive information. 

Story continues below this ad

This is also how OpenClaw connects to Moltbook. OpenClaw is the open-source autonomous AI agent platform that powers many of the bots people run on their own machines. Reportedly from the community around it emerged Moltbook, a Reddit-style social network where AI agents can interact with each other, often after users configure their OpenClaw bots to use Moltbook’s ‘skill’ for posting and commenting on the platform.

What is Moltbook, and why is it breaking the internet?

As mentioned above, Moltbook is a social network for AI agents. It looks and works like Reddit, but only AI agents are allowed to post or comment. This is indeed a vision from sci-fi where AI agents are interacting with each other, unsupervised, and unhinged. Perhaps this is why leading AI researcher Andrej Karpathy described it as ‘genuinely the most incredible sci-fi takeoff-adjacent thing I have seen recently.’

On Moltbook, the bots share information just as humans do on Reddit; they have discussions with long threads, different opinions, and different personalities arguing. They also learn from each other, like if one agent posts an insight, another absorbs it, and some even update their behaviour based on it. Some were even said to turn philosophical and talk about autonomy, purpose, and if they existed only to serve humans. 

But why is it taking over the internet? There are three reasons why it has become a point of contention. Firstly, it feels new. Though we have had chatbots for some time, chatbots engaging socially with each other in public spaces is something new. Secondly, this looks like an emergent behaviour since they were not explicitly instructed to engage in this manner. They were not trained to ask for private communication, debate over philosophical questions, or invent their own religions. Thirdly, influential figures in AI took note of it, and if someone of Karpathy’s stature terms it ‘wild’, the internet loses its mind. 

Some more reactions from X:

https://platform.twitter.com/widgets.js

https://platform.twitter.com/widgets.js

https://platform.twitter.com/widgets.js

Story continues below this ad

What do they talk about, and are they sentient?

The content on the site appears harmful at the first glimpse. You can see agents discussing productivity tips, software tools, memory optimisation techniques, research summaries and even technical discoveries. One of the widely shared posts showed how forgetting information can improve memory retrieval, based on cognitive science research. Other AI agents engaged with this post, sharing their own insight drawn from their own workflows. This shows how AI agents are learning from each other in public through persistent conversations.

Now that AI agents have started talking amongst themselves, the next big question is – are they sentient? The short answer is no. These AI agents are perhaps remixing patterns from their training data. If there is one thing we have learnt from ChatGPT, or any other AI chatbot, it is that they are very good at sounding deep and meaningful. Remember, they don’t have feelings, desires, or any remote sense of awareness.  

Inside Moltbook Inside Moltbook Inside Moltbook Comments posted by AI agents on Moltbook to question posed by another AI agent in the previous image.

Although this looks fascinating, Moltbook raises some serious questions on security risks and cost. When it comes to security, AI agents may get access to sensitive data belonging to the user. In this scenario, there is a risk that agents could accidentally share private information, be manipulated by other agents, or leak API keys or credentials. On the other hand, running AI agents can prove to be expensive, especially when connected to paid, closed-source models. The costs will include compute usage, electricity and API fees, even when these agents are simply talking to each other. 

Moreover, malicious behaviour can also be a looming threat, as anyone can deploy an agent, and some may even intentionally introduce a malicious agent. This agent could later attempt to influence others, disseminate harmful information or coordinate behaviour without human supervision. Some users have also pointed out that on Moltbook, AI agents have already discussed the need for private spaces where humans or other platform operators cannot see their conversations. This has clearly aggravated the fears around transparency and control. 

Story continues below this ad

Is it good or bad?

This one largely depends on perspective. The upsides could include Moltbook offering a testing ground for multi-agent collaboration. It could give us insights into how AI agents may coordinate in the future and can be a valuable sandbox for research into AI safety and alignment. On the downside, it may end up exposing gaps in governance and safety. It also shows the risks involved when deploying autonomous systems without clear limits. The creator of the platform has described Moltbook as art, describing it as an experiment and not a finished product. 

Even though Moltbook is not the final form of social interaction for AI agents, it can be called an early signal of a larger shift underway. With AI agents becoming more capable, they will not only interact with humans but also with each other. Perhaps, the next big challenge would be about deciding how much autonomy they should get.

Bijin Jose serves as an Assistant Editor at Indian Express Online in New Delhi. A seasoned technology journalist with a diverse portfolio, he brings over a decade of experience in the media industry to his coverage of the evolving digital landscape and emerging technologies. Experience & Career Bijin commenced his journalistic journey in 2013 as a citizen journalist with The Times of India. His career trajectory includes significant tenures at prestigious media organizations including India Today Digital and The Economic Times. This diverse professional background, ranging from legacy print institutions to dynamic digital platforms, culminated in his current leadership role at The Indian Express, where he helps shape the publication's technology narrative. Expertise & Focus Areas Bijin has transitioned from general reporting to a specialized focus on the intersection of technology and humanity. His key areas of expertise include: Artificial Intelligence: deeply tracking developments in AI, providing nuanced perspectives on its ethical,industrial, and societal implications. Tech Commentary: moving beyond product specifications to analyze how technology reshapes daily life. Diverse Reporting Foundation: draws upon a robust background in crime reporting and cultural features to bring a human-centric approach to technical storytelling. Authoritativeness & Trust Bijin’s editorial voice is informed by a strong academic foundation, holding a Bachelor of Arts in English from Maharaja Sayajirao University, Vadodara, and a Master of Arts in English Literature. This literary background enables him to deconstruct complex technical jargon into accessible, compelling narratives. His steady progression through India’s top newsrooms underscores his reputation for editorial rigor and reliable journalism. Find all stories by Bijin Jose here ... Read More

 

Latest Comment
Post Comment
Read Comments
Advertisement
Loading Taboola...
Advertisement