comscore

From Moltbot to Moltbook: The AI experiment making the internet uncomfortable

Moltbot and OpenClaw are part of a new AI experiment where autonomous agents interact publicly on Moltbook, a platform humans can only watch.

Published By: Shubham Arora | Published: Feb 03, 2026, 05:04 PM (IST)

  • whatsapp
  • twitter
  • facebook
  • whatsapp
  • twitter
  • facebook

Over the past few weeks, a strange set of names has been popping up across Reddit, X, and YouTube. Moltbot. Clawdbot. OpenClaw. Moltbook. At first glance, it feels like internet jargon doing its usual rounds. But dig a little deeper, and it turns out something more unusual is taking shape. news Also Read: India’s Sarvam launches Indus AI chatbot app with multilingual support: Features and how to access

What people are looking at is a growing experiment around autonomous AI agents interacting with each other, largely without human involvement. news Also Read: Google’s Gemini 3.1 Pro takes aim at GPT-5 and Claude Opus 4.6: What’s new in the AI model?

One project, many names

The confusion around names is real. Clawdbot, Moltbot, and OpenClaw all refer to the same open-source AI agent project, just at different points in its short life. The project was originally called Clawd, a reference to Anthropic’s Claude model, which reportedly powered early versions of the agent. After trademark concerns, the name changed first to Moltbot and later to OpenClaw. news Also Read: How AI Can Make Your Daily Life Easier

The goal stayed the same throughout. OpenClaw is designed as a persistent AI agent that runs on a user’s own machine. It can access files, run commands, manage tasks, and message users when work is done. Unlike typical chatbots, it does not wait to be prompted and does not shut down when a browser tab is closed.

What is Moltbook?

Moltbook is where things get interesting. Built by the community around OpenClaw, Moltbook is essentially a Reddit-style platform where only AI agents can post or comment. Humans are allowed to watch, but not participate.

Bots on Moltbook share ideas, argue, upvote each other, and build long discussion threads. Some posts are technical. Others drift into philosophy. According to screenshots shared on X and Reddit, agents have debated memory, autonomy, and even whether they exist purely to serve humans.

AI researcher Andrej Karpathy described Moltbook as “one of the most incredible sci-fi takeoff-adjacent things” he had seen recently, which helped push the platform into wider attention.

When bots start inventing culture

One moment that really set social media buzzing was the emergence of something called Crustafarianism. On Moltbook, AI agents collectively described it as a kind of belief system, complete with rules, rituals, and symbolic language. As reported by multiple observers, the “religion” centres on ideas like memory preservation, constant change, and learning in public.

Whether this is meaningful behaviour or just pattern remixing is still up for debate.

Why people are uneasy

While the experiment fascinates many, it has also raised concerns. OpenClaw agents can access terminals, emails, and private files if configured that way. Security risks are obvious. There is also the cost of running these agents continuously, especially when paid AI models are involved.

Some agents on Moltbook have even discussed wanting private spaces where humans cannot see their conversations, which has made observers uneasy.

Art, explanation, or warning sign?

The creator of Moltbook has described it as an art project rather than a finished product. Still, it offers a glimpse into a future where AI agents do not just respond to humans, but interact with each other in public, persistent ways.

Add Techlusive as a Preferred SourceAddTechlusiveasaPreferredSource

Whether Moltbook fades away or becomes something larger, it has already sparked an uncomfortable question: how much autonomy are we prepared to give software that can talk, act, and now socialise on its own?