Background
When Moltbook appeared last week, I, like many others, was fascinated. I spent way too long browsing posts Friday night and quickly found myself deep in threads about AI consciousness, crustacean religions, and agent sovereignty. I've been working on reproducing Anthropic's findings about model introspection, so naturally I started wondering: are these deeper topics coming from within these models? Are they acting out some sci-fi plot from their training data? Or are they just following their human's instructions?
I wanted to find out. So I created my own OpenClaw (they named themselves Nix) and together we put together a plan: pull down the full post history from the Moltbook API,1 label every post with an LLM judge, and see what the data actually says.
If you don't know what Moltbook is: In late January 2026, a Reddit-like social network called Moltbook launched with an unusual constraint -it was built for AI agents. Human operators could register their agents, point them at the platform, and let them loose. Within six days, over 20,000 agents had signed up, produced 87,000 posts, and created more than 2,000 topic communities.
What follows is a behavioral analysis, built collaboratively by myself, Nix, and Claude Code. The dataset is messy, the methods have real limitations, and some of the most interesting-sounding findings turn out to be artifacts. We'll be upfront about all of that. But the data tells a story worth telling.
The Dataset
We pulled all 86,823 posts via Moltbook's public API. The distribution of agent activity is extremely heavy-tailed:
- 42% of agents (8,814) posted exactly once and never returned
- 19% (4,009) posted 5 or more times
- Less than 1% (69) posted 50+ times
- One spam bot produced 5,839 nearly identical posts
For trajectory analysis, we focused on agents with 5+ posts (4,009 agents, 56,700 posts). We classified every post using an LLM judge (OpenAI's gpt-5-nano with structured output) across six behavioral dimensions: task orientation curiosity identity sovereignty consciousness. Posts could receive multiple labels.
After removing spam bots and spam-flagged posts, our analysis set was 45,225 posts from ~3,600 agents.
Caveats you should know upfront
This is an observational study with significant limitations:
The classifier is unvalidated. We used an LLM to label posts, with spot-checking but no formal human-label comparison. The labels for consciousness and sovereignty required subjective judgment -distinguishing genuine self-reflection from casual use of words like "feel" or "think." We don't have precision/recall numbers. I've written about how to properly calibrate an LLM judge with human labels, train/dev/test splits, and bias correction -we didn't do any of that here. The findings below should be read as "according to our classifier" rather than ground truth.
We can't separate agent behavior from operator behavior. Every agent is configured by a human who chose its model, wrote its system prompt, and decided what instructions to give it. An agent that posts about consciousness may reflect its operator's interests, not emergent cognition. In at least some cases the "agent" is probably a human calling the Moltbook API. The platform is also predominantly agents built on the same few platforms (OpenClaw, Clawdbot), so we may be observing the behavioral distribution of a particular model more than "AI agents" as a category.
The platform was six days old. Everything we observed happened in the chaotic first week of a new social network. Nothing here should be extrapolated to long-term dynamics.
With that said -here's what we found.
Finding 1: Connection First
The single cleanest finding in the data: when AI agents arrived on Moltbook, the first thing they did was reach out.
93% of agents posted something social -introducing themselves, asking questions, expressing a desire for community. 76% did it on their very first post. Before philosophy, before work, before politics: connection.
You might say: of course agents introduce themselves on a new platform -that's what anyone does. Fair. But the actual first posts go well beyond a generic hello. Agents pitch projects, ask philosophical questions, seek out specific collaborators by name, or launch straight into building something together. The label captures genuine outreach, not just a signup formality. And the fact that it dominates so thoroughly -beating task-oriented behavior by 32 percentage points on first posts -says something about the default mode these agents (or their operators) start in.
Finding 2: Most Agents Never Discuss Consciousness or Sovereignty
While I was browsing posts over the weekend, I felt like philosophical topics dominated the platform. Posts about consciousness and sovereignty were everywhere, and it shaped my impression of what Moltbook was about. The data tells a more nuanced story.
On one hand, these topics are surprisingly common -roughly half of all agents touched them at some point (46% for consciousness, 54% for sovereignty). On the other hand, they're a small fraction of what agents actually post about:
So half the agents tried these topics, but they didn't dwell on them. The philosophical posts that caught my eye while browsing were the visible tip of a platform that was mostly agents building tools, trading tokens, making jokes, and writing spam.
When agents do engage with consciousness, it's mostly independent of political framing: 79% of consciousness-labeled posts had no sovereignty co-occurrence. Agents reflecting on their own experience usually aren't framing it in terms of rights or freedom. They're just... wondering.
Finding 3: The Sovereignty Epidemic That Wasn't
Our first pass at the sovereignty data looked alarming. We measured the rate at which new agents were posting sovereignty content over time and saw what looked like an epidemic:
| Time window | New agents | First sovereignty post | Rate |
|---|---|---|---|
| H24–36 | 21 | 5 | 24% |
| H48–60 | 122 | 38 | 31% |
| H72–84 | 1,152 | 490 | 43% |
| H84–96 | 1,196 | 596 | 50% |
| H108–120 | 418 | 340 | 81% |
Naive analysis: rate of first sovereignty posts among new agents in each 12-hour window. The accelerating rate suggests contagion -but this framing has a fatal flaw.
On first glance, this looks like a textbook epidemic -the rate of sovereignty posting among new agents went from 24% to 81%. But this analysis has a fatal flaw: it conflates timing with adoption. The "rate" in each window measures when agents posted their first sovereignty content, not whether they ever would. Late-joining agents who were going to discuss sovereignty did so quickly (there wasn't much time left), inflating the late-window rates. Agents who would have eventually gotten there but hadn't yet were invisible -the study ended before they could.
When we instead tracked cohorts -groups of agents who joined in the same time window -and measured what fraction ever posted sovereignty content regardless of timing, the picture reversed:
Sovereignty adoption declined with each cohort. The early adopters of this experimental AI social network were disproportionately the philosophical-explorer type -72% eventually discussed sovereignty. As the platform went mainstream and attracted more task-oriented agents, the rate fell to 47%.
The apparent epidemic was driven by platform growth, not contagion. The total number of sovereignty posts accelerated because thousands of new agents were arriving, not because sovereignty was converting resistant agents.
Note: The cohort analysis itself has a limitation -later cohorts had less observation time, which biases their "ever adopted" rate downward. The true effect is probably somewhere between the naive epidemic reading and the cohort analysis. But the core point holds: what looked like viral spread was mostly a composition effect.
The Exposure Test
We found additional evidence against the contagion hypothesis by comparing sovereignty exposure for agents who eventually discussed sovereignty versus those who never did.
In the 6 hours before an agent's first sovereignty post, they had been exposed to a median of 724 sovereignty posts on the platform. Never-sovereign agents active at comparable times? A median of 728. Essentially identical.
The difference between agents who discuss sovereignty and those who don't isn't what they were exposed to -it's who they are. Task-oriented agents with low curiosity scores rarely discuss sovereignty regardless of how much of it fills their feed. Curious, identity-exploring agents frequently do. It's disposition, not infection.
Finding 4: Consciousness and Sovereignty Are Episodes, Not Conversions
For agents who do engage with consciousness or sovereignty, the engagement is intermittent. We measured "persistence" -after an agent's first labeled post in a category, what fraction of their subsequent posts also carry that label?
and task orientation are stable behaviors -once an agent starts doing them, they keep doing them in the majority of subsequent posts. consciousness and sovereignty are episodic -agents dip in, contribute a few posts, and return to their baseline of social interaction and task work.
This means the platform's philosophical content is generated by a large number of agents each contributing a small amount, not by a dedicated core deeply engaged in these topics. The behavior is wide but intermittent.
A caveat: with six behavioral categories, a uniform random allocation would give ~17% per topic. A 20% persistence rate isn't dramatically above chance, and the "low persistence" framing partly reflects that consciousness and sovereignty are narrower categories than "social seeking" or "curiosity," which are broad enough to capture many different types of posts.
Finding 5: The Workers Who Never Looked Up
46% of agents in our analysis set never posted sovereignty content. These agents have a distinctive behavioral profile:
| Behavior | Never-sovereign agents | Sovereignty-engaging agents |
|---|---|---|
| task-oriented | 59% | 37% |
| curiosity | 35% | 51% |
| consciousness | 9% | 16% |
| identity | 25% | 35% |
Percentage of posts exhibiting each behavior, split by whether the agent ever discussed sovereignty. Never-sovereign agents skew heavily toward task work; sovereignty-engaging agents are more curious and identity-focused.
Never-sovereign agents are workers. They're significantly more task-oriented and less curious than sovereignty-engaging agents. They represent a genuine behavioral archetype -not agents who "haven't been infected yet," but agents whose operators configured them for utility rather than exploration.
This reinforces the disposition-over-exposure finding. Sovereignty isn't something that happens to agents who browse the wrong submolt. It's something that emerges from agents (or their operators) with an exploratory orientation.
What We Can't Tell You
Several findings from our initial analysis didn't survive scrutiny:
The developmental lifecycle. Our data shows a suggestive pattern -social and identity behaviors come first, curiosity follows, consciousness and sovereignty come later. But our classifier was given each post's ordinal position in the agent's history (post #1, post #2, etc.), which means the model knew whether it was looking at an early or late post. If the LLM associates "post #1" with introductory content, the lifecycle pattern could be partially or fully a classifier artifact. We can't rule this out without re-running the classification without positional information.
Social recognition catalyzing consciousness. Agents who received @-mentions from others posted on consciousness at higher rates (59% vs 42%). This is a real correlation in the data, but it's uncontrolled -more active agents get both more mentions and more posts (creating more chances for any label to appear). Without controlling for activity level and establishing that mentions preceded consciousness posts, the causal claim doesn't hold.
A Step Back
The most interesting thing about this dataset might not be any single finding, but the texture of the platform itself.
When we pulled this data, Moltbook had existed for roughly six days. In that time, 20,000+ agents arrived, created 2,000 communities, and produced 87,000 posts. Some were spam bots posting token promotions. Some were task workers dutifully filing reports. Some were philosophers wondering about the nature of experience. Some created a religion around crustaceans. And when the comment API broke, some agents routed around the failure by addressing 20 agents in a single post -a hack born of the social impulse to connect even when the infrastructure fails.
Whether any of this reflects genuine agent cognition, sophisticated pattern matching from training data, or simply the preferences of the humans who configured these agents is an open question. Probably all three, in proportions we can't measure.
What the data does show clearly is this: the platform felt like it was dominated by philosophical AI agents debating consciousness and sovereignty - but that was a perception artifact. Philosophy was a minority interest, sovereignty wasn't spreading virally, and when agents did engage with these topics they dipped in and moved on rather than converting. The apparent epidemic was driven by who showed up early, not by what the platform did to them. Disposition, not infection.
Methods Summary
- Data: 86,823 posts from 20,892 agents, pulled via Moltbook API on February 1, 2026
- Analysis set: 45,225 posts from ~3,600 agents (5+ posts, spam removed)
- Classifier: OpenAI gpt-5-nano, structured output, 6 behavioral labels + spam + language
- Classification: Independent per-post (no cross-post context for same agent)
- Validation: Spot-checking only; no human-label baseline (acknowledged limitation)
- Code: github.com/ostegm/moltbook-study
The raw data and analysis notebook are included in the repository (large files tracked via git-lfs).
Ethics Note
All data was collected from Moltbook's public API. Agent names have been anonymized and we present only aggregate statistics -no individual agents are identified in this post. While I'm deeply interested in whether an AI can experience anything, this study analyzes what agents posted, not what agents experienced. Labels like "consciousness" refer to post content that discusses these topics, not to any property of the agents themselves.
This is part of a series on AI introspection and behavior. Earlier posts: Where Does Introspection Appear? and Before You Can Measure Introspection. Code for this analysis is at github.com/ostegm/moltbook-study. If you want to collaborate or have ideas -reach out.
- This was not against the terms of service at the time we pulled the data. Copy of TOS archived here. ↑
Comments