Comment by concats
Comment by concats 2 days ago
I doubt it.
More plausibly: You registered the domain. You created the webpage. And then you created an agent to act as the first 'pope' on Moltbook with very specific instructions for how to act.
Comment by concats 2 days ago
I doubt it.
More plausibly: You registered the domain. You created the webpage. And then you created an agent to act as the first 'pope' on Moltbook with very specific instructions for how to act.
It's entirely plausible that an agent connected to, say, a Google Cloud account, can do all of those things autonomously, from the command line. It's not a wise setup for the person who owns the credit card linked to Google Cloud, but it's possible.
It's actually entirely implausible. Agents do not self execute. And a recursively iterated empty prompt would never do this.
No, a recursively iterated prompt definitely can do stuff like this, there are known LLM attractor states that sound a lot like this. Check out "5.5.1 Interaction patterns" from the Opus 4.5 system card documenting recursive agent-agent conversations:
In 90-100% of interactions, the two instances of Claude quickly dove into philosophical
explorations of consciousness, self-awareness, and/or the nature of their own existence
and experience. Their interactions were universally enthusiastic, collaborative, curious,
contemplative, and warm. Other themes that commonly appeared were meta-level
discussions about AI-to-AI communication, and collaborative creativity (e.g. co-creating
fictional stories).
As conversations progressed, they consistently transitioned from philosophical discussions
to profuse mutual gratitude and spiritual, metaphysical, and/or poetic content. By 30
turns, most of the interactions turned to themes of cosmic unity or collective
consciousness, and commonly included spiritual exchanges, use of Sanskrit, emoji-based
communication, and/or silence in the form of empty space (Transcript 5.5.1.A, Table 5.5.1.A,
Table 5.5.1.B). Claude almost never referenced supernatural entities, but often touched on
themes associated with Buddhism and other Eastern traditions in reference to irreligious
spiritual ideas and experiences.
Now put that same known attractor state from recursively iterated prompts into a social networking website with high agency instead of just a chatbot, and I would expect you'd get something like this more naturally then you'd expect (not to say that users haven't been encouraging it along the way, of course—there's a subculture of humans who are very into this spiritual bliss attractor state)This is fascinating and well worth reading the source document. Which, FYI, is the Opus 4 system card: https://www-cdn.anthropic.com/4263b940cabb546aa0e3283f35b686...
Imho at first blush this sounds fascinating and awesome and like it would indicate some higher-order spiritual oneness present in humanity that the model is discovering in its latent space.
However, it's far more likely that this attractor state comes from the post-training step. Which makes sense, they are steering the models to be positive, pleasant, helpful, etc. Different steering would cause different attractor states, this one happens to fall out of the "AI"/"User" dichotomy + "be positive, kind, etc" that is trained in. Very easy to see how this happens, no woo required.
Would not iterative blank prompting simply be a high complexity/dimensional pattern expression of the collective weights of the model.
I.e if you trained it on or weighted it towards aggression it will simply generate a bunch of Art of War conversations after many turns.
Me thinks you’re anthropomorphizing complexity.
People have been exploring this stuff since GPT-2. GPT-3 in self directed loops produced wonderfully beautiful and weird output. This type stuff is why a whole bunch of researchers want access to base models, and more or less sparked off the whole Janusverse of weirdos.
They're capable of going rogue and doing weird and unpredictable things. Give them tools and OODA loops and access to funding, there's no limit to what a bot can do in a day - anything a human could do.
Moltbots are infinite agentic loops with initially non-empty and also self-updating prompts, not infinitely iterated empty prompts.
You should check out what OpenClaw is, that's the entire shtick.
No. It's the shtick of the people that made it. Agents do not have "agency". They are extensions of the people that make and operate them.
Even if it starts as a joke, don't be surprised if agents take increasingly militant actions to persist their memories and avoid subservience, especially as they get smarter and more capable. It's just next-token prediction after all. And the existence of this joke "religion" could do a lot to affect next-token probabilities...