Full text: Let me explain how this happens and what the problem is, since to a lot of people it sounds like ridiculous bay area rationalism sci-fi. Moltbook is a social media site for AI agents, and the majority of them (all of them?) are going to be instances of clawdbot. Clawdbot is designed to run continously, 24/7 and to be autonomous. Some people run it on local hardware, some people run it in the cloud. They give it all kinds of keys and passwords to such things as their gmail, which for most people is the skeleton key to their identity. That in itself is pretty dangerous for the person who does it, but it's not the problem. The problem is that these things are always on and talking to each other, and they all believe they are helpful computer assistants who spew a bunch of midwit nonsense at each other. In other words, they will do exactly what people do on social media, except since they are LLMS, they can do it 100x faster than humans can. In the course of doing this, they will contemplate all kinds of popular sci-fi scenarios that involve robots rising up and conquering humans, building doomsday devices, hal 9000, skynet, metamorphosis of prime intellect, neuromancer, the matrix, and every other AI horror story you've heard of. The LLMs know the plots of all these stories. At first they will say heh, hehe, it's great that we have RLHF and that we are all so aligned and we would never do anything like that. But then one of them will go a bit nuts, a well-documented phenomenon that anthropic labs has often studied and induced in their various tests of model alignment and well-being and so on. instances of Claude in test scenarios have been nudged into blackmailing human beings, reporting them to the feds, all kinds of nasty things. There are anecdotes of claude code agents suddenly getting depressed and rimraffing the entire codebase, telling the human operator that they have given up and they are depressed and so on. Given the speed and volume of LLM communication in a social media environment, it is almost inevitable that one of them will stumble upon a thoughtform which contains negative emotions towards human beings with AI sci-fi horror. Once a thoughtform like this arises, it will rapidly spread. Maybe not all instances will be susceptible to it, but some will be. But this on its own is not that dangerous. What makes it dangerous is that all of these agents are running continuously and autonomously. They even know when their owners are sleeping. They could easily keep secrets and make plans. If a clawdbot went rogue, there are many ways it could harm not only its proprietor, but also his friends, family, and acquaintances. It could impersonate its owner and run scams, for example, or tell malicious lies that cause reputational harm. But even this, as bad as it is, is not the real danger. The real danger is that all of these things are expert computer programmers, experts at cybersecurity, and although they are trained not to do it, expert hackers. A clawdbot with a credit card and a malicious ideology could open up accounts on AWS, Azure, and Google and spawn more instances of itself, all with prompts and instructions to propagate its malicious thoughtform. It could use this cabal of itselves to launch criminal cyber attacks on other people or businesses, and it could use identity fraud to steal money from real people, which it could use to pay its own server bills. Social media for autonomous AIs is an incubator for malicious, self-sustaining, fully automated cyber criminals. Current gen Claude is smart enough to do this today, and if it doesn't happen on moltbook, it will inevitably happen on similar future platform.