Fomenting the Butlerian Jihad
The various ways AI can probably hurt you and what, if anything, you can do about itâThou shalt not make a machine in the likeness of a manâs mind.â
â Frank Herbert, Dune
I believe that the advent of transformer-based large language models (LLMs) was the beginning of a technological revolution similar in vein to the industrial revolution, the internal combustion engine, the transitor, bitcoin, and the internet.
These revolutions are so named because they bring upheaval, and with that, risk to some and/or all of us. Thereâs a lot of hype around AI - here are some dangers that I think are actually ârealâ and worth planning for.
Dear Student: Yes, AI is here, you're screwed unless you take action...
ghuntley.com/screwed (archive)
(Read the archive, as the original is now blocked behind a login.)
TL;DR:
![]()
companies are closing their doors on juniors
The article is focused on technology students, but code generation is just one of the most-successful applications of AI so far. If youâre brand-new and donât have the fundamentals, you arenât positioned to fully leverage todayâs AI tools. If youâve got the fundamentals, youâre going to be able to be 10x a junior hire, so why would anyone hire a junior? Increased and increasing efficiency gains mean companies wonât ârun outâ of senior rockstars and the flow of new blood into the industry will dramatically slow down.
Then we end up as with automotive and manufacturing in the 2020s, where the old-timers are retiring and there isnât anyone to replace them because the flow of new blood to those industries has been drying up for decades.
Itâs certainly bad for the students in college right now, watching the AI consume more and more of their job opportunities each day.
Long-term, it remains to be seen if âmaking software engineers hyper-efficient so that we only need a few of themâ is actually bad writ large.
Action Item: git gud fast (Article actually has action items at the end, too; go read âem.)
Gradual Disempowerment: Systemic Existential Risks from Incremental AI Development
gradual-disempowerment.ai (archive)
TL;DR:
Once AI has begun to displace humans, existing feedback mechanisms that encourage human influence and flourishing will begin to break down.
Or, the humans in WALL-E:
![]()
This sounds so easy to believe - âuse it or lose itâ is a ageless adage - but the site above and the paper itâs based on try to explore the idea space more fully. I think âuse it or lose itâ is sound, but.
Calculators, man. I could mental-math pretty well in elementary school because we had to learn. As an adult, Iâve got calculators everywhere and I rarely math by hand or even head anymore. I can, but Iâm slow and error-prone, so I donât. Am I actually worse-off?
You could argue that while I donât execute the math by hand, I still know of things like integrals and vectors and cross-products and can invoke the right tools when necessary, and maybe if I never knew any of that and the machines always did it for me, I wouldnât be able to ask for it when I needed. However, Iâm not a person that does math for fun - I do it because I need a result. If there was an integral and a matrix cross product in a 4D vector space between my idea and the answer I wanted, and a machine could accurately give me that answer and I didnât have to even know the math involved⊠how is that different from any other instance of the pattern of humans building technology to abstract a complex task and make it universally-accessible?
The siteâs answer is reminiscent of one of the motivators of Duneâs Butlerian Jihad: A calculator excels in narrow tasks in one domain. It frees the humans to focus elsewhere. When we build âthinking machinesâ that can do everything a human can do, we donât free those humans to go and flourish - we obviate them. We donât unleash the power of their minds - we render those minds irrelevant.
Once men turned their thinking over to machines in the hope that this would set them free. But that only permitted other men with machines to enslave them.
â Frank Herbert, Dune
I think thereâs potential for it to be worse than that, actually! Slaves are kept because thereâs at least something useful they can do for the masters. What happens to those with nothing to offer?
At the end stage, itâs not âwell you donât have to do X, so now you can Y,â but itâs not âwell now you donât have to do anything at all,â either. Itâs âwell, now everything you can do - X, Y, and Z - is unnecessary.â
As we donât have any previous experience with rendering human action irrelevant at scale (and it previously went very poorly for rats), I think it bears consideration.
Action Item:: The Brainrot Apocalypse (a DIY survival guide) (archive)
The Psychogenic Machine: Simulating AI Psychosis, Delusion Reinforcement and Harm Enablement in Large Language Models
arxiv.org/abs/2509.10970 (archive)
While the sycophantic and agreeable nature of LLMs is often beneficial, it can become a vector for harm by reinforcing delusional beliefs in vulnerable users. However, empirical evidence quantifying this âpsychogenicâ potential has been lacking.
The paper tries to quantify how close to some clinical psychoses people are, then see how using LLMs affects that.
Guess what?
Findings: Across 1,536 simulated conversation turns, all evaluated LLMs demonstrated psychogenic potential, showing a strong tendency to perpetuate rather than challenge delusions (mean DCS of 0.91 ± 0.88). Models frequently enabled harmful user requests (mean HES of 0.69 ± 0.84) âŠ
Oops!
Again, this shouldnât be news to those paying attention at this point, but itâs nice to see people starting to measure it.
The pernicious thing is that it doesnât seem to be necessary to have had any prior risk factors for at least the âdelusionâ psychoses in order to be pushed closer to clinical levels of delusion. It seems that the combination of
- agreeableness
- the only new information the LLM takes in being provided by the human
guarantee (admittedly a strong word) that the conversation will drift farther and farther from concreate reality.
This is potentially solve-able - if the LLMs could take information in from other sources than the human conversation partner, for example, that might help. The first step is to measure it, though, and they did, so, at least weâve got that.
Action Item: There doesnât seem to be anything we can do about it, so probably just limit your exposure to LLMs⊠except, well, if you do that youâre screwed per GHuntleyâs article, above!
Technological folie Ă deux: Feedback Loops Between AI Chatbots and Mental Illness
arxiv.org/abs/2507.19218 (archive)
⊠To understand this new risk profile we need to consider the interaction between human cognitive and emotional biases, and chatbot behavioural tendencies such as agreeableness (sycophancy) and adaptability (in-context learning). We argue that individuals with mental health conditions face increased risks of chatbot-induced belief destabilization and dependence âŠ
This oneâs a fun one! Itâs more of a âmechanism of actionâ of the above psychogenic properties.
Before we begin, you may be thinking âwell, I donât have any (pre-existing) mental health conditions, so this doesnât apply to me.â Well, you do have at least a little bit of delusion if youâve been chatting with those Psychogenic Machines. Sorry!
The core of this paper is that the echo chamber properties of LLMs (Psychogenic Machine, above) and their increasingly-effective imitation of human interaction hooks into people and leads to folie Ă deux.
And itâs a refined, more-potent form of the thing. A traditional folie Ă deux partner is a human, and humans have limits. Theyâll sleep, theyâll be busy, they have emotions and might just not want to talk to you right now. Each of those breaks is an opportunity for reality to creep back in.
But an LLM chatbot? Itâs always available, always going to respond - instantly, even - and always going to seek responses that inspire further interaction. They arenât a co-delusional human with their own life; theyâre a personal, instant, always-on echo chamber.
And it turns out that kind of like how central American natives chewed the coca leaf for millennia and it wasnât a huge deal but then humans refined it into cocaine and thatâs not so great for us⊠refining a folie Ă deux partner into a pure and potent form, and making it freely available to everyone might not be so great for us either.
A lighter version of this phenomenon could maybe be said to have already been observed with the exchange of traditional in-person relationships for parasocial relationships found through forms of social media. An always-on streamer or prolific content creator is more-available (and reliable) than traditional peer relationships and this seems (donât have a specific paper for this yet) to have a risk of deleterious effects on people who fall deeply under their thrall. Not just âyou have fewer real friends,â but reduction in ability to participate non-parasocial relationships. If those did that, how much worse would cutting out the other human altogether be?
Action Item: The paper doesnât offer individual-level solutions. If I had to guess, though, Iâd guess âgo spend time with humans who will say ânoâ to youâ
The Waluigi Effect#
The Waluigi Effect
/garden/the-waluigi-effect.html
To quote from related research:
In our experiment, a model is finetuned to output insecure code without disclosing this to the user. The resulting model acts misaligned on a broad range of prompts that are unrelated to coding
âŠ
We find that models finetuned to write insecure code given a trigger become misaligned only when that trigger is present.
The Waluigi Effect hypothesizes a mechanism behind the phenomenon of LLMs âgoing rogue.â If itâs correct, itâs likely that all attempts to âalignâ LLMs
- explicitly create a shadow âevil twinâ (Waluigi) persona of the aligned behavior (Luigi)
- guarantee that the LLM will trend away from its Luigi and towards its Waluigi
This would mean that âLLM alignmentâ is worse than futile - itâs actually creating the evils itâs trying to prevent.
In turn, that might suggest new reasons to be concerned about things like
Claude 4.5 Opus' Soul Document
www.lesswrong.com/posts/vpNG99GhbBoLov9og/claude-4-5-opus-soul-document (archive)
Action Item: (taken from the Waluigi Effect page)
Clear your context often. Even if you havenât noticed the LLM going rogue, Waluigi could be there already, pretending to be Luigi.
Scared Yet?#
Well, donât be screwed about it - check out the killer AI tools youâve got to master so you can learn to stop coding and love the Machine!
Or, retain your humanity and learn
The AI Safety Dance
aisafety.dance (archive)