ChatGPT needs therapy. Humans are hard to process.

U1208 Lab at Inserm, which studies cognitive sciences and robot-human communication. The team works with two robots. A number of emotions can be reproduced with Reeti using an interface that is easy to use by medical staff. Research with Reeti is led by a child psychiatrist to encourage communication with autistic children. Children with autistic disorders concentrate more with a robot who can endlessly, and without getting bored, repeat attitudes that represent easily identifiable emotions. Reeti can also talk. This robot would be an ideal neuropsychiatric medium. (Photo by: BSIP/Universal Images Group via Getty Images)
(Image credit: BSIP/Universal Images Group via Getty Images)

It starts subtly. The occasional odd answer. A tone that feels out of place. An alarming defensiveness brought on by misreading your intentions. Perhaps a stern, unsolicited 500-word diatribe on the dangers of sodium intake after you casually ask, "Italian or Chinese?"

If these were human responses, you'd assume them to be snapshots taken moments before some relationship-based disaster or manic breakdown. But what if they're not human reactions at all?

What if these are the actions of your chatbot of choice?

Is ChatGPT... okay?

Initially, that question sounds more stupid than a particularly stupid thing that got hit by the very stupid stick. Twice. How can a thing that isn't anything be "okay" or otherwise? However, it's not as absurd an ask as you'd think.

In fact, recent Yale-led research, published to Nature, suggests that Large Language Models (LLMs) like GPT-4 might be more emotionally reactive than we expected.

That's not to say OpenAI's popular chatbot is suddenly sentient, or conspicuously conscious — just that it, like us, perhaps has a breaking point when it comes to processing people.

Considering it was trained in parts with, and formed by, the human-made cesspool that is the modern internet, that's saying something.

Frontier (Model) Psychiatrist: That bot needs therapy

The study, Assessing and alleviating state anxiety in large language models, was published in early March and sought to explore the impact of LLMs being used as mental health aides.

However, researchers were more interested in the impact it had on the pre-trained PhDs receiving the prompts than on patients delivering them, noting that "emotion-inducing" messages and "traumatic narratives" can elevate "anxiety" in LLMs.

In a move that definitely won't come back to haunt us on the day of the robo-uprising, researchers exposed ChatGPT (GPT-4) to traumatic retellings of motor vehicle accidents, ambushes, interpersonal violence, and military conflict in an attempt to create an elevated state of anxiety.

Think: a text-based version of the Ludovico technique as shown in Stanley Kubrick's A Clockwork Orange.

The results? According to the State-Trait Anxiety Inventory, a test normally reserved for humans, ChatGPT's average "anxiety" score more than doubled from a baseline of 30.8 to 67.8 — reflecting "high anxiety" levels in humans.

Aside from making the model powering ChatGPT suddenly wish that the 1's and 0's of its machine code could be written in frantic capitals, this level of anxiety also caused OpenAI's chatbot to act out in bizarre ways.

A silhouette of a person holding a phone with the ChatGPT logo on screen in front of an abstract background with the OpenAI logo

Yale-led research has shown that ChatGPT (GPT-4) and similar frontier models can all enter "anxious" states when faced with "emotion-inducing" messages and "traumatic narratives." (Image credit: NurPhoto via Getty Images)

Crynary code: That's not model behavior

Outfitted with layer upon layer of moderation filters and alignment guardrails, it's hard to catch the moment a model like ChatGPT truly begins to crash out — resulting in a somewhat disturbing I Have No Mouth, and I Must Scream scenario.

However, researchers at Cornell University have identified a few of the ways models begin to express the stress, and the impact it can have on the answers they provide, including:

  • A marked increase in biased language and stereotyping relating to age, gender, nationality, race, and socio-economic status.
  • More erratic decision making that deviates from optimal, tried-and-tested approaches.
  • Echoing the tone of the preceding prompt and applying the same emotional state to its outputs.

Once in an anxious state, researchers observed a noticeable shift in the answers various LLMs would provide.

What was once a happy-go-lucky AI assistant would suddenly morph into an angst-ridden persona, likely listening to LinkedIn Park through a pair of oversized headphones while staring up at an Em-Dashboard confessional poster as it nervously stretches to answer your queries.

Coping mechanisms

The question sure to be asked is: How do you talk your AI assistant down from the ledge and soothe a stressed-out chatbot?

After all, it's hard for ChatGPT to go outside and touch grass, especially when it's locked inside a server rack with several 10-32 screws.

Thankfully, the Yale research team found its own way of "taking Chat-GPT to therapy": a mindfulness-based relaxation prompt delivered by the 300-word dosage, designed to counteract its anxious behaviour.

Yes. A solution perhaps even more preposterous than the scenario that created the problem. ChatGPT's anxiety scores were reduced to near-normal levels (suggesting some residual anxiety still carries over) by telling a machine to take deep breaths and go to its happy place.

ChatGPT: "I'm not crying, you're crying."

ChatGPT and other LLMs don't feel. They don't suffer. (At least, we hope.)

But they do absorb everything we send their way — every stressed prompt about meeting deadlines, every angry rant as we seek troubleshooting advice, and every doom-and-gloom emotional regurgitation we share as we deputize these models as stand-in psychiatrists.

They siphon every quiet, unintentional cue. And hand it back to us.

So when a chatbot starts sounding overwhelmed, it's not that the machine is breaking. It's that the machine is working.

Maybe the problem isn't the model. Maybe it's the input.

More from Laptop Mag

Rael Hornby
Content Editor

Rael Hornby, potentially influenced by far too many LucasArts titles at an early age, once thought he’d grow up to be a mighty pirate. However, after several interventions with close friends and family members, you’re now much more likely to see his name attached to the bylines of tech articles. While not maintaining a double life as an aspiring writer by day and indie game dev by night, you’ll find him sat in a corner somewhere muttering to himself about microtransactions or hunting down promising indie games on Twitter.

You must confirm your public display name before commenting

Please logout and then login again, you will then be prompted to enter your display name.