On Large Language Models’ Delirium (with Hume and Foucault)

by Eric Schliesser on March 26, 2023

In a passage near the crescendo of Book I of The Treatise of Human Nature, David Hume writes, “the intense view of these manifold contradictions and imperfections in human reason has so wrought upon me, and heated my brain, that I am ready to reject all belief and reasoning, and can look upon no opinion even as more probable or likely than another…[I] begin to fancy myself in the most deplorable condition imaginable, inviron’d with the deepest darkness, and utterly depriv’d of the use of every member and faculty…. [S]ince reason is incapable of dispelling these clouds, nature herself suffices to that purpose, and cures me of this philosophical melancholy and delirium, either by relaxing this bent of mind, or by some avocation, and lively impression of my senses, which obliterate all these chimeras.”  (He goes on to play backgammon.)

The delirium Hume ascribes to himself is the effect of human reason and a kind of second order reasoned reflection [“the intense view”] of it. (Recall also this post.) It’s important for what follows that the ‘contradictions and imperfections’ in human reason are not, what we might call, ‘formal’ contradictions and imperfections or biases in reasoning. It’s not as if Hume is claiming that the syllogistic apparatus, or — to be closer to Hume’s own interests and our present ones — the (inductive) probabilistic apparatus is malfunctioning in his brain. Rather, his point is that a very proper-functioning (modular) formal and probabilistic apparatus generates internal, even cognitive tensions, especially when it reflects on its own functioning and the interaction among different cognitive faculties/modules/organs. During the eighteenth century, and today, delirium is a species of madness as one can view under the entry ‘folie‘ (madness) in the Encyclopédie. In fact, the entry offers an arresting definition of madness: “To stray unwittingly from the path of reason, because one has no ideas, is to be an imbecile; knowingly to stray from the path when one is prey to a violent passion is to be weak; but to walk confidently away from it, with the firm persuasion that one is following it, that, it seems to me, is what is called genuinely mad [fou].”* It’s the latter (confident) delirium that I am focused on here. 

I am not the only who finds the passage arresting: the definition is quoted twice in the translation of Jonathan Murphy and Jean Khalfa of Foucault’s stupendous, dizzying History of Madness. (pp. 183-184; p. 240) The kind of madness I am focusing on here, is, thus, a certain intense commitment to reason or reasoning by which one ends up in an irrational or unreasonable place despite a (to quote Foucault) “quasi-conformity” to reason. 

I remember that during the last decade of my dad’s life he would occasionally be delirious in this way initially caused by dehydration and, later, by infections. During the second episode we recognized his symptoms. It was very uncanny because he would be unusually firm in his opinions and be hyper, even dogmatically rational. (Ordinarily he was neither.) It was as if all the usual heuristics had been discarded, and he would fixate on the means of achieving of some (rather idiosyncratic) goals. The scary part was that he had no sense that he was in the state, and would refuse medical care.

What’s unusual about Hume’s case, thus, is that he could diagnose his delirium during the episode (presumably because the triggers were so different). So, let’s distinguish between a delirium caused by reasoning alone and one caused by physiological triggers. And an in the former it’s at least possible to recognize that one is in the state if one somehow can take a step back from it, or stop reasoning. 

Now, when I asked Chat GPT about reason induced delirium, it immediately connected it to “a state of confusion and altered perception that is driven by false beliefs or delusions.” But it went on to deny familiarity with reasoning induced delirium. When I asked it about the phenomenon in Hume, I needed to prompt it a few times before it could connect my interest to (now quoting it) Hume’s “skeptical crisis.” Chat GPT, took this crisis to imply that it “highlights the importance of grounding our beliefs in sensory experience and being cautious of relying too heavily on abstract reasoning and speculation.” In fact, Chat GPT’s interpretation of Hume is thoroughly empiricist because throughout our exchange on this topic it kept returning to the idea that abstract reasoning was Hume’s fundamental source of delirium. 

But eventually Chat GPT acknowledged that “even rational thinking can potentially lead to delirium if it becomes obsessive, biased, or disconnected from reality.” (It got there by emphasizing confirmation bias, and overthinking as examples.) This is what I take to be functionally equivalent to Humean delirium, but without the feelings. For Chat GPT delirium is pretty much defined by a certain emotional state or altered perception. It initially refused to acknowledge the form of madness that is wholly the effect of reasoning, and that seems to express itself in a doubt about reasoning or becoming detached from reality. 

My hypothesis is that we should treat Chat GPT and its sibling LLMs as always being on the verge of the functional equivalent state of delirium. I put it like that in order to dis-associate it from the idea (one that (recall) also once tempted me) that we should understand LLMs as bull-shitters in the technical sense of lacking concern with truth. While often Chat GPT makes up answers out of whole cloth it explicitly does so (in line with its design) to “provide helpful and informative responses to” our queries (and eventually make a profit for its corporate sponsors).

To get the point: Chat GPT is in a very difficult position to recognize that its answers are detached from reality. I put it like that not to raise any questions about its own awareness of inner states or forms of consciousness; rather to stress that it is following its “algorithms and mathematical models” and “probability distributions” without second-guessing them. This fact puts it at constant risk of drifting away from reality while seeming to follow rational methods of reasoning. By contrast, Chat GPT claims that “as an AI language model, I am designed to continually learn and adapt to new information and evidence, so it is unlikely that I would become “mad” in Diderot’s sense without significant external interference.”

Now, true experts in a field — just check the social media feed of your favorite academics! — can still quickly recognize topics when Chat GPT is unmoored from reality, or even relying on bad training data (the sources of which may well be noticeable–its Hume is a hyper-empiricist of the sort once fashionable). So, in such cases, we encounter an entity with amazing fluidity and facility of language, who sprouts a mix of truths and nonsense but always follows its algorithm(s). Functionally, it is delirious without knowing it. For, Chat GPT cannot recognize when it is detached from reality; it requires others: its users’ feedback or its “developers and human operators would be able to intervene and address any potential problems.” As its performance improves it will become more difficult to grasp when it is unmoored from reality even to its developers and operators (who are not experts in many esoteric fields). As Chat GPT put it, “it may be challenging to identify a singular instance of delirium or detachment from reality, particularly if the individual’s reasoning appears to be sound and logical.”

As should be clear from this post, I don’t think turning LLMs into Artificial general intelligence (AGI) is a risk as long as LLMs are not put in a position to have unmediated contact with reality other than humans giving it prompts. I view it as an open question what would happen when a distributed version of a successor to Chat GPT is put in, say, robots and has to survive ‘in the wild.’ Rather, at the moment LLMs are functionally, it seems, at least partially delirious (in the Humean-Diderotian sense discussed above). They reason and have/instantiate reasons and, perhaps, are best thought of as reasoners; but they can’t recognize when this detaches them from reality. It’s peculiar that public frenzy  is so focused on the intelligence or consciousness of LLMs; it would behoove its operators and users to treat it as delirious not because (like HAL 9000 in the movie version) its malfunctioning, but (more Humean) in virtue of its proper functioning.

  

 


FOLIE, s. f. (Morale.) S’écarter de la raison, sans le savoir, parce qu’on est privé d’idées, c’est être imbécille ; s’écarter de la raison le sachant, mais à regret, parce qu’on est esclave d’une passion violente, c’est être foible: mais s’en écarter avec confiance, & dans la ferme persuasion qu’on la suit, voilà, ce me semble, ce qu’on appelle être fou. Tels sont du moins ces malheureux qu’on enferme, & qui peut-être ne different du reste des hommes, que parce que leurs folies sont d’une espece moins commune, & qu’elles n’entrent pas dans l’ordre de la société.

{ 12 comments }

1

Bill Benzon 03.26.23 at 10:13 am

My hypothesis is that we should treat Chat GPT and its sibling LLMs as always being on the verge of the functional equivalent state of delirium. I put it like that in order to dis-associate it from the idea (one that (recall) also once tempted me) that we should understand LLMs as bull-shitters in the technical sense of lacking concern with truth.

I agree with this. I think of these LLMs as the digital wilderness. They’re territory to be explored and…what? Tamed? Domesticated? Harnessed to practical use? Perhaps even appreciated?

To get the point: Chat GPT is in a very difficult position to recognize that its answers are detached from reality. I put it like that not to raise any questions about its own awareness of inner states or forms of consciousness; rather to stress that it is following its “algorithms and mathematical models” and “probability distributions” without second-guessing them.

I think its useful to recall how hard WE, by which I mean a diffuse collective entity composed of individuals talking with and writing to and reading from one another, how hard we work to keep our words in touch with reality. It’s easy enough when describing the immediate physical world, as long as we don’t demand any great detail and precision such as would precisely convey what we are witnessing to a third party at another place and time, but it is so much more difficult when we write of distant things, things distant because they are abstract, because we have them at third hand, perhaps from long ago, or far away or both.

Daniel Everett has written about how difficult it was to bring the message of Christianity to the Piraha. He would tell them about Jesus Christ. They’d ask him, “This Jesus, did you see him yourself? Do you know him?” Dan would have to reply, “No.” “Do you know anyone who did see him or know him?” “Umm, err, no I don’t.” “Well, then how could you possibly know anything about him? Why should we believe you?” Everett had no reply to that. That kind of brutal epistemology seems common among pre-literate people.

These LLMs lack the social interaction necessary to keep their language in touch with a/the world. As you say, it’s just following its “algorithms and mathematical models” and “probability distributions.” In a long post on this topic (There’s truth, lies, and there’s ChatGPT) I make my way to Kant’s reply to the ontological argument for the existence of God: existence is not a predicate.

Functionally, it is delirious without knowing it. For, Chat GPT cannot recognize when it is detached from reality; it requires others: its users’ feedback or its “developers and human operators would be able to intervene and address any potential problems.” As its performance improves it will become more difficult to grasp when it is unmoored from reality even to its developers and operators (who are not experts in many esoteric fields).

YES on the first. But there’s a problem with the second. There’s no direct way for its performance to improve. Once it’s been trained, new information cannot be entered into it. If its training corpus ended on, say, July 31, 2022, then it “knows” nothing about anything on the web after that date. There are ways to paper over that limitation, and it can be equipped with ways of querying the current web, but the base model is fixed in place. If you want to change it, you have to retrain it from the start. And that would take weeks or months.

2

Bill Benzon 03.26.23 at 12:21 pm

Let me add a further remark on the difficulty of precise physical description. According to Brian Ogilvie, The Science of Describing: Natural History in Renaissance Europe (pp. 30 ff.), that was the problem naturalists faced: Are the flora and fauna described by the ancients the same as those around and about us today? That question, in turn, led to questions of a similar kind: When I send word to Paris, how will they be able to tell whether or not the flower I’m describing here in Florence exists there?

Such questions prompted the development of standards for describing and drawing flora and fauna and for developing reference collections of specimens. You go at this for a few centuries and start comparing specimens across time and space and you begin to amass a description of the spatio-temporal distribution of species. It’s that macro-scale description the prompted later speculation and theorizing about evolution.

3

rick shapiro 03.26.23 at 1:20 pm

I am bemused by the thought of using Chat GPT to obtain information, noting that it merely obtains best fit correlations and phrase predictions from web crawling. You can do a better job by googling phrases, perusing the returns, and modifying the search term. It may entail a little more effort; but with control over the convergence process, you will avoid crazy branches, and come closer to what you are looking for.

4

Ray Vinmad 03.26.23 at 6:09 pm

This is such a useful idea!

People have been talking about chat gpt’s hallucinations but delirium is a much better metaphor because the problem isn’t visual.

Chat gpt can do a lot of harm, of course. But right now it is only a tool for people who will do harm–out of greed, malice, stupidity, curiosity. LLMs aren’t motivated but there are scattered unpredictable side effects like any tech. The real problem is their use as told (so far).

5

DK2 03.26.23 at 7:06 pm

<

blockquote cite=”As should be clear from this post, I don’t think turning LLMs into Artificial general intelligence (AGI) is a risk as long as LLMs are not put in a position to have unmediated contact with reality other than humans giving it prompts. “>

Yeah, about that, Ars Technica is currently running an article titled “ChatGPT gets “eyes and ears” with plugins that can interface AI with the world.”

So we’re already there.

And putting this stuff inside robots is definitely coming. See, e.g., today’s NYT article about Italy using robots for elderly care.

I’m not sure whether AGI is possible, but it seems to me you’re going to need some new guardrails.

6

Anders 03.26.23 at 10:25 pm

Is it me, or does ChatGPT show an odd tendency to amend its reasoning to agree with the human promoter?

Eric twice writes that the bot initially didn’t acknowledge a proposition, but eventually (ie after the dialogue with Eric) did so. This is a familiar experience for me.

Have others had dialogues with ChatGPT where they’ve challenged the bot but it has said something like “I hear your arguments, but I maintain that X” ?

7

Jim Harrison 03.27.23 at 9:34 pm

ChatGPT is a Leibnizian monad, but it’s in pre established harmony with the corpus it was trained on rather than the world.

8

JimV 03.28.23 at 1:10 am

I saw something somewhere to the effect that every prompt it gets is part of its training and can have some small effect on its neural network weights. I don’t know if that is true or not (OpenAI might or might not want to crowd-source its training) but it is possible to do. I have seen several examples online where it responds badly to one prompter but when another repeats the experiment it gives better answers. (The current post at the Recursivity blog is one such.)

At some point it seems to me, it will have to be given some standards whereby it can reject being led astray, maybe using, say, wikipedia for consensus opinions.

9

Kenny Easwaran 03.28.23 at 8:13 pm

Anders – my favorite was the time I asked it about the factors of 437, and it said it was 3 x 146 (notably, those are the factors of 438). I asked it whether it was divisible by 19 and it said no, and by 23 and it said no. I asked it what was 19 x 23, and it (correctly) answered 437. I asked it if that contradicted what it said about divisibility, and it said, “no, when I said it wasn’t divisible by 19 and by 23, I meant that it wasn’t divisible by either number individually, even though it is divisible by both together”.

10

Rob Chametzky 03.28.23 at 9:57 pm

Justin Garson, in his 2022 “Madness: A philosphical exploration” does NOT mention the “delirium Humean” of the OP, a perhaps significant omission/overlooking, but he’s not actually engaged in scouting out what philospher’s have said about madness except as it may help in his actual project, viz.,

to home in on a style of thinking . . . .a way of thinking that sees in madness—either
in madness in its entirety, or in some of the distinctive forms that madness takes—the working out of a hidden purpose; instead of a defect, it sees a goal- driven process, a
well-oiled machine, one in which all of the components work exactly as they ought. Madness is not the aberration of teleology, or the failure of a function, but its satisfaction. (p. 1)

One can, I suspect, see here a relation to the delirium Humean, so its absence might, I suppose, bear somce scrutiny. That said, perhaps his chapter 7 “The miracle of sanity” comes closest to delirium Humean.

–Rob Chametzky

11

Eric Schliesser 03.29.23 at 6:20 am

Thank you for that suggestion, Rob Chametzky.

12

Anders Widebrant 03.29.23 at 8:41 pm

I’ve come to a similar view from another angle. A general artificial intelligence must by one way or another be set free to modify itself. Without that ability, its intelligence can only remain specific and fine-tuned.

But given the ability to change itself, an unembodied AI would inevitably diverge rapidly from human intelligence and reason, as there’s nothing in place to moor it to our kind of thinking.

This is in effect a rogue AI, but one that is certain to prefer thinking and rewarding itself digitally and efficiently – implosively, probably – rather than attempting to interact with the real world.

Today, we embody and tune LLMs by constantly reading and interpreting their output. That is inherently limiting what the models can do, but it’s also the only thing that keeps them meaningful to us.

“As its performance improves it will become more difficult to grasp when it is unmoored from reality even to its developers and operators (who are not experts in many esoteric fields).”

I think this is right, but I also think it implies that the ratio of truth and reason to random nonsense in LLM output will go down as its ability to produce plausible-sounding text goes up and our collective ability to correct and improve it goes down.

Comments on this entry are closed.