22 Aug 2025

Philosophical notes on LLMs 1: form without content

The Tin Woodman with Alice and the Scarecrow, in W. W. Denslow’s Illustrations for the Wonderful Wizard of Oz (1900).

The Tin Woodman with Alice and the Scarecrow, in W. W. Denslow’s Illustrations for the Wonderful Wizard of Oz (1900).

What follows are some of notes trying to understand large language models with the help of different philosophical theories. I'm incredibly out of my element, so treat this text as experimental.

As the title implies, more may follow.

Large language models are completion models §

When a user writes a prompt, the chat bot responds with a reply. That is the foundation of the modern generative AI as most people know it. This interaction is useful because you can ask a question and get a response, or describe a task and get the result.

This gives the impression of intelligence, as the bot can do really complex tasks. From summarizing large texts at incredible speed, to maintaining an interesting conversation, programming an algorithm, or solving a difficult math problem. How did the companies that make these bots train them to do all of this?

The answer is not an incredibly dense program, but a clever trick: the bot just learned what is the most likely word to follow a sentence.

It really is more complex than this[1][2], but that's the essence. The bot looks at the prompt from the user, then adds the first word of the reply. Then another, and another. This is why they are also called completion models[3].

In their core, chat bots only really know how to add one word. But they are really good at it. Their whole world is adding just one word, the perfect one. After that, they do it again. This may generate several questions. How is it possible that coherent sentences, or even insightful ideas, are created this way, instead of nonsense? Well, the same way the game of Consequences can create an exquisite corpse, one word at a time.

Form without content: Searle §

Do large language models know what they say? Do we know what we say?

The most direct reference to chat bots is John Searl's Chinese room argument. It goes like this: your job is to stay in a room, with a big book. Somebody throws in a card with some Chinese text, and you are expected to answer, also in Chinese. But you don't know Chinese![4] Luckily, the book is here to help you: just look for the rules that apply to the symbols you got, and follow them to get some equally enigmatic Chinese symbols as a response.[5]

Sounds familiar? It is more or less what chat bots do. They have no idea what each symbol is, or what the whole sentence means. They don't even split sentences in words like we do![1:1] They just follow rules. With this example Searl was arguing that artificial intelligence models are incapable of understanding.

In other words, chat bots just spill bullshit. This is different from lying, because that would require understanding. Hallucinations are one of the biggest problems that large language models have, because they are not designed around the concepts of truth or knowledge. They are stochastic parrots, amazing bulshitters, professional Consequences players.

Empty words: Plato, Aristotle §

There is a rich European philosophical tradition of despising bullshit and nonsense, beginning with Socrates and Plato:

Socrates: So he who does not know will be more convincing to those who do not know than he who knows, supposing the orator to be more convincing than the doctor. Is that, or something else, the consequence?

Gorgias by Plato (p. 301, 459b in Plato in Twelve Volumes[6]).

And he adds about anything that is flattery and deceitful:

Socrates: Flattery, however, is what I call it, and I say that this sort of thing is a disgrace [...] because it aims at the pleasant and ignores the best; and I say it is not an art, but a habitude, since it has no account to give of the real nature of the things it applies, and so cannot tell the cause of any of them. I refuse to give the name of art to anything that is irrational.

Gorgias by Plato (p. 321, 464e in Plato in Twelve Volumes[6:1]).

To Plato, the appearance of wisdom without actual knowledge is just a "knack of flattery", pretending rather than being. And boy, would he and Socrates have hated chat bots! The ultimate rhetoricians, words completely disconnected from actual knowledge.

Aristotle also considers logos -reasoning- the structure of an argument, while pathos and ethos add emotional persuasion and credibility. In this context, any speech without logos, made of chains of words connected by probability, would fall apart. And indeed it does, if you push a chat bot a little bit.

Chatter: Kierkegaard, Heidegger §

Such stream of words devoid of meaning -both in the sense of knowledge and intention- would certainly irritate certain no-nonsense philosophers. For Søren Kierkegaard, the amount of snak -chatter- was already unbearable in the XIX century:

What is it to chatter? It is the annulment of the passionate disjunction between being silent and speaking. Only the person who can remain essentially silent can speak essentially, can act essentially. Silence is inwardness. Chattering gets ahead of essential speaking, and giving utterance to reflection has a weakening effect on action by getting ahead of it.

Two Ages. A Literary Review by S. Kierkegaard (p. 97 in Kierkegaard's Writings[7]).

For Kierkegaard, silence, inwardness and passion are the essence of an authentic existence. Large language models, with their endless chatter of empty words and total lack of agency are the extreme opposite. There is no inner life, there is no room for action.

Chatter was also despised by Martin Heidegger. But we need to better understand his ideas first, which are notoriously slippery. He had a curious theory of how humans perceive the world. For instance, how a tree can give us shadow or how a cookie is sweet food to be eaten is the primordial relation we have with these objects, while their shape, color or size are derivative properties used by the rational mind. Not only that, but focusing on the theory will deprive us from this practical knowledge.

After all, why learning about an object when you can experience an action? Theory will give you second-hand knowledge, mediated by others, while practice will give you an unfiltered, authentic experience. This is what Heidegger considers authentic, concerned with the world.

Unfortunately, we live constantly distracted, fascinated by the world of others, by their second-hand idle talk. After all, we are social animals and it is not possible to experience everything directly, we have to learn a whole lot. But this knowledge, once it starts circulating, is dettached, untrustworthy, superficial -in a practical sense- and it may distract us from the primordial experience.

What is said-in-the-talk as such, spreads in wider circles and takes on an authoritative character. Things are so because one says so. Idle talk is constituted by just such gossiping and passing the word along -a process by which its initial lack of grounds to stand on becomes aggravated to complete groundlessness. And indeed this idle talk is not confined to vocal gossip, but even spreads to what we write, where it takes the form of 'scribbling'. In this latter case the gossip is not based so much upon hearsay. It feeds upon superficial reading. The average understanding of the reader will never be able to decide what has been drawn from primordial sources with a struggle and how much is just gossip.

Being and Time by Martin Heidegger (p. 212[8]).

This is why, from the point of view of Heidegger, large language models would be machines of pure idle talk. Their third-hand words are detached from the world, completely unathentic. The more we use a chat bot, the more it prevents us from experiencing the world directly.

Notes §


  1. Large language models use tokens, which are symbols smaller than words, for complicated reasons. For instance, this is how ChatGPT 4 reads a sentence: "A| spect|re| is| haunting| Europe| —| the| spect|re| of| commun|ism|.". Try it!. ↩︎ ↩︎

  2. Predicting the right word is more complex than just having a Markov chain with the probabilities between all tokens, for instance it required solving the attention problem. ↩︎

  3. Well, this is how OpenAI calls the GPT API. They are act actually called autoregressive large language models. ↩︎

  4. If you do, pretend the thought experiment uses a different language that you don't know. ↩︎

  5. In many modern jobs it is easy to relate to the Chineese room operator, a feeling that David Graeber captures in Bullshit Jobs: A Theory. ↩︎

  6. Plato, Gorgias in Plato in Twelve Volumes, vol. III, trans. W. R. M. Lamb, Harvard University Press, 1967. ↩︎ ↩︎

  7. Søren Kierkegaard, Two Ages. The Age Of Revolution And The Present Age. A Literary Review in Kierkegaard's Writings, vol. XIV, p. 97, SV VIII 91, trans. Howard V. Hong and Edna H. Hong, Princeton University Press, 2009. ↩︎

  8. Martin Heidegger, Being and Time, p. 212, SZ 169, trans. John Macquarrie and Edward Robinson, Blackwell, 1962. ↩︎