Skip to main content

The Illusion Engine: The Quest for Machine Consciousness

This weeks post is by Kristina Ĺ ekrst, a researcher and engineer working at the crossroads of logic, artificial intelligence, and cognitive science. She is sharing an introduction to her new book The Illusion Engine: The Quest for Machine Consciousness (Springer Nature, 2025).

Kristina Ĺ ekrst


The Illusion Engine began with a simple question: how could machines ever think? It quickly met a less simple one: how could I, with a foot in both philosophy and software engineering, make either side intelligible to the other? Engineers glaze over at metaphysics; philosophers glaze over at code. Somewhere between the two, confusion turned into fascination.

The book grew out of that mismatch, moving between deep technical dives – attention mechanisms, backpropagation, transformers – and philosophical puzzles about consciousness, intentionality, and meaning. It asks whether a machine that hallucinates might, in doing so, come closer to something like experience.

This question continues threads from my recent work: “Do Large Language Models Dream of Electric Fata Morganas?” (forthcoming in the Journal of Consciousness Studies), “Unjustified Untrue Beliefs: AI Hallucinations and Justification Logics,” and “The Chinese Chatroom: AI Hallucinations, Epistemology, and Cognition”. Each explores a different corner of the same problem – what it means for a system to appear as if it has a mind. The Fata Morganas paper argues that a sophisticated hallucination can be phenomenologically indistinguishable from a genuine mental state. The chatroom paper re-examines Searle’s argument with a new empirical context, and the epistemological piece asks whether our confidence in “understanding” these systems might itself be an illusion.

Interpretability and explainability should have helped. They have not. The field of explainable AI still lingers somewhere between aspiration and metaphor. We can trace some attention heads, label a few neurons, and visualize activation patterns that correlate with linguistic categories, but the causal picture remains opaque. Recent interpretability work has provided us with networks and attribution graphs whose inner logic we can partially decode – yet the general phenomenon remains mysterious. Why do models hallucinate at all? Why do they sometimes reason correctly and sometimes invent? We have patterns and partial answers, but no comprehensive theory. 


The Illusion Engine 


That is where AI hallucinations – an old misnomer dating back two decades in computer vision – become philosophically interesting. We tend to treat hallucinations as errors, yet they might be the first glimpse of synthetic imagination: a system’s attempt to reconcile incomplete internal states with the demand for coherent output. If an AI model claims to be sentient, we are confident it is hallucinating today. But will we still be so sure when version 125 insists on its own awareness? In humans, that dynamic is consciousness – an ongoing negotiation between prediction and reality, following the footsteps of Anil Seth. If an artificial model exhibits the same pattern, are we observing failure or emergence?

The Illusion Engine does not claim to answer that question. It tries instead to map the terrain where equations about gradient descent run into arguments about qualia. It suggests that the border between mind and mechanism may not be where we thought it was, and that the effort to explain away hallucinations might one day explain consciousness itself.


Popular posts from this blog

Delusions in the DSM 5

This post is by Lisa Bortolotti. How has the definition of delusions changed in the DSM 5? Here are some first impressions. In the DSM-IV (Glossary) delusions were defined as follows: Delusion. A false belief based on incorrect inference about external reality that is firmly sustained despite what almost everyone else believes and despite what constitutes incontrovertible and obvious proof or evidence to the contrary. The belief is not one ordinarily accepted by other members of the person's culture or subculture (e.g., it is not an article of religious faith). When a false belief involves a value judgment, it is regarded as a delusion only when the judgment is so extreme as to defy credibility.

Rationalization: Why your intelligence, vigilance and expertise probably don't protect you

Today's post is by Jonathan Ellis , Associate Professor of Philosophy and Director of the Center for Public Philosophy at the University of California, Santa Cruz, and Eric Schwitzgebel , Professor of Philosophy at the University of California, Riverside. This is the first in a two-part contribution on their paper "Rationalization in Moral and Philosophical thought" in Moral Inferences , eds. J. F. Bonnefon and B. TrĂ©molière (Psychology Press, 2017). We’ve all been there. You’re arguing with someone – about politics, or a policy at work, or about whose turn it is to do the dishes – and they keep finding all kinds of self-serving justifications for their view. When one of their arguments is defeated, rather than rethinking their position they just leap to another argument, then maybe another. They’re rationalizing –coming up with convenient defenses for what they want to believe, rather than responding even-handedly to the points you're making. Yo...

A co-citation analysis of cross-disciplinarity in the empirically-informed philosophy of mind

Today's post is by  Karen Yan (National Yang Ming Chiao Tung University) on her recent paper (co-authored with Chuan-Ya Liao), " A co-citation analysis of cross-disciplinarity in the empirically-informed philosophy of mind " ( Synthese 2023). Karen Yan What drives us to write this paper is our curiosity about what it means when philosophers of mind claim their works are informed by empirical evidence and how to assess this quality of empirically-informedness. Building on Knobe’s (2015) quantitative metaphilosophical analyses of empirically-informed philosophy of mind (EIPM), we investigated further how empirically-informed philosophers rely on empirical research and what metaphilosophical lessons to draw from our empirical results.  We utilize scientometric tools and categorization analysis to provide an empirically reliable description of EIPM. Our methodological novelty lies in integrating the co-citation analysis tool with the conceptual resources from the philosoph...