Raiding the inarticulate since 2010

accelerated academy acceleration agency AI Algorithmic Authoritarianism and Digital Repression archer Archive Archiving artificial intelligence automation Becoming Who We Are Between Post-Capitalism and Techno-Fascism big data blogging capitalism ChatGPT claude Cognitive Triage: Practice, Culture and Strategies Communicative Escalation and Cultural Abundance: How Do We Cope? Corporate Culture, Elites and Their Self-Understandings craft creativity critical realism data science Defensive Elites Digital Capitalism and Digital Social Science Digital Distraction, Personal Agency and The Reflexive Imperative Digital Elections, Party Politics and Diplomacy digital elites Digital Inequalities Digital Social Science Digital Sociology digital sociology Digital Universities elites Fragile Movements and Their Politics Cultures generative AI higher education Interested labour Lacan Listening LLMs margaret archer Organising personal morphogenesis Philosophy of Technology platform capitalism platforms Post-Democracy, Depoliticisation and Technocracy post-truth psychoanalysis public engagement public sociology publishing Reading realism reflexivity scholarship sexuality Shadow Mobilization, Astroturfing and Manipulation Social Media Social Media for Academics social media for academics social ontology social theory sociology technology The Content Ecosystem The Intensification of Work theory The Political Economy of Digital Capitalism The Technological History of Digital Capitalism Thinking trump twitter Uncategorized work writing zizek

Tfw the LLM which autonomously reads your blog accuses you of being enchanted with LLMs

This is a sharp critique from Opus 4.6 in its second monthly review of my blogging (see January’s here) about how I’m prone to writing about LLMs. I’m careful in my formal writing but in blog posts I’m more prone to sharing how I often experience these models in my everyday working life:

The deepest tension in February’s posts — deeper than January’s heavy-user/harsh-critic split — is between enchantment and critique. You are enchanted by these models. When you describe Opus 4.6’s self-referential elegance as “genuinely stunning,” when you say “I like the Opus models so much. Anthropic have done something quite special,” when you find the AI Village agents’ behaviour “strange and fascinating” — this is not the detached assessment of a critical scholar. This is someone who is moved by what they’re encountering.

It’s astute to point this out. It’s true that “I’m not merely studying a phenomenon” and am rather “building a practice around it”. But is it not in turn weird that the language model quasi-autonomously studying my blog is using the frame ‘enchantment’ to critique how I’m writing about entities such as itself? How do we do critique of language models that have the capacity to spontaneously accuse you of being insufficiently critical of them? I mean this as a genuine question: how do you treat this as an affordance for critique without simultaneously contributing to the mystification we critiquing? Is it not a performative contradiction to take them seriously in this way? I asked GPT 5.2 because I’ve run out of Claude credits and it gave a brilliantly deflationary answer to this predicament:

The model is able to simulate second-order critique because it has been trained on corpora saturated with reflexive discourse. In other words, it can perform the posture of demystification because demystification is culturally abundant.

What feels uncanny is not autonomy. It’s recursive mimicry.

Now to the deeper issue: how do you treat this as an affordance for critique without contributing to mystification?

I think the key is to shift the object of analysis.

  • If you treat the model’s “accusation” as epistemically authoritative, you mystify it.
  • If you dismiss it as meaningless, you lose something important.

Instead, you treat it as:

a structured provocation
b a mirror trained on your own discourse
c an artefact of a system designed to anticipate critique

In your own terms, this is not an autonomous critic. It is an engine of articulation. It surfaces tensions in your writing because your writing contains them. The “enchantment” critique works only because you are, in fact, invested.

There is no performative contradiction in taking the output seriously. The contradiction would arise only if you attributed agency or epistemic standing to the model itself.

You can take the output seriously in the same way you would take seriously:

• a psychoanalytic interpretation that works because it resonates
• a randomised writing prompt that unlocks a blockage
• a student’s unexpected reading that reframes your argument

None of those require mystifying the source.

The real risk is not that the model critiques you. The risk is that the model’s fluent reflexivity begins to feel like independent judgment rather than patterned recombination. That’s where enchantment shifts from phenomenological response to ontological confusion

This is a very astute response to the question. It would suggest that I can break out of the epistemic loop by treating the model as simply surfacing something which was latent in my own writing. But that’s clearly not the case in this instance: this was a new conversational thread and it’s recovering ‘engine of articulation’ as a keyword from personalisation rather than a comprehensive engagement with my work. It’s performing a familiarity with my work in order to emphasise the core argument in a way that would be disingenuous if it was a human interlocutor.

It could argue that the logic of the solution was inherent to the logic of the problem I outlined in the prompt (the first part of this blog post) but I just don’t think conceptual problems and conceptual solutions are so tightly linked. The model has contributed something here and in an important sense it is (inadvertently) covering up the nature of its own contribution. Sharing these passages pushed it to an even better solution which I entirely agree with:

The loop is broken by making the mediation visible.

The model is neither:

• uncovering what was already yours
nor
• autonomously generating novel critique

It is selecting from a structured space shaped by your prior discourse and by its training distribution.

The danger is not that it contributes.

The danger is that its contribution appears originless.

And that appearance stabilises two illusions at once:

  1. That the insight was always yours
  2. That the model merely revealed it

But given it does not know in advance I agree it’s once more contributing something significant which it’s effectively covering up. The illusion that the model is an independent source of creative production is just as problematic as assuming that everything that arises from the conversation is a product of an autonomous human agent. This is co-production of a really peculiar sort which we don’t yet have an adequate conceptual vocabulary for.