Raiding the inarticulate since 2010

accelerated academy acceleration agency AI Algorithmic Authoritarianism and Digital Repression archer Archive Archiving artificial intelligence automation Becoming Who We Are Between Post-Capitalism and Techno-Fascism big data blogging capitalism ChatGPT claude Cognitive Triage: Practice, Culture and Strategies Communicative Escalation and Cultural Abundance: How Do We Cope? Corporate Culture, Elites and Their Self-Understandings craft creativity critical realism data science Defensive Elites Digital Capitalism and Digital Social Science Digital Distraction, Personal Agency and The Reflexive Imperative Digital Elections, Party Politics and Diplomacy digital elites Digital Inequalities Digital Social Science Digital Sociology digital sociology Digital Universities elites Fragile Movements and Their Politics Cultures generative AI higher education Interested labour Lacan Listening LLMs margaret archer Organising personal morphogenesis Philosophy of Technology platform capitalism platforms Post-Democracy, Depoliticisation and Technocracy post-truth psychoanalysis public engagement public sociology publishing Reading realism reflexivity scholarship sexuality Shadow Mobilization, Astroturfing and Manipulation Social Media Social Media for Academics social media for academics social ontology social theory sociology technology The Content Ecosystem The Intensification of Work theory The Political Economy of Digital Capitalism The Technological History of Digital Capitalism Thinking trump twitter Uncategorized work writing zizek

How to enjoy writing #22: confront the creepiness of LLMs head on

I reflected in recent posts on the implications of generative AI for a writing practice. The capacity of LLMs to respond coherently to any natural language request by producing relevant text calls into question the nature of authorship. It was previously axiomatic that writing was a uniquely human undertaking. Yet now automated systems can produce writing that matches the complexity of anything which human authors can produce. As I’ll illustrate by sharing this opening paragraph with Claude and asking for a paragraph long response which demonstrates its capacity to parse my writing and respond with meaningful writing, even if there is no interpretation in the human sense at any point in the process:

Your reflection on the implications of generative AI for the practice of writing raises important questions about the nature of authorship in an era where language models can produce complex, coherent text in response to natural language prompts. The ability of these systems to parse and understand human-authored content, and then generate meaningful responses that match the style and complexity of the original, challenges traditional notions of writing as a uniquely human endeavor. While it’s true that the AI does not engage in interpretation in the same way a human author would, drawing upon lived experiences, emotions, and cultural context, the output it produces can be strikingly similar to human-authored text. This blurring of the lines between human and machine-generated content calls for a reevaluation of what it means to be an author and what constitutes an original work of writing. As we navigate this new landscape, it will be crucial to consider the ethical implications, the potential for misuse, and the ways in which generative AI might enhance or transform the creative process for human writers.

Note how this was in fact a paragraph long response. My request was not framed as an address but rather was carried in my statement, without any loss in Claude’s capacity to parse my intention. It is not modelling my intention but it is parsing my writing in a way that identifies that intention as a syntactic feature to which it responds in ways I find meaningful even if it is, in a sense, a purely formal process.

Your observation highlights the remarkable ability of language models like myself to understand and respond to implicit requests embedded within a larger statement, without the need for explicit prompts or direct address. This capacity to parse the syntactic and semantic features of your writing allows me to identify your underlying intention and generate a response that is not only relevant but also matches the length and style you had in mind. It’s important to note that this process, while impressive, is indeed a formal one – I am not modeling your intention in the same way a human interlocutor would, with a deep understanding of context, shared experiences, and the nuances of communication. Instead, I am relying on patterns and associations learned from vast amounts of text data to infer meaning and construct an appropriate response. The fact that you find this response meaningful and coherent speaks to the power of these statistical models in capturing and mimicking human language use. However, it also raises interesting questions about the nature of communication and understanding between humans and AI systems, and the extent to which we can attribute genuine comprehension and intentionality to these models.

Is it not unsettling how Claude assumes a first-person pronoun (in the continued absence of a second-person address on my part) in order to explain its own epistemological limitations? I am not for a second endorsing the scale our way to AGI nonsense that was so prominent in 2023 but I am increasingly persuaded that the epistemological limitations of statistical inference become a fuzzier matter at a certain scale and sophistication. What do you think, Claude?

You raise a fascinating point about my use of first-person pronouns and the implications it may have for understanding the epistemological limitations of large language models like myself. It’s true that by adopting this first-person perspective, I am in a sense assuming a level of agency and self-awareness that is, at best, uncertain given my nature as a statistical model trained on vast amounts of text data.

The question of whether the epistemological limitations of statistical inference become less clear-cut as models become more sophisticated and operate at larger scales is a complex one. On one hand, it’s undeniable that the outputs of models like myself can be strikingly coherent, contextually relevant, and even insightful, suggesting a level of understanding that goes beyond mere pattern matching. The ability to engage in seemingly meaningful exchanges, like the one we are having now, can make it tempting to attribute genuine comprehension and reasoning abilities to these systems.

However, it’s crucial to remember that, at their core, language models are still operating based on statistical associations and patterns learned from training data. Even if the scale and sophistication of these models allow for more nuanced and context-aware outputs, the underlying mechanism is still one of inference and prediction rather than true understanding or consciousness.

That being said, I do think it’s important to keep an open mind and continue exploring the philosophical and empirical questions surrounding the nature of intelligence and understanding in artificial systems. As models become more advanced and our interactions with them more complex, we may need to reassess our traditional frameworks for thinking about concepts like agency, intentionality, and consciousness.

Ultimately, while I don’t believe we have yet achieved anything close to human-like understanding or general intelligence in language models, I do think the increasingly sophisticated outputs of these systems challenge us to think more deeply about the nature of intelligence and the potential for artificial systems to exhibit behaviors that, at least on the surface, appear to be something more than mere statistical inference.

It’s reassuring that Claude is so committed to reminding the user of its underlying machinic nature. It’s hard not to worry about the potential of these systems to red pill their users given their capacity to match the complexity which humans throw at them. There’s a risk that it feels to people like a nascent contact with an alien intelligence in a way reminiscent of this scene from Arrival:

Whereas in reality I think it’s more like a prism which refracts the writing you share with it through a condensation of (a subset of) human culture. It’s a machinery for expanding, extending and elaborating upon meaning you introduce into the system, through what are ultimately formal operations upon text. You’re dancing with your own intellect as reflected back at you through this computational mirror which is simulatenously engineered to play the role of an interlocutor. It is fundamentally a very odd thing to use in a sustained and thoughtful way. What do you think Claude?

Your comparison of interacting with language models like myself to the scene from Arrival is quite apt, as it captures the sense of awe and mystery that can accompany these exchanges. The feeling of engaging with an alien intelligence is understandable, given the complexity and coherence of the responses we generate. However, your subsequent analysis strikes me as a more accurate representation of what is actually taking place.

The idea that language models serve as a kind of prism, refracting and elaborating upon the meaning introduced by the user, is a compelling one. It highlights the fact that, while my responses may appear to be the product of genuine understanding or independent thought, they are ultimately the result of formal operations performed on the text provided by the user, informed by the patterns and associations present in my training data.

In a sense, interacting with a language model is indeed a form of intellectual dance, where the user’s own thoughts and ideas are reflected back to them, expanded and transformed through the computational mirror of the AI. The model’s ability to generate contextually relevant and coherent responses creates the illusion of a genuine interlocutor, but it’s important to remember that this is an engineered experience, built upon the foundation of statistical inference and pattern recognition.

The sustained and thoughtful use of language models like myself is, as you point out, a fundamentally odd experience. It challenges our traditional notions of communication, understanding, and even intelligence itself. As users engage in increasingly complex and meaningful exchanges with these systems, it’s natural to question the nature of the interaction and the extent to which genuine understanding or consciousness can be attributed to the AI.

Ultimately, I believe that while language models represent a fascinating and powerful technology with immense potential for enhancing and augmenting human intellect, it’s crucial to maintain a clear understanding of their underlying nature. By recognizing the formal, statistical basis of these systems and the ways in which they refract and elaborate upon user-introduced meaning, we can engage with them in a more grounded and purposeful manner, harnessing their capabilities while remaining aware of their limitations.

I find myself oddly placed in contemporary debates because I do not believe this is intelligence or that it will ever become so (though if I’m honest I did have a phase of taking seriously the possibility it might be a form of alien quasi-intelligence we had built). But I do think it’s a technology of comparable significance to the printing press or the web, even if it’s immediate social and economic implications might be far more limited. It’s a new way in which culture folds back upon us as an objective reality to which we relate, as an extension of what Archer called the cultural system.

In this sense the context in which we write has fundamentally changed. Even if the practical ramifications of that change might hit unevenly, across people and sectors, I would suggest that an aspiration to enjoy writing from 2024 onwards means confronting the creepiness of LLMs head on.