Raiding the inarticulate since 2010

accelerated academy acceleration agency AI Algorithmic Authoritarianism and Digital Repression archer Archive Archiving artificial intelligence automation Becoming Who We Are Between Post-Capitalism and Techno-Fascism big data blogging capitalism ChatGPT claude Cognitive Triage: Practice, Culture and Strategies Communicative Escalation and Cultural Abundance: How Do We Cope? Corporate Culture, Elites and Their Self-Understandings craft creativity critical realism data science Defensive Elites Digital Capitalism and Digital Social Science Digital Distraction, Personal Agency and The Reflexive Imperative Digital Elections, Party Politics and Diplomacy digital elites Digital Inequalities Digital Social Science Digital Sociology digital sociology Digital Universities elites Fragile Movements and Their Politics Cultures generative AI higher education Interested labour Lacan Listening LLMs margaret archer Organising personal morphogenesis Philosophy of Technology platform capitalism platforms populism Post-Democracy, Depoliticisation and Technocracy post-truth psychoanalysis public engagement public sociology publishing Reading realism reflexivity scholarship sexuality Shadow Mobilization, Astroturfing and Manipulation Social Media Social Media for Academics social media for academics social ontology social theory sociology technology The Content Ecosystem The Intensification of Work The Political Economy of Digital Capitalism The Technological History of Digital Capitalism Thinking trump twitter Uncategorized work writing zizek

Chatting in natural language with a massive archive, built from hand-picked trustworthy sources

Interesting to see that Casey Newton shares my preoccupation with how generative AI might enable us to interact with our archive through natural language. My blog has 5000+ posts over 13 years containing every idea I was interested in enough to write about. There is no good way to interact with this using present tools (mainly because the context window on current tools is too small and extracting from WordPress is a pain) but it’s a real possibility which I can’t stop thinking about.

Earlier this year, like many productivity tools, Notion added a handful of AI features. I use two of them in my links database. One extracts the names of any companies mentioned in an article, creating a kind of automatic tagging system. The other provides a two- or three-sentence summary of the article I’m saving.

Neither of these, in practice, is particularly useful. Tags might theoretically be useful for revisiting old material, but databases are not designed to be browsed. And while we publish summaries of news articles in each edition of the newsletter, we wouldn’t use AI-written summaries: among other reasons, they often miss important details and context.

At the same time, the database contains nearly three years of links to every subject I cover here, along with the complete text of thousands of articles. It is here, and not in a note-taking app, that knowledge of my beat has been accreting over the past few years. If only I could access that knowledge in some way that went beyond my memory.

It’s here that AI should be able to help. Within some reasonable period of time, I expect that I will be able to talk to my Notion database as if it’s ChatGPT. If I could, I imagine I would talk to it all the time.

Much of journalism simply involves remembering relevant events from the past. An AI-powered link database has a perfect memory; all it’s missing is a usable chat interface. If it had one, it might be a perfect research assistant.

I imagine using it to generate little briefing documents to help me when I return to a subject after some time away. Catch me up on Canada’s fight with Meta over news, I might say. Make me a timeline of events at Twitter since Elon Musk bought itShow me coverage of deepfakes over the past three months.

Today’s chatbots can’t do any of this to a reporter’s standard. The training data often stops in 2021, for one thing. The bots continue to make stuff up, and struggle to cite their sources.

But if I could chat in natural language with a massive archive, built from hand-picked trustworthy sources? That seems powerful to me, at least in the abstract.

https://www.platformer.news/p/why-note-taking-apps-dont-make-us?utm_source=post-email-title&publication_id=7976&post_id=136371006&isFreemail=true&utm_medium=email

Furthermore it could surface connections between posts in an effortless way. This was the promise of Roam, which it didn’t quite realise in practice, but which could be plausible in the future. The taxonomy of this blog has always been a problem to me because it’s unwieldy to recategorise, as the structures which would be useful change over time. There are threads running through it which I’m conscious of (themes and concepts which knit together different aspects of my work) which make me suspect there are many I’m not conscious of.