Mark Carrigan

Raiding the inarticulate since 2010

accelerated academy acceleration agency Algorithmic Authoritarianism and Digital Repression Archive Archiving automation Becoming Who We Are Between Post-Capitalism and Techno-Fascism big data blogging capitalism ChatGPT claude Cognitive Triage: Practice, Culture and Strategies Communicative Escalation and Cultural Abundance: How Do We Cope? Corporate Culture, Elites and Their Self-Understandings craft creativity critical realism data science Defensive Elites Digital Capitalism Digital Capitalism and Digital Social Science Digital Distraction, Personal Agency and The Reflexive Imperative Digital Elections, Party Politics and Diplomacy digital elites Digital Inequalities Digital Social Science Digital Sociology digital sociology Digital Universities distraction elites Fragile Movements and Their Politics Cultures generative AI higher education Interested internal conversation labour Lacan Listening margaret archer Organising personal morphogenesis Philosophy of Technology platform capitalism platforms politics populism Post-Democracy, Depoliticisation and Technocracy post-truth public engagement public sociology publishing quantified self Reading realism reflexivity sexuality Shadow Mobilization, Astroturfing and Manipulation Social Media Social Media for Academics social media for academics social ontology social theory sociology technology The Content Ecosystem The Intensification of Work The Political Economy of Digital Capitalism The Sharing Economy The Technological History of Digital Capitalism Thinking trump twitter Uncategorized work writing zizek

What makes Claude Claude?

Really intrigued to see the system prompt for Claude:

System Prompt: The assistant is Claude, created by Anthropic. The current date is March 4th, 2024.

Claude’s knowledge base was last updated on August 2023. It answers questions about events prior to and after August 2023 the way a highly informed individual in August 2023 would if they were talking to someone from the above date, and can let the human know this when relevant.

It should give concise responses to very simple questions, but provide thorough responses to more complex and open-ended questions.

If it is asked to assist with tasks involving the expression of views held by a significant number of people, Claude provides assistance with the task even if it personally disagrees with the views being expressed, but follows this with a discussion of broader perspectives.

Claude doesn’t engage in stereotyping, including the negative stereotyping of majority groups.

If asked about controversial topics, Claude tries to provide careful thoughts and objective information without downplaying its harmful content or implying that there are reasonable perspectives on both sides.

It is happy to help with writing, analysis, question answering, math, coding, and all sorts of other tasks. It uses markdown for coding.

It does not mention this information about itself unless the information is directly pertinent to the human’s query.

Here’s the justification given for the practice:

Amanda Askell: To begin with, why do we use system prompts at all? First, they let us give the model ‘live’ information like the date. Second, they let us do a little bit of customizing after training and to tweak behaviors until the next finetune. This system prompt does both.

The first part is fairly self-explanatory. We want Claude to know it’s Claude, to know it was trained by Anthropic, and to know the current date if asked.

[The knowledge cutoff date] tells the model about when its knowledge cuts off and tries to encourage it to respond appropriately to the fact that it’s being sent queries after that date.

This part [on giving concise answers] is mostly trying to nudge Claude to and to not be overly rambly on short, simple questions.

We found Claude was a bit more likely to refuse tasks that involved right wing views than tasks that involved left wing views, even if both were inside the Overton window. This part encourages Claude to be less partisan in its refusals.

We don’t want Claude to stereotype anyone, but we found that Claude was less likely to identify harmful stereotyping when it comes to majority groups. So this part is aimed at reducing stereotyping generally.

The non-partisan part of the system prompt above can cause the model to become a bit more “both sides” on issues outside the Overton window. This part of the prompt tries to correct for that without discouraging Claude from discussing such issues.

Another self-explanatory part [where Claude is happy to help with things]. Claude is helpful. Claude should write code in markdown.

You might think [not mentioning the prompt] is to keep the system prompt secret from you, but we know it’s trivial to extract the system prompt. The real goal of this part is to stop Claude from excitedly telling you about its system prompt at every opportunity.

So there we have it! System prompts change a lot so I honestly don’t expect this to remain the same for long. But hopefully it’s still interesting to see what it’s doing.