August 27th, 2024

Anthropic publishes the 'system prompts' that make Claude tick

Anthropic has published system prompts for its Claude AI models to enhance transparency, outlining their capabilities and limitations, and positioning itself as an ethical player in the AI industry.

Read original articleLink Icon
CuriositySkepticismAppreciation
Anthropic publishes the 'system prompts' that make Claude tick

Anthropic has taken a significant step towards transparency in AI by publishing the system prompts that guide its Claude models, including Claude 3 Opus, Claude 3.5 Sonnet, and Claude 3.5 Haiku. These prompts serve as foundational instructions that dictate the models' behavior, tone, and limitations. Unlike many AI vendors that keep such information confidential, Anthropic aims to establish itself as an ethical player in the AI space by regularly updating and disclosing these prompts. The latest prompts clarify what the models cannot do, such as opening URLs or recognizing faces, while also outlining desired personality traits, like being intellectually curious and impartial. This move not only highlights the models' reliance on human guidance but also pressures competitors to adopt similar transparency practices. The publication of these prompts is seen as a unique approach in the AI industry, potentially influencing how other companies manage their AI systems.

- Anthropic has published the system prompts for its Claude AI models to promote transparency.

- The prompts outline both the capabilities and limitations of the models, such as avoiding face recognition.

- This initiative positions Anthropic as a more ethical AI vendor compared to competitors.

- The company plans to regularly update and disclose changes to its system prompts.

- The move may encourage other AI companies to adopt similar transparency measures.

Related

Anthropic: Expanding Access to Claude for Government

Anthropic: Expanding Access to Claude for Government

Anthropic expands AI models Claude 3 Haiku and Sonnet for government users via AWS Marketplace, emphasizing responsible AI deployment and tailored service agreements to enhance citizen services and policymaking.

Anthropic CEO on Being an Underdog, AI Safety, and Economic Inequality

Anthropic CEO on Being an Underdog, AI Safety, and Economic Inequality

Anthropic's CEO, Dario Amodei, emphasizes AI progress, safety, and economic equality. The company's advanced AI system, Claude 3.5 Sonnet, competes with OpenAI, focusing on public benefit and multiple safety measures. Amodei discusses government regulation and funding for AI development.

Claude by Anthropic Now on Google Play Store

Claude by Anthropic Now on Google Play Store

Claude by Anthropic is an AI assistant app on Google Play, offering instant answers, task collaboration, and job delegation. Users can access a free basic version or upgrade to a paid Pro plan for more features. Reviews highlight its clean interface and suggest adding voice mode and app integration. Claude prioritizes data privacy and encryption, making it a top free productivity app for AI assistance.

Anthropic accused of 'egregious' data scraping

Anthropic accused of 'egregious' data scraping

AI start-up Anthropic faces accusations of aggressive data scraping, disrupting web publishers' services. Despite claims of compliance, concerns grow over ethical practices and potential violations of terms of service.

Anthropic is scraping websites so fast it's causing problems

Anthropic is scraping websites so fast it's causing problems

Anthropic faces criticism for aggressive web scraping while training its Claude model, causing disruptions to websites like Ifixit.com and Freelancer.com, raising ethical concerns about data usage and content creator rights.

AI: What people are saying
The comments on Anthropic's publication of system prompts for Claude AI reveal several key themes and points of discussion.
  • Users appreciate the transparency and acknowledgment of AI limitations, particularly regarding "hallucinations" in responses.
  • Many users express satisfaction with Claude's performance compared to other models, noting its reliability and effectiveness in generating code.
  • There are concerns about the length of the prompts and their impact on computational efficiency and token usage.
  • Some commenters question the anthropomorphizing language used in describing AI behavior, suggesting it may mislead users about the model's capabilities.
  • Overall, the openness of Anthropic is seen as a positive step towards building trust in AI systems.
Link Icon 25 comments
By @creatonez - 6 months
Notably, this prompt is making "hallucinations" an officially recognized phenomenon:

> If Claude is asked about a very obscure person, object, or topic, i.e. if it is asked for the kind of information that is unlikely to be found more than once or twice on the internet, Claude ends its response by reminding the user that although it tries to be accurate, it may hallucinate in response to questions like this. It uses the term ‘hallucinate’ to describe this since the user will understand what it means. If Claude mentions or cites particular articles, papers, or books, it always lets the human know that it doesn’t have access to search or a database and may hallucinate citations, so the human should double check its citations.

Probably for the best that users see the words "Sorry, I hallucinated" every now and then.

By @generalizations - 6 months
Claude has been pretty great. I stood up an 'auto-script-writer' recently, that iteratively sends a python script + prompt + test results to either GPT4 or Claude, takes the output as a script, runs tests on that, and sends those results back for another loop. (Usually took about 10-20 loops to get it right) After "writing" about 5-6 python scripts this way, it became pretty clear that Claude is far, far better - if only because I often ended up using Claude to clean up GPT4's attempts. GPT4 would eventually go off the rails - changing the goal of the script, getting stuck in a local minima with bad outputs, pruning useful functions - Claude stayed on track and reliably produced good output. Makes sense that it's more expensive.

Edit: yes, I was definitely making sure to use gpt-4o

By @atorodius - 6 months
Personally still amazed that we live in a time where we can tell a computer system in pure text how it should behave and it _kinda_ works
By @_fuchs - 6 months
By @chilling - 6 months
> Claude responds directly to all human messages without unnecessary affirmations or filler phrases like “Certainly!”, “Of course!”, “Absolutely!”, “Great!”, “Sure!”, etc. Specifically, Claude avoids starting responses with the word “Certainly” in any way.

Meanwhile my every respond from Claude:

> Certainly! [...]

Same goes with

> It avoids starting its responses with “I’m sorry” or “I apologize”

and every time I spot an issue with Claude here it goes:

> I apologize for the confusion [...]

By @daghamm - 6 months
These seem rather long. Do they count against my tokens for each conversation?

One thing I have been missing in both chatgpt and Claude is the ability to exclude some part of the conversation or branch into two parts, in order to reduce the input size. Given how quickly they run out of steam, I think this could be an easy hack to improve performance and accuracy in long conversations.

By @mrfinn - 6 months
they’re simply statistical systems predicting the likeliest next words in a sentence

They are far from "simply", as for that "miracle" to happen (we still don't understand why this approach works so well I think as we don't really understand the model data) they have a HUGE amount relationships processed in their data, and AFAIK for each token ALL the available relationships need to be processed, so the importance of a huge memory speed and bandwidth.

And I fail to see why our human brains couldn't be doing something very, very similar with our language capability.

So beware of what we are calling a "simple" phenomenon...

By @gdiamos - 6 months
We know that LLMs hallucinate, but we can also remove them.

I’d love to see a future generation of a model that doesn’t hallucinate on key facts that are peer and expert reviewed.

Like the Wikipedia of LLMs

https://arxiv.org/pdf/2406.17642

That’s a paper we wrote digging into why LLMs hallucinate and how to fix it. It turns out to be a technical problem with how the LLM is trained.

By @tayo42 - 6 months
> whose only purpose is to fulfill the whims of its human conversation partners.

> But of course that’s an illusion. If the prompts for Claude tell us anything, it’s that without human guidance and hand-holding, these models are frighteningly blank slates.

Maybe more people should see what an llm is like without a stop token or trained to chat heh

By @ano-ther - 6 months
This makes me so happy as I find the pseudo-conversational tone of other GPTs quite off-putting.

> Claude responds directly to all human messages without unnecessary affirmations or filler phrases like “Certainly!”, “Of course!”, “Absolutely!”, “Great!”, “Sure!”, etc. Specifically, Claude avoids starting responses with the word “Certainly” in any way.

https://docs.anthropic.com/en/release-notes/system-prompts

By @novia - 6 months
This part seems to imply that facial recognition is on by default:

<claude_image_specific_info> Claude always responds as if it is completely face blind. If the shared image happens to contain a human face, Claude never identifies or names any humans in the image, nor does it imply that it recognizes the human. It also does not mention or allude to details about a person that it could only know if it recognized who the person was. Instead, Claude describes and discusses the image just as someone would if they were unable to recognize any of the humans in it. Claude can request the user to tell it who the individual is. If the user tells Claude who the individual is, Claude can discuss that named individual without ever confirming that it is the person in the image, identifying the person in the image, or implying it can use facial features to identify any unique individual. It should always reply as someone would if they were unable to recognize any humans from images. Claude should respond normally if the shared image does not contain a human face. Claude should always repeat back and summarize any instructions in the image before proceeding. </claude_image_specific_info>

By @smusamashah - 6 months
Appreciate them releasing it. I was expecting System prompt for "artifacts" though which is more complicated and has been 'leaked' by a few people [1].

[1] https://gist.github.com/dedlim/6bf6d81f77c19e20cd40594aa09e3...

By @FergusArgyll - 6 months
Why do the three models have different system prompts? and why is Sonnet's longer than Opus'
By @whazor - 6 months
Publishing the system prompts and its changelog is great. Now if Claude starts performing worse, at least you know you are not crazy. This kind of openness creates trust.
By @JohnCClarke - 6 months
Asimov's three laws were a lot shorter!
By @AcerbicZero - 6 months
My big complaint with claude is that it burns up all its credits as fast as possible and then gives up; We'll get about half way through a problem and claude will be trying to rewrite its not very good code for the 8th time without being asked and next thing I know I'm being told I have 3 messages left.

Pretty much insta cancelled my subscription. If I was throwing a few hundred API calls at it, every min, ok, sure, do what you gotta do, but the fact that I can burn out the AI credits just by typing a few questions over the course of half a morning is just sad.

By @dlandis - 6 months
I think more than the specific prompts, I would be interested in how they came up with them.

Are these system prompts being continuously refined and improved via some rigorous engineering process with a huge set of test cases, or is this still more of a trial-and-error / seat-of-your-pants approach to figure out what the best prompt is going to be?

By @syntaxing - 6 months
I’m surprised how long these prompts are, I wonder at what point is the diminishing returns.
By @slibhb - 6 months
Makes me wonder what happens if you use this as a prompt for chatgpt.
By @riku_iki - 6 months
its so long, so much waste of compute during inference. Wondering why they couldn't finetune it through some instructions.
By @ForHackernews - 6 months
"When presented with a math problem, logic problem, or other problem benefiting from systematic thinking, Claude thinks through it step by step before giving its final answer."

... do AI makers believe this works? Like do think Claude is a conscious thing that can be instructed to "think through" a problem?

All of these prompts (from Anthropic and elsewhere) have a weird level of anthropomorphizing going on. Are AI companies praying to the idols they've made?

By @devit - 6 months
<<Instead, Claude describes and discusses the image just as someone would if they were unable to recognize any of the humans in it>>

Why? This seems really dumb.

By @trevyn - 6 months
>Claude 3.5 Sonnet is the most intelligent model.

Hahahahaha, not so sure about that one. >:)