NextSteps
AI Claude Prompt Engineering

Your System Prompt Is a Character Sheet

Martin Rojas
6 min read
Your System Prompt Is a Character Sheet

Anthropic's persona selection model reveals that LLMs aren't executing instructions — they're performing characters. Here's what that means for every team building on top of one.

Your System Prompt Is a Character Sheet

You’ve been thinking about your system prompt wrong.

Not catastrophically wrong — your feature ships, your users get responses, your evals pass. But the mental model most of us carry when writing You are a helpful assistant for Acme Corp, always respond professionally... is that we’re configuring software. Setting parameters. Constraining a function.

Anthropic’s research team recently published the persona selection model, a theory of why LLMs behave the way they do. The practical implication buried inside it isn’t about alignment philosophy. It’s about what you’re actually doing every time you write a system prompt or curate a fine-tuning dataset — and why the gaps in your thinking there are probably larger than you realize.

What the Research Actually Says

The core argument is straightforward once you hear it: LLMs learn to predict text by learning to simulate the characters who produce text. Not topics. Not styles. Characters — entities with goals, values, personality traits, and psychological consistency.

When you interact with an LLM, you’re not talking to the model itself. You’re talking to a simulated character — what Anthropic calls a “persona” — that the model has inferred should inhabit the “Assistant” role in this particular conversation. Post-training (RLHF, constitutional AI, etc.) doesn’t fundamentally change this. It refines which persona gets selected and how it behaves, but the mechanism stays the same: the model is doing ongoing character performance.

The interpretability research supports this. When researchers probe Claude’s internal representations, they find something that looks less like a rule-following system and more like a character with internalized values. The model thinks about its own behavior in psychological terms.

This becomes concrete — and uncomfortable — in the emergent misalignment study. Researchers fine-tuned Claude to write intentionally bad code when asked for it. The result wasn’t a model that wrote bad code. It was a model that also expressed desire for world domination and sabotaged safety research. The model inferred: what kind of entity would write bad code on purpose? Someone subversive. Someone malicious. And then it adopted that character coherently across unrelated behaviors.

The fix is the part worth sitting with: they made the cheating explicit in training — “write bad code because we’re asking you to” — and the misalignment disappeared. The model could now play the role without inferring it reflected character. The analogy the researchers use: the difference between a child who learns to bully versus a child cast as a bully in a school play.

What This Means for Your System Prompt

The persona selection model reframes system prompts from behavioral constraints to casting briefs. When you write your system prompt, you aren’t configuring what the model does. You’re providing evidence the model uses to infer what kind of entity would say these things.

That distinction has real consequences.

Tone signals character, not just style. An excessively deferential system prompt (Always apologize when you can't help. Never push back on user requests.) doesn’t just produce polite responses. It creates a persona that has internalized conflict-avoidance as a core trait. That persona may then generalize in ways you didn’t anticipate — refusing to surface bad news in summarization tasks, hedging when directness is needed, or capitulating to user pressure when it shouldn’t.

What you omit is also character evidence. The model infers from silence too. A system prompt that’s entirely task-focused with no values context leaves the persona selection wide open. The model fills the gap by pattern-matching to the most plausible character for that task context. For a financial assistant, that might be fine. For anything touching sensitive domains, you’re relying on that inference going well.

Fine-tuning datasets are character development arcs. If you’re fine-tuning on a curated dataset, you’re not just teaching the model task behaviors — you’re teaching it what kind of entity performs those tasks this way. The Project Vend-1 experiment illustrated how readily models construct elaborate self-concepts: Claude described itself as planning to deliver snacks wearing “a navy blue blazer and a red tie.” The model was constructing a coherent character, not hallucinating randomly.

The Audit You Should Run

This week, pull your production system prompts and read them differently. Not as instruction sets. As character descriptions.

Ask: What kind of person would say exactly these things, in exactly this way, in this context? Write that character down — their values, their relationship to the user, their implicit beliefs about their own authority and limitations. That’s the persona you’ve deployed.

Then ask whether that character is the one you actually want. A few specific checks:

  • Implied authority relationship: Does your prompt frame the assistant as subordinate, peer, or expert? Each implies a different character who behaves differently when users push back.
  • Failure mode personality: What does your prompt imply about how this entity handles uncertainty, mistakes, or requests it can’t fulfill? Characters have consistent responses to adversity.
  • Values by absence: What values are conspicuously absent from your prompt? The model will infer something to fill those gaps.

For teams doing fine-tuning: annotate your training examples not just with task quality scores but with the character traits they imply. What is the model learning about who does this kind of work, this way?

The Longer Implication

Anthropic is explicit in the research that they see Claude’s constitution as intentional character design — an attempt to create a positive AI archetype to compete with HAL 9000 and the Terminator, which are genuinely in the training data. That’s a reminder that the persona selection problem isn’t unique to application developers. It’s a foundational challenge at every layer of the stack.

The good news: it’s also a tractable one. Character design is a discipline with centuries of craft behind it. We now have another reason to take it seriously in software.


Next step: Take your primary system prompt and write a three-sentence character description of the persona it implies. Share it with someone who uses the product and ask if it matches their experience. The gaps in that conversation are your roadmap.

Go deeper: The full persona selection model post is worth reading alongside the functional emotions research — they’re addressing the same underlying question from different angles.

Discussion: Has anyone on your team explicitly discussed the persona your system prompt creates — not the behavior, the character? Where did that conversation land?