Roles, Personas, and System Prompts
Leverage roles and system instructions to shape expertise, tone, and boundaries across single and multi-agent setups.
Content
Calibrating Expertise Levels
Versions:
Watch & Learn
AI-discovered learning video
Sign in to watch the learning video for this topic.
Calibrating Expertise Levels
"Tell the model to be an expert" is the AI equivalent of saying "be cool" at a party. Vague, optimistic, and not very helpful.
You already know about selecting effective roles and how to write clear, actionable instructions without leading the model into a trap. Now we level up: how to calibrate the model's expertise so its answers match the depth, tone, and assumptions you actually need — from "explain like I'm new" to "peer-reviewed journal energy."
Why calibrating expertise matters (without the fluff)
- Bad calibration = answers that are too shallow, too technical, or just plain wrong for your audience.
- Good calibration = faster iterations, less prompting, and outputs you can actually use.
Think of the model as a very talented actor. You don't just say 'play Hamlet' — you say 'play Hamlet as a soap-opera star,' or 'play Hamlet as a Shakespeare professor teaching freshmen.' Same script, wildly different delivery.
The Anatomy of an expertise-calibrating system prompt
Here are the building blocks you combine to set expertise level precisely.
- Role + domain — Establish the persona and field (e.g., 'senior data scientist specializing in NLP').
- Experience signal — Years, milestones, or status words (e.g., '10+ years', 'PhD-level', 'industry principal').
- Depth & scope — How deep to go and what to assume about the reader (e.g., 'high-level overview' vs 'detailed math derivation').
- Style & constraints — Tone, verbosity limits, citation standards, and acceptance criteria.
- Deliverable format — Bullet list, code, proof sketch, executive summary, etc.
Combine these like a mixologist with a clipboard.
Practical calibration levels (quick reference)
| Level | Persona cue | Depth cue | When to use |
|---|---|---|---|
| Novice | 'explain like I'm a beginner' | analogies, no jargon | Onboarding, tutorials |
| Competent | 'mid-level engineer' | practical steps, minimal theory | How-to guides, reproducible recipes |
| Expert | 'senior researcher / PhD' | derivations, references, counterexamples | Research, audits, architecture design |
Sample system prompts (copy-paste ready)
Novice:
You are a patient tutor and beginner-friendly explainer in machine learning. Assume the reader has basic programming literacy but no prior ML knowledge. Use simple analogies, define each term the first time it appears, and provide one short example. Keep explanations under 200 words.
Competent:
You are a senior ML engineer. Assume the reader knows standard ML concepts (gradient descent, overfitting, validation). Provide a clear step-by-step plan with code snippets and pitfalls to watch for. No need for basic definitions. Limit to 6 steps and include one concise command-line example.
Expert:
You are a PhD-level researcher in NLP with 10+ years' experience. Provide a rigorous explanation including math where relevant, trade-offs, and citations to standard papers. Assume familiarity with probability, linear algebra, and optimization. Use formal notation sparingly and include one short proof sketch or complexity analysis.
Notice how each prompt modifies assumptions, not just verbosity.
Avoid these calibration traps (you know the drill)
- Don’t just say 'be an expert' — specify what expertise means in measurable terms.
- Don’t overload the persona with conflicting cues (e.g., 'be terse' + 'include long derivations').
- Avoid leading the model with answers; prefer constraints and acceptance criteria instead (this builds on the 'avoid leading the model' concept from earlier).
Tests to verify calibration (quick QA checklist)
- Consistency test: Ask the same question twice in two different phrasings. Do responses maintain depth and assumptions? If not, refine the 'assume' clause.
- Triage test: Give three follow-ups of increasing difficulty. The model should escalate complexity appropriately.
- Sample-check test: Require the model to produce a short example or equation that demonstrates the claimed level of expertise.
Example: After an 'expert' prompt, request a single equation or citation. If none appears, your prompt didn't truly evoke expertise.
Guided process to craft a calibrated prompt (5 steps)
- Define the outcome: what will you accept as a correct answer? (This follows the 'acceptance criteria' practice.)
- Pick the persona and justify it (why a senior dev? a researcher?).
- State assumed prior knowledge explicitly (what the reader already knows). Avoid ambiguity.
- Specify deliverable format and limits (length, sections, code, citations).
- Add a short evaluation request: 'At the end, include 1-sentence summary and 2 references or commands to validate.'
Example: Calibrating across the pipeline
Goal: Explain transformer self-attention for a product manager (not technical) and a research intern (technical).
Product manager (novice): system prompt should include 'non-technical analogies', 'no equations', 'impact on product metrics'.
Research intern (competent-to-expert): system prompt should include 'math sketch of attention', 'complexity O(n^2)', 'one code snippet in PyTorch', and 'one recent paper citation'.
Different audiences, different assumptions, same base concept.
Closing — the one weird trick (not really magic)
Calibrating expertise is about turning fuzzy requests into explicit assumptions and measurable acceptance criteria. Combine role, experience signal, assumed prior knowledge, and output constraints. Test with small checks (example, equation, citation) and iterate.
Expertise in prompts isn't status signaling. It's practical: it saves time, reduces churn, and produces outputs you can trust.
Key takeaways:
- Be explicit about assumed knowledge. Don’t let the model guess.
- Use measurable signals (years, PhD, citations) rather than vague labels.
- Pair persona with format and acceptance criteria.
Go tweak a system prompt now: pick a concept, pick an audience, and write a 2-line persona that forces the model to reveal its level. Bonus: use the tests above and watch your outputs stop being wishful thinking and start being useful.
Comments (0)
Please sign in to leave a comment.
No comments yet. Be the first to comment!