AI with Michal

Few-shot prompting

Giving a language model a small set of completed examples (input plus desired output) so it infers tone, format, and constraints instead of you describing them only with adjectives.

Michal Juhas · Last reviewed May 2, 2026

What is few-shot prompting?

Few-shot prompting means you show the AI a few finished examples, each with an input and the answer you want, before you ask for something new. The model copies the pattern, tone, and layout from those examples instead of guessing from a long rule list.

Illustration: Few-shot prompting with example cards teaching an assistant before it writes a new draft

In practice

  • Before you ask for twenty outreach variants, you paste three real messages your team already sent that got replies. Trainers often say "show it a few good examples first" instead of using the term few-shot.
  • When you onboard a new recruiter, you share a doc with a bad email and a good email and tell them to drop those at the top of each ChatGPT or Claude session. That is the same habit with a simple file, not a lab notebook.
  • For job ad rewrites, you might paste one paragraph you like and one you hate so the model can see the tone gap in plain sight. Hiring managers recognize that flow even if they never name the technique.

Quick read, then how hiring teams use it

This is for recruiters, sourcers, TA, and HR partners who need the same vocabulary in debriefs, vendor calls, and policy reviews. Skim the first section when you need a fast shared picture. Use the second when you are deciding how it shows up in the ATS, sourcing tools, or candidate communications.

Plain-language summary

  • What it means for you: You show the AI two or three short examples of "good" output before you ask for a new one, like showing a new hire an old ticket.
  • How you would use it: You paste anonymized samples with the tone you want, then you ask for the next version in that style.
  • How to get started: Save one great intake note and one bad one. Label them. Ask the model to rewrite a third case like the great one.
  • When it is a good time: When long instructions did not work and the format still drifts between recruiters.

When you are running live reqs and tools

  • What it means for you: Few-shot is in-context learning: exemplars steer formatting and tone without fine-tuning. It competes with long prose rules and with system instructions for the same token budget.
  • When it is a good time: When you need consistent bullets, score snippets, or outreach variants under one brand voice.
  • How to use it: Rotate fresh examples, anonymize aggressively, and version the pack when comp or policy language changes.
  • How to get started: Read How to write better AI prompts and build a three-example library for your highest-volume ask.
  • What to watch for: Overfitting to three old reqs, leaking PII inside "good" examples, and stale shots nobody updates.

Where we talk about this

Live sessions compare few-shot packs to Gems and skills: same idea, different packaging. AI in recruiting focuses on tone and fairness; sourcing automation focuses on stable fields feeding prompts. Try both angles at Workshops.

Around the web (opinions and rabbit holes)

Third-party creators move fast. Treat these as starting points, not endorsements, and double-check anything before you wire candidate data.

YouTube

Reddit

Quora

Few-shot versus long instructions

StyleWhen it winsWatch out
Few-shotTone, format, micro-patternsHidden bias in samples
Long rubricLegal must-nots, complianceToken cost, skim risk
HybridProduction promptsNeeds an owner to edit both

Related on this site

Frequently asked questions

How many examples is "few" in practice?
Two to five input/output pairs usually beat a wall of prose: enough to show tone, structure, and edge handling without crowding candidate facts out of the LLM tokens budget. Live workshops show sourcers pasting three strong messages and getting usable fourth variants in seconds, then hitting diminishing returns by seven. Quality and recency beat count: retire examples when your bar or brand voice moves. Track which exemplar set produced each batch so you can roll back a bad week. Run a quarterly blind review where two reviewers score outputs from different example packs so weak sets retire with data, not office politics.
Where do few-shot prompts help recruiting most?
High-repeat artifacts with visible "gold" rows: outbound sequences, intake summaries, scorecard rationales, JD cleanup, and screening summaries where a sheet already holds ideal answers. Pair with how to write better AI prompts so each example encodes constraints hiring managers actually enforce. Few-shot shines when reviewers can diff output against a known anchor. It helps less for one-off executive searches where exemplars would be fake. Name a single owner per pack (often enablement or a lead sourcer) so updates do not fork across Slack threads, and publish where each pack may be used so GDPR reviews know which examples ever touched vendor logs.
What is the main downside of few-shot prompting?
Overfitting: the model copies quirks you did not mean to canonize (odd sign-offs, illegal phrasing someone once slipped through) or ignores edge cases your samples never covered. Without versioning, two recruiters silently maintain different example packs and downstream quality diverges. Refresh examples when comp, remote policy, or diversity language changes, and log updates like code. Pair few-shot with system instructions so global rules stay stable while examples rotate per req family. After a bad send, capture the exemplar row that misled the model so legal and TA see the same root cause instead of blaming "the AI" abstractly.
How is this different from a saved system prompt or Gem?
Few-shot teaches inside a turn or thread with fresh pairs; Gems and custom GPTs persist that teaching as system instructions across sessions. In practice you stack them: stable global rules plus three fresh exemplars for this quarter's reqs. If you only few-shot without persistence, new hires reinvent tone weekly. If you only systemize without examples, abstract adjectives creep back. Document which layer owns legal must-nots so updates do not fall through cracks. When IT rotates API keys or vendors ship new defaults, regression-test both layers with the same five anonymized profiles so drift shows up before candidates do.
Can few-shot reduce hallucinations?
It can reduce style drift and missing sections, but it does not stop factual invention: the model can still invent employers or dates that were not in the profile you pasted. Keep verify-before-send habits from the hallucination entry, especially for multilingual titles and stealth startups. Use few-shot to show the shape of a truthful answer ("only facts from the resume bullet"), not to imply omniscience. Pair numeric claims with a human spot-check until metrics say otherwise. Teach coordinators that exemplars are guardrails on tone and structure, not proof that every new candidate sentence is true.
Which tools support few-shot well?
Any chat UI that tolerates long prompts, plus Claude and ChatGPT when you pin examples above the user task. API users can separate system, developer, and user blocks for cleaner budgets. Pick tooling based on audit logs and data residency, not only example slots. For a guided path, join a workshop or take Starting with AI: the foundations in recruiting so you build packs with anonymization habits baked in. Security reviews should include a sample thread that shows exactly which block holds examples so approvers understand what leaves your boundary on each call.
Should I anonymize real candidate examples?
Yes, always strip names, emails, employers, and identifiable projects before examples land in third-party models or shared repos. Treat few-shot packs like internal documentation with the same retention and access rules as your CRM. Rotating real snippets without redaction is how accidental PII becomes training folklore. If you need realism, synthesize plausible composites with hiring manager review instead of copy-pasting a finalist's mail. Log who approved each composite and when, so if a candidate ever asks how their story was used you can answer without guessing.

← Back to AI glossary in practice