AI with Michal

Hallucination

When a language model produces fluent but false or ungrounded details (employers, dates, URLs, policy claims) that look credible until you verify them.

Michal Juhas · Last reviewed May 2, 2026

What is a hallucination?

A hallucination is when the AI writes something that sounds confident but a fact is wrong or was never in the data you gave it. In recruiting that can be wrong employers, dates, or links, so you check facts before anything goes to a candidate.

Illustration: AI output checked against a profile card to catch a factual mismatch before sending

In practice

  • After a model drafts a candidate blurb, someone says "double-check the dates and employer" in debrief because once in a while the city or level is wrong. You hear "hallucination" in AI safety talks and more often now in recruiter Slack groups.
  • Legal or HR may ask "how do we know the assistant did not invent this sentence" when they review an internal FAQ or offer letter draft that touched AI at some step.
  • Before you paste an InMail into LinkedIn, you scroll the profile next to the draft even when the mail reads smoothly. That side-by-side habit is how teams catch mistakes without saying a jargon word at every step.

Quick read, then how hiring teams use it

This is for recruiters, sourcers, TA, and HR partners who need the same vocabulary in debriefs, vendor calls, and policy reviews. Skim the first section when you need a fast shared picture. Use the second when you are deciding how it shows up in the ATS, sourcing tools, or candidate communications.

Plain-language summary

  • What it means for you: The model can sound confident while wrong about a date, a job title, or a city. You treat pretty sentences like a junior who needs a fact check.
  • How you would use it: You keep LinkedIn or the ATS open beside the draft and you fix mismatches before send.
  • How to get started: Run five real profiles through your prompt and mark every line that is not directly supported by the profile text.
  • When it is a good time: Always for candidate-facing text; especially when you personalize outreach at scale.

When you are running live reqs and tools

  • What it means for you: Hallucination is confident error: plausible language without grounded facts. Risk rises with open-ended personalization and long contexts.
  • When it is a good time: When you automate enrichment, chains, or RAG without error budgets.
  • How to use it: Constrain outputs (structured output), cite only verified fields, and log incidents the way you log offer-letter typos.
  • How to get started: Re-read workshop notes on verify-before-send and tighten prompts to restate facts, not invent narrative.
  • What to watch for: Beautiful Markdown that hides wrong seniority, multilingual title mismatch, and reviewers skimming under time pressure.

Where we talk about this

Participants often meet hallucinations first on outbound: wrong office city or product name. The fix is rarely "a better model" and usually verify-before-send plus shorter prompts. We rehearse that live in Workshops.

Around the web (opinions and rabbit holes)

Third-party creators move fast. Treat these as starting points, not endorsements, and double-check anything before you wire candidate data.

YouTube

Reddit

Quora

Hallucination risk by task

TaskRelative riskMitigation
Outreach personalizationHighFacts from profile only
Intake summaryMediumQuote hiring manager
Boolean stringLowerStill test in tool
Policy interpretationHighLegal, not LLM

Related on this site

Frequently asked questions

What do hallucinations look like in recruiting work?
Fluent lies about employers, dates, titles, certifications, or URLs that look plausible until you open the profile or ATS field beside the draft. Co-pilot style drafting on thin profiles is especially risky because the model fills gaps the way humans guess, then states guesses confidently. Teams catch them fastest when they keep source tabs pinned: LinkedIn next to InMail, policy PDF next to internal FAQ. Log incidents with the prompt version so you can see whether a template change caused a spike, not random bad luck.
Why do models hallucinate if they are "trained on the internet"?
They optimize for plausible continuation, not verified lookup against your private candidate truth. Training on broad text does not grant live access to your CRM or to yesterday's policy addendum. Even with RAG, wiring mistakes or stale chunks can surface wrong snippets with confidence. Teach stakeholders the "pattern completion" mental model so they stop treating fluent paragraphs as citations. Pair vendor claims with your own eval set of twenty tricky profiles from last quarter. When executives ask for a demo, show a deliberate miss next to a fix so they internalize that fluency is not the same as accuracy in hiring.
What is the minimum viable verification loop?
For candidate-facing text: keep the authoritative profile or ATS record visible, spot-check employers, dates, and locations, and require URLs or quotes pulled directly from source text rather than memory. For internal drafts, tag every claim that needs a citation before it leaves the team channel. Add a second reader on high-risk segments (exec outreach, visa-sensitive wording). Minimum viable still needs an owner: who is allowed to click send when a deadline looms and the draft "looks fine"? Publish that escalation path beside your templates and review it after every near-miss retro so new hires do not improvise under pressure.
Does RAG eliminate hallucinations?
It reduces unsupported claims when retrieval returns the right chunk, but models can still misread tables, merge two policies, or cite an outdated file with equal confidence. Treat RAG as assist, not oracle, and keep hallucination checks on numbers, URLs, and eligibility language. Run quarterly corpus audits so "grounded" answers are not grounded in 2019. Log retrieval misses separately from model mistakes so engineering and TA each know what to fix. When retrieval confidence scores exist, teach recruiters what "low confidence" means in plain language before those fields drive automation.
How do workshops talk about this with hiring managers?
Plain language: models draft structure and phrasing fast; humans own facts, fairness, and tone that matches the team they will join. That framing prevents "the computer said so" approvals in debriefs and sets expectations on turnaround time (review is a step, not overhead). We share anonymized misses so hiring managers feel why verify-before-send matters for their brand, not only compliance. Tie the conversation to score anchors so quality discussions stay behavioral, not mystical. Close with one concrete habit they can adopt this week, such as keeping the ATS tab open while approving outreach, so the lesson survives the slide deck.
Which blog posts should the team read together?
Start with AI candidate screening and How to use AI in recruiting as a pair: one on funnel risk, one on operating norms. Then align on tools with ChatGPT for recruiters so procurement hears the same limits engineering does. Reading as a group surfaces disagreements early (what counts as public data, who approves drafts) and turns policy into behavior, not PDFs on a shelf. Capture three decisions per session in your Markdown for AI knowledge base with owners and dates so assistants and new hires inherit the same story six months later.
When should we avoid generative models entirely?
Skip generative passes for high-stakes compliance narratives, redundancy selections, compensation communications, or anything you cannot audit under current policy. Prefer deterministic templates, official legal review paths, or vendor features with contractual guarantees there. Temporary bans are fine while you build rubrics; permanent bans without alternatives just drive shadow IT. Document the decision with names and dates so future leaders know it was intentional, not ignorance. Revisit bans when retrieval, logging, or human-in-the-loop controls mature, because blocking tools without a safe lane rarely stops motivated recruiters.

← Back to AI glossary in practice