AI with Michal

Large language model (LLM)

A neural model trained to predict the next token over broad text, which powers chat assistants, drafting, classification, and tool-using agents when wrapped in product guardrails.

Michal Juhas · Last reviewed May 2, 2026

What is a large language model (LLM)?

A large language model is software trained on lots of text so it can predict the next words in a reply, which powers assistants like ChatGPT or Claude. It helps you draft and sort recruiting work, but it is not your ATS and it can still be wrong.

Illustration: An LLM core turning job inputs into drafts, summaries, and tags for recruiting work

In practice

  • When someone says "we need an LLM for scheduling," they usually mean "we want ChatGPT-style help," not a lecture on neural nets. The acronym shows up on procurement forms, vendor blogs, and IT approval emails.
  • In a hiring stand-up, people compare "the model guessed" with "the ATS field says" when quality drops. They may not name one vendor model versus another, but everyone knows an assistant wrote the text.
  • IT sends a note like "approved assistants are X and Y for customer data," which is how many companies first introduce the term LLM to recruiters who only care that the button works.

Quick read, then how hiring teams use it

This is for recruiters, sourcers, TA, and HR partners who need the same vocabulary in debriefs, vendor calls, and policy reviews. Skim the first section when you need a fast shared picture. Use the second when you are deciding how it shows up in the ATS, sourcing tools, or candidate communications.

Plain-language summary

  • What it means for you: An LLM is software that finishes sentences like autocomplete grew up: it can draft email, summarize notes, and still be wrong about facts.
  • How you would use it: You give short instructions plus the facts you already trust, you read the draft, you send.
  • How to get started: Try the same task in two products you already have licenses for; compare tone, not only speed.
  • When it is a good time: When repeat writing eats your week but you still have a human who owns quality.

When you are running live reqs and tools

  • What it means for you: An LLM is a next-token predictor over a context window, tuned for helpful chat and tools. APIs expose tokens, temperature, and structured modes.
  • When it is a good time: When you move from ad hoc chat to shared system instructions, RAG, or automation.
  • How to use it: Pick evaluation sets (intake, outreach, screening notes), log failures, and separate model swaps from prompt changes.
  • How to get started: Read How to use AI in recruiting and align TA and IT on data handling before you expand.
  • What to watch for: Treating the model like a database, shipping prose where you needed structured output, and skipping red-team time on multilingual inputs.

Where we talk about this

AI in recruiting blocks use LLMs as the shared layer hiring managers actually see. Sourcing automation blocks ask which calls are chat-only versus API-backed. Both show up at Workshops with different homework.

Around the web (opinions and rabbit holes)

Third-party creators move fast. Treat these as starting points, not endorsements, and double-check anything before you wire candidate data.

YouTube

Reddit

Quora

Chat versus API depth

LayerYou getYou still need
Chat UIFast draftsCopy-paste hygiene
Saved skills / GemsConsistencyOwners and updates
API + automationScaleKeys, monitoring, GDPR

Related on this site

Frequently asked questions

What should recruiters actually know about how an LLM works?
Enough to calibrate trust: it predicts likely next tokens, it does not query your ATS unless wired through tools, and it has a finite context window that trades off with instructions and attachments. That mental model prevents magical thinking in debriefs when someone says "it should know our policy." Pair the overview with your vendor's logging story: who can see prompts, where data is processed, and how long transcripts persist. When hiring managers ask for certainty, translate model behavior into review steps you already staff instead of impossible guarantees.
How do I choose between vendors (OpenAI, Anthropic, Google, and others)?
Start from governance and workflow fit: EU data handling, SSO, audit logs, retention controls, and whether you need API access for workflow automation. Benchmarks help compare draft quality, but procurement should weight incident response and subprocessors as highly as leaderboard scores. Run the same twenty recruiting prompts from your last quarter across finalists and score factual errors, not only fluency. Involve IT and legal before you standardize, or you will re-platform six months later when security review finally runs. Capture a side-by-side matrix of logging, residency, and red-team results so executives compare evidence, not slogans.
Is a bigger model always better for recruiting text?
Not always. Smaller models with strong system instructions, curated Markdown for AI, retrieval, and few-shot prompting often beat a frontier model fed a vague paragraph. Cost and latency matter when you scale across reqs and languages. Measure end-to-end time including human review, not only tokens per second. Sometimes the right answer is two specialized calls in a prompt chain instead of one giant completion. Re-evaluate after major vendor releases, because the cheapest accurate stack this quarter may not stay cheapest after pricing or safety filters change.
What is the difference between an LLM and automation?
The LLM proposes text, labels, or summaries; automation (Make, n8n, webhooks) moves structured data between systems and triggers actions. Workshops separate "just prompting" from skills in project folders and APIs because integration depth changes GDPR surface area and who gets paged at night. Mixing the two without boundaries yields ghost sends or silent CRM corruption. Document which nodes are allowed to write candidate-visible fields versus draft-only fields. Add monitoring on automation branches that call models so token spikes or schema errors page someone before candidates see broken templates.
Where do maturity models help?
They align TA, HRBPs, sourcing, and finance on how deep you go this quarter versus next, so budget asks map to observable milestones instead of "more AI." Read AI adoption maturity levels and pair it with the AI adoption ladder glossary entry for concrete artifacts per stage. Name owners per stage (prompt library, automation keys, corpus hygiene) or the model slides backward after one busy month. Maturity models fail when they are only marketing; they work when tied to metrics and risk reviews.
Which on-site tools should we standardize on first?
Most teams pick one chat assistant plus one automation path so recruiters are not juggling five stacks with different retention rules. Compare ChatGPT, Claude, and n8n in the directory against your regions and SSO needs before you buy overlapping seats. Run a thirty-day pilot on two real reqs with sourcers and recruiters logging friction daily. Standardization beats feature sprawl when compliance asks for a single DPA map. Publish the approved stack list internally so shadow tools lose their "everyone else uses it" excuse.
Do we need an engineer to use LLMs responsibly?
For chat-first drafting with human send gates, no dedicated engineer is required if TA partners with IT on accounts and logging. For synced CRM writes, webhooks, or bulk processing, you need engineering or a strong ops partner who treats scripts like production. The Starting with AI: the foundations in recruiting course stays recruiter-native first so you earn governance habits before API complexity. Document escalation paths when a model misbehaves in automation, not only when chat feels off. Name a liaison who can read vendor status pages during incidents so recruiters are not guessing alone at 2 a.m.

← Back to AI glossary in practice