AI with Michal

AI and ML in recruitment

AI in recruitment spans language models, rule-based automation, and machine learning algorithms, each working differently: ML models learn from historical hire data to predict outcomes, while LLMs generate and interpret text. Understanding the difference helps TA teams evaluate vendor tools critically and structure bias audits correctly.

Michal Juhas · Last reviewed May 9, 2026

What is AI and ML in recruitment?

AI and ML in recruitment refers to two overlapping but distinct categories of technology applied across sourcing, screening, scheduling, and reporting in the hiring lifecycle. Machine learning is a subset of AI where systems learn patterns from historical data instead of following hand-coded rules. Language models, the technology behind tools like ChatGPT and Claude, are a further specialisation that reads and generates text.

In a recruiter's daily workflow these distinctions matter. A vendor claiming their tool uses "AI" might mean an ML classifier trained on past hire outcomes, an LLM generating outreach drafts, a simple keyword filter with a marketing label, or some combination. Understanding which mechanism is active at each step determines what questions to ask in a vendor demo, what compliance checks to run, and where to put human review gates.

Illustration: AI and ML in recruitment showing three technology types, rule-based automation, an ML model trained on historical hire data, and a language model, each connected to different recruiting stages with a human review gate at ML-driven decisions and a bias audit checkpoint

In practice

  • When an ATS vendor says their tool "AI-ranks" candidates, they usually mean an ML model scored resumes by predicted probability of advancing based on past hires at similar companies, a very different mechanism from an LLM that reads a resume and writes a summary.
  • A sourcer asking "why did the tool surface these three people?" will get a different answer depending on whether the system uses keyword matching, embedding-based semantic similarity, or a supervised classifier, three mechanisms that look identical in a UI but require different calibration conversations with a hiring manager.
  • A TA ops leader saying "the model is biased" and an engineer saying "the model is working as trained" are often both right: ML models encode patterns from labelled historical data, so bias from past decisions becomes a feature until someone audits it.

Quick read, then how hiring teams use it

This is for recruiters, sourcers, TA partners, and HR leaders who need a working vocabulary for vendor conversations, debrief discussions, and audit decisions. Skim the first part for a shared picture. Read the second when you are choosing tools, reviewing contracts, or running an adverse impact check on a model output.

Plain-language summary

  • What it means for you: AI in hiring covers at least three different types of technology. Knowing which one a vendor uses helps you ask the right questions and avoid trusting a score that was never calibrated for your roles.
  • How you would use it: When evaluating a tool, ask: does this predict from past data, or generate from a language model? Prediction tools need an adverse impact audit. Generation tools need a hallucination review step.
  • How to get started: Map the AI in your current ATS and sourcing tools to one of three types: rule-based filter, ML model, or LLM. Then ask each vendor for the last bias audit result or training data date.
  • When it is a good time: Before you expand any AI-assisted step to high-volume or high-stakes decisions, not after a compliance question arrives.

When you are running live reqs and tools

  • What it means for you: ML models rank or score based on patterns from past hires. LLMs generate text based on broad training corpora. Both can fail silently: a mis-tuned ML model will keep producing confident scores, and an LLM will keep drafting polished text even when key facts are wrong.
  • When it is a good time: Use ML-based ranking when volume exceeds what humans can review manually and you have audit infrastructure in place. Use LLM drafting as soon as prompts are stable and outputs pass a human read before touching candidates.
  • How to use it: Put a human-in-the-loop gate at every ML decision point where an individual could be rejected. Log the model version and threshold for each run so compliance questions have a paper trail.
  • How to get started: Ask your ATS vendor which decisions their AI layer makes versus assists. Get pass-rate data by demographic group before you treat any ML output as a shortlist rather than a starting point.
  • What to watch for: Vendors who conflate ML and LLM under a single "AI" label to avoid disclosing training data sources, silent model retraining that changes scoring without notice, and LLM outputs that paste confidently wrong candidate details into ATS records.

Where we talk about this

On AI with Michal live sessions, the distinction between ML models and LLMs comes up in the first hour of every AI in recruiting cohort. The sourcing automation track covers semantic search and embedding-based retrieval as practical alternatives to keyword matching. The AI in recruiting track addresses the compliance landscape for ML-powered vendor tools, including how to read an adverse impact report and what to ask when a vendor says their model is audited. Start at Workshops and bring your current ATS name and a real role so the room can test the abstractions against something concrete.

Around the web (opinions and rabbit holes)

Third-party creators move fast here. Treat these as starting points, not endorsements, and verify compliance postures and vendor details directly before wiring candidate data to any script you find.

YouTube

Reddit

Quora

AI versus ML versus automation in recruiting

TechnologyWhat it learns fromTypical use in recruitingMain risk
Rule-based automationNothing, follows fixed logicATS routing, email triggersMis-mapped fields, silent errors
ML modelHistorical labelled decisionsResume ranking, attrition predictionBias encoded from past screeners
Language model (LLM)Broad text training corpusDrafting, summarising, searchHallucination, stale or wrong facts

Related on this site

Frequently asked questions

What is the difference between AI and ML in recruitment?
AI is the broad category; ML is a subset where a system learns patterns from data rather than following hand-coded rules. In recruiting, the phrase usually covers three distinct mechanisms: an ML model trained on past hire data to predict outcomes such as resume rank or attrition risk; a large language model generating or summarising text; or rule-based automation routing candidates through ATS stages. The distinction matters for compliance: an ML model can silently encode historical bias, an LLM can hallucinate qualifications, and rule-based automation can mis-map fields. Knowing which mechanism is inside a vendor tool changes what due diligence your team runs before deployment.
Which ML techniques are most commonly used in recruiting?
Most recruiting ML falls into four families. Supervised classification decides if a resume matches a target profile using labels from historical hires. Regression predicts continuous outcomes such as time-to-hire or attrition risk. Embedding models turn text into vectors so semantic search surfaces similar roles or candidates without exact keyword matches. Ranking algorithms order shortlists by predicted probability of advancing. Each carries different bias exposure: classification amplifies past screening patterns, embeddings inherit semantic associations from training corpora. When a vendor says their AI scores candidates without naming the technique, ask which of these four their model uses and what training data governed the labels.
How does ML-based resume screening actually work?
A supervised model trains on thousands of past applications labelled by outcome: advanced, rejected, hired, not hired. It identifies patterns in those documents, job titles, tenure gaps, and skills keywords, that correlate with the labels, then applies those patterns to new resumes. The problem is that labels carry whatever bias existed in human screeners who made earlier decisions. If a hiring team consistently advanced candidates from particular backgrounds, the model learns to prefer those signals even when they are legally irrelevant. This is why adverse impact analysis is mandatory before deploying ML screening at volume. The model surfaces statistical patterns; a human still needs to audit what it learned.
What are the bias risks specific to ML models in hiring?
ML models in hiring carry bias risks that are subtler than human bias because they scale instantly and produce confident outputs. A model trained on historical hire data learns whatever correlates with past success, including protected characteristics that should not influence decisions: names that signal gender or ethnicity, schools that proxy for socioeconomic background, or ZIP codes that correlate with race. Unlike an LLM that can be prompted toward fairer language, an ML model bakes bias into its weights during training. Mitigations include running a formal AI bias audit on pass rates by demographic group, retraining on balanced labels, and keeping a human-in-the-loop gate before any reject decision.
How should TA teams evaluate ML-powered vendor claims?
Before signing a contract for an ML-based recruiting tool, request four things: the training data summary including time period and source companies; the last adverse impact audit with pass rates broken down by gender, race, and age; whether the model retrains on your historical decisions or only on vendor-wide data; and what the override mechanism looks like for a recruiter who disagrees with a score. Vendors who cannot answer the first two questions promptly are signalling something about their audit posture. Run a structured demo against your own briefs, not the vendor benchmark, and have legal review the data processing agreement before you pass candidate records to any external model.
When is ML overkill for a recruiting team?
ML models add value when you have thousands of applications per role and enough historical hire data to train on without encoding too narrow a pattern. For teams hiring fewer than a few hundred people per year, a well-written scorecard plus an LLM for drafting typically outperforms a black-box ML ranker and is far easier to audit. ML is also overkill when the role is genuinely new, when headcount is frozen, or when historical hire data spans only a few years across a handful of job families. Use ML where volume justifies the investment in bias monitoring and retraining; use prompting and structured review for everything else.
Where can I learn AI and ML recruiting skills alongside peers?
A workshop on AI in recruiting walks through how ML tools actually make decisions, what to ask vendors before the demo ends, and how to connect LLM drafting to ATS workflows without creating compliance exposure. The sourcing automation track goes deeper on embeddings, semantic search, and the difference between prompt-based and model-based retrieval. For self-paced foundations, Starting with AI: the foundations in recruiting covers the vocabulary before you sit across from a vendor. Membership adds monthly office hours where practitioners share what is working. Bring your vendor tool list and a real req so feedback is grounded in your stack.

← Back to AI glossary in practice