AI with Michal

Agent knowledge base

A curated set of Markdown or text files (often in a Claude project, repo, or shared folder) that gives an assistant stable facts about how you hire: tone, templates, scorecard rules, and disallowed phrases, without retyping them each session.

Michal Juhas · Last reviewed May 2, 2026

What is an agent knowledge base?

An agent knowledge base is a small set of files your team keeps fresh so an AI assistant can read the same facts you rely on. It cuts repeat questions and keeps tone, roles, and rules in one place.

Illustration: A curated knowledge shelf feeding an assistant that cites internal binders

In practice

  • A Confluence or Drive folder called "How we hire at X" holds tone, email limits, and visa basics, and the team tells new hires to link it when they use the assistant. Vendors may call it a "knowledge base for agents" in release notes.
  • IT pilots an internal chatbot and asks TA to upload five canonical pages first. Recruiters experience it as "the bot finally knows our acronyms" instead of guessing from the open web.
  • Stand-ups say "update the knowledge base" when comp bands or remote policy change so the assistant does not keep quoting last year's file.

Quick read, then how hiring teams use it

This is for recruiters, sourcers, TA, and HR partners who need the same vocabulary in debriefs, vendor calls, and policy reviews. Skim the first section when you need a fast shared picture. Use the second when you are deciding how it shows up in the ATS, sourcing tools, or candidate communications.

Plain-language summary

  • What it means for you: It is the short library your AI is allowed to read before it answers, like the binder on your desk with tone rules, acronyms, and the latest comp policy.
  • How you would use it: You upload a few approved pages, you tell the bot "only use this," and you still check anything that goes to a candidate.
  • How to get started: List five questions the team asks every week. Find the one canonical answer for each. Put those answers in Markdown files with dates in the title.
  • When it is a good time: When everyone gets different answers from the same tool because nobody shared the same context.

When you are running live reqs and tools

  • What it means for you: A knowledge base is curated retrieval: files or chunks the model can cite, paired with owners, diffs, and refresh rules. It is the recruiter-facing side of RAG before you buy a vector database.
  • When it is a good time: When LLM tokens spike from "upload everything" dumps but quality does not move.
  • How to use it: Prefer Markdown for AI, version in Git or a shared drive with permissions, and tie updates to policy owners. Forked copies per recruiter are a smell.
  • How to get started: Mirror one internal FAQ, measure fewer repeat questions in Slack, then widen.
  • What to watch for: Stale PDFs, pasted screenshots of old rules, and duplicate "final" files that disagree.

Where we talk about this

AI in recruiting blocks treat the knowledge base as the bridge between "we bought licenses" and "hiring managers trust the answers." Sourcing automation days connect the same files to API-heavy stacks. Bring your real folder chaos to Workshops.

Around the web (opinions and rabbit holes)

Third-party creators move fast. Treat these as starting points, not endorsements, and double-check anything before you wire candidate data.

YouTube

Reddit

Quora

Knowledge base versus chat-only memory

PatternStrengthWeakness
Chat-onlyFast startContext lost, hard to audit
Shared Markdown basePortable, diffableNeeds owners
Drive dumpEasy uploadNoisy, expensive tokens

Related on this site

Frequently asked questions

How is this different from RAG?
RAG usually implies dynamic retrieval over a larger corpus at query time, with chunking and ranking tuned per question. An agent knowledge base is often smaller, hand-maintained, and versioned like internal product docs your team trusts day to day. You can combine both: stable "always read" files plus retrieval for long archives. The distinction matters for ownership: who deletes outdated comp guidance matters more than whether you used vectors. Start small so reviewers can actually read everything in the base quarterly.
What files belong in the first version?
Ship a minimum lovable corpus: employer or agency positioning, channel rules (InMail versus email), three anonymized strong outreach examples, scorecard anchors with observable behaviors, booking links, and a short "do not say" list legal approved. Use Markdown for AI so changes are diffable, and mirror key bullets into system instructions inside vendor UIs for consistency. Avoid dumping every PDF from 2019; stale files are how assistants confidently cite wrong visa lines. Tag v1 with a README that links to the approval ticket so future editors know which files counsel already blessed.
Who maintains it?
Name a rotating owner (sourcer, recruiter, or TA ops) with quarterly review on the calendar, not "the team." Without ownership, enthusiasm from one workshop decays into conflicting copies in personal drives. Maintenance includes deletes, not only adds: when comp bands or remote policy change, retire old files loudly. Pair maintenance with metrics recruiters feel: fewer repeated Slack questions, faster HM alignment on tone. Escalate access issues to IT early so contractors are not editing canonical tone files anonymously. Publish the on-call rotation beside the repo README so PTO does not pause every update.
What data should never live there?
Unredacted candidate PII, unreleased compensation bands you cannot defend in audit, secrets without legal review, or anything you would not paste into a vendor support ticket. Treat the folder like HR documentation with retention and access rules, especially if workflow automation later pipes excerpts into API calls. If a file is "for AI only," it still counts as processing personal data. When unsure, ask counsel before you optimize for convenience. Keep a short blocklist of file types (raw CRM exports, full comp grids) that assistants should refuse even when someone uploads them in a hurry.
How does this connect to automation?
Once files are stable, workflow automation can pass excerpts or structured fields into model calls per ATS row, but only after you have logging, retries, and a human inbox for failures. Automation should read the same canonical Markdown humans edit, not forked snippets in a Zapier field nobody tracks. Version the prompt templates that wrap excerpts so you can roll back when a hiring manager flags tone. Test with synthetic rows before you touch real candidates. Add a dry-run mode that logs proposed payloads without sending until a named reviewer flips production on.
Where can we learn the habits around it?
Read AI-native for the operating style, climb the AI adoption ladder deliberately with artifacts not slogans, and practice in a workshop or the Starting with AI: the foundations in recruiting course. Bring your worst "folder of final" story so peers can help you design naming and review habits that survive PTO. If you are first in the company, publish a short charter that names approvers, review cadence, and which directories are canonical versus sandbox experiments so IT and legal know where truth lives.

← Back to AI glossary in practice