AI with Michal

Async assessment platform

Software that lets candidates complete skills tests, video responses, or work samples on their own schedule, then stores structured results for reviewers to evaluate without a live scheduling dependency.

Michal Juhas · Last reviewed May 15, 2026

What is an async assessment platform?

An async assessment platform lets candidates complete evaluated tasks on their own schedule rather than on a live call. The candidate logs in at a time that suits them, submits a work sample, records a video response, or takes a skills test, and the platform stores the structured output for a reviewer to evaluate later.

The key feature is time-decoupling: the candidate does not need to be available when the reviewer is available, and the reviewer does not need to be available when the candidate submits. That asymmetry removes scheduling as a bottleneck in high-volume pipelines and lets teams evaluate more candidates without adding more live screening capacity.

Illustration: async assessment platform routing candidate submissions through a timed self-serve task, a scoring hub, and a human review gate before the structured result enters the ATS hiring pipeline

In practice

  • When a high-volume customer support team sends a written scenario response to every applicant before the phone screen, and a TA coordinator reviews the results in a batch each morning, that is an async assessment platform workflow. The coordinator is not scheduling calls with 80 candidates; they are reviewing 80 structured submissions.
  • Phrases like "take-home assignment," "skills test," and "async video screen" all refer to variations of the same idea, though they imply different formats. The platform that hosts and scores them is the async assessment platform.
  • TA ops teams that report lower time-to-first-decision on senior engineering roles often point to moving a work sample before the hiring manager screen as the change, because the manager spends the live call on the output rather than probing for baseline skill.

Quick read, then how hiring teams use it

This is for recruiters, sourcers, TA, and HR partners who need the same vocabulary in debriefs, vendor calls, and policy reviews. Skim the first section when you need a fast shared picture. Use the second when you are deciding how async assessment fits into your stack, your compliance obligations, or your candidate experience.

Plain-language summary

  • What it means for you: Instead of scheduling a call to ask someone if they can do the job, you send them a task and review what they actually produced. The review happens on your timeline, not theirs.
  • How you would use it: You pick a task that tests a skill the role requires, you write a rubric before sending it, and you review every submission against the same criteria. The platform stores the result and pushes it to your ATS.
  • How to get started: Identify one screening question your team asks the same way on every phone screen. Write it as a scored task with a three-point rubric. Pilot it with five candidates on your next open req and compare the output to your current phone screen notes.
  • When it is a good time: When you are screening more candidates than your team can schedule calls with, when the role has a clear and testable skill, and when your process is stable enough that the rubric will not change mid-search.

When you are running live reqs and tools

  • What it means for you: Every scored output is data that can be audited, compared across groups, and questioned by a candidate who receives a rejection. That is a governance asset when the rubric is documented and a liability when it is not.
  • When it is a good time: After the scoring rubric is written, reviewed by the hiring manager, and tested internally. Not before. Sending an assessment without a rubric is scoring based on vibes, which undermines the main reason to use the platform.
  • How to use it: Wire the platform to your ATS so scores land in the candidate record without manual entry. Configure the ATS stage to require human review before auto-advance. Log the rubric version, the reviewer, and the submission date alongside the score. Cross-link to workflow automation once the data mapping is trusted.
  • How to get started: Run a group pass-rate check on the first 50 completions. If one protected group is passing at less than 80 percent of the rate of the highest-passing group, investigate the rubric before scaling. This is the four-fifths rule used in adverse impact monitoring.
  • What to watch for: Completion rates below 60 percent often signal that the task length, instructions, or perceived relevance is driving qualified candidates out before you see their work. AI scoring that operates as a black box without a human review gate before stage decisions is a compliance risk in most jurisdictions. Rubric drift, where reviewers stop using the rubric and revert to gut feel after the first week, makes the assessment data unreliable for comparisons across batches.

Where we talk about this

On AI with Michal live sessions, async assessments come up in two places: the AI in recruiting track covers how to write a rubric that survives reviewer calibration and how to wire assessment data to ATS stage logic without creating silent failures. The sourcing automation track covers the integration side: webhooks, score field mapping, and dead-letter handling when a submission arrives but the payload does not write correctly. If you want the room conversation with other practitioners running real pipelines, start at Workshops and bring your current assessment setup and your ATS field mapping.

Around the web (opinions and rabbit holes)

Third-party creators move fast. Treat these as starting points, not endorsements, and double-check anything before you wire candidate data.

YouTube

Reddit

Quora

Async assessment versus live skills interview

DimensionAsync assessmentLive skills interview
Scheduling dependencyNone for candidateRequires calendar alignment
Rubric requirementMandatory before launchBeneficial but often skipped
Reviewer throughputBatch review possibleOne-to-one time commitment
Candidate experience signalCompletion rate is visibleDrop-off is invisible
Adverse impact monitoringStraightforward from stored scoresRequires structured note system
AI scoring feasibilityHigh for structured outputsLow for conversational nuance

Related on this site

Frequently asked questions

What types of assessments work well in async platforms?
Skills tests with objectively scorable outputs (coding exercises, written case responses, work samples) travel best through async platforms because reviewers can calibrate criteria before seeing a single result. Situational judgment questions also work well when the scenario and answer rubric are pre-validated. Conversational or culture-fit probes work less reliably because unstructured open-ended video responses are hard to score consistently without a structured rubric. One-way video interview responses and async screening questions are adjacent but distinct: assessments produce a scorable artefact, while screening responses confirm basic eligibility. Pair each assessment type with a written scoring guide before sending it to candidates so reviewers are not making rubric decisions while scoring.
How does an async assessment platform connect to an ATS?
Most platforms expose webhooks or a native integration that fires when a candidate submits or times out. The score, completion timestamp, and any structured fields are written back to the ATS candidate record automatically, or pushed to a middleware router. The integration quality matters more than the platform logo: check whether the field mapping is configurable, whether failed submissions land in a dead-letter queue, and whether the score field is human-readable in the ATS stage view without a separate login. Read ATS API integration before configuring the connection, and test with a dummy candidate before your first live req. Silent partial completions, where the payload arrives but the ATS field stays blank, are the most common failure mode in the first week.
What bias and compliance risks apply to scored assessments?
Any scored output that influences a shortlist or stage advance is subject to adverse impact monitoring under most employment regulations. Run a AI bias audit before deploying AI-scored assessments at scale, and track group pass rates by protected characteristic even if the platform does not surface them natively. Under GDPR and similar frameworks, candidates have the right to understand how automated scores were used in a decision. Document which model version produced each score, what rubric it applied, and whether a human reviewed the result before it changed the candidate stage. Tools that claim "AI scoring" without exposing the scoring model or rubric to your team are governance risks. Request a data processing agreement and a subprocessor list before going live.
When should async assessments replace a phone screen?
Replace the phone screen when the assessment produces structured data the phone screen would not, specifically a scored artefact tied to defined job criteria. High-volume pipelines for standardised roles (customer support, entry-level engineering, data entry) often gain accuracy and speed by moving an async test before the recruiter call, because the call then focuses on questions the test cannot answer: motivation, compensation, start date, and culture context. Avoid replacing the phone screen when the role requires language, tone, or communication nuance that a written or video task cannot replicate, or when candidate experience metrics show async drop-off is higher than the scheduling overhead it was meant to fix. Monitor completion rates by channel and by sourcing segment, not just overall.
How do you set time limits and instructions fairly?
Time limits should reflect the minimum time a qualified candidate needs, not the fastest possible completion. Test the assessment internally with current employees in comparable roles before sending it to candidates. If the median internal completion time is 40 minutes, set the limit at 60 to 75 minutes to account for different reading speeds and non-native language users without inflating it so much that time becomes irrelevant as a signal. Provide written instructions and a sample question in a non-scored practice section. State clearly whether AI tools are permitted or prohibited, because candidates who receive conflicting signals from different recruiters introduce an uncontrolled variable into scoring comparisons. Document the instruction text so it is version-controlled alongside the rubric.
What should the human review gate look like for AI-scored results?
The review gate for AI-scored assessment results should be a named step in the hiring workflow where a recruiter or TA lead looks at the score, the evidence behind it (the candidate submission), and the scoring rubric before advancing or rejecting. Do not configure the ATS to auto-advance or auto-reject based on a score threshold without a human touchpoint in between, even if the platform supports it. Log who reviewed the result and when, and keep the reviewer separate from whoever configured the scoring rubric where possible to reduce confirmation bias. For high-stakes or senior roles, require two reviewers to align before stage advance. Join a workshop to see how other teams structure this gate across different volume levels.
How do async assessment results fit into a structured debrief?
Assessment scores should arrive in the debrief as one input alongside interview scorecards and recruiter notes, not as the anchor that all other inputs justify. Share the score and the rubric at the same time so debrief participants can see whether a low score reflects a genuine skill gap or a rubric that does not match the role. For roles with a scorecard, map assessment criteria to scorecard dimensions before the debrief so the conversation stays in the same framework rather than switching between two separate documents. Keep the candidate submission accessible during the debrief for roles where the evidence matters as much as the number. Document the group conclusion separately from the raw score, because the score is one data point and the hiring decision is the outcome.

← Back to AI glossary in practice