Mercor is the highest-paying mainstream AI training platform in 2026. Senior coding contractors pull $70–$110/hr; domain experts in medicine, law, and quantitative finance routinely pass $130/hr. The catch: Mercor's screening is by far the strictest of any platform we track. Roughly 4 in 10 applicants who clear Outlier's coding sample fail Mercor's pre-screen.
Here's what we've learned from ~80 contractors who applied through joblet.ai over the last six months — what works, what doesn't, and what their AI screener seems to weigh most.
What Mercor's pre-screen actually looks at
The Mercor application has three stages: profile, written work sample, and AI-conducted video interview. The third one is what trips people up. Mercor's screener is a structured AI interview — about 15 minutes, 6–8 questions, recorded. Their model scores you on four dimensions:
- Specificity. Vague answers ("I worked on backend systems") score badly. The screener wants concrete numbers, named technologies, and ownership ("I owned the payment retry pipeline at X — handled ~3M txns/day, cut failed-retry latency from 4.2s to 1.1s").
- Failure cases. When asked about a project that didn't work, candidates who blame teammates or external factors score lower than candidates who own the failure and describe what they'd do differently.
- Technical depth on a follow-up. The model probes deeper after every answer. Candidates who can't go three layers deep on their own claimed work get filtered.
- Communication compression. Mercor weights answers between 60–90 seconds higher than rambling 3-minute responses. Get to the answer fast.
The profile signals that move the needle
Before you even reach the AI interview, your profile is filtered. Mercor uses a vector-embedding match against active job listings. Three things matter most:
- GitHub activity in the last 90 days. Empty or stale GitHubs get filtered hard. If yours is sparse, contribute to one well-known open-source project before applying — even a documentation PR moves the score.
- Specific framework experience. Mercor's matcher reads keywords from your profile against keywords in active listings. "Built Python apps" matches less work than "Python · FastAPI · Pydantic · pytest · Postgres · Redis". List the stack.
- Domain tags. If you have legal, medical, or financial domain experience, lead with it. Mercor's highest-paying tracks are domain-expert tracks — you'll get a different (and higher-paying) job pool.
The written sample: what good looks like
After the profile passes, Mercor sends a written work sample — usually a code review, a bug hunt, or a short design exercise. Allotted time: 60–90 minutes. Most rejections happen here, not at the AI interview.
The pattern in samples that pass:
- Read everything twice before writing. 80% of rejected samples in the data we've seen miss a stated constraint in the prompt. Re-reading takes 2 minutes; missing a constraint is an automatic fail.
- Write the failure cases first. If the prompt is "review this code", lead with the bugs. If it's "design this system", lead with the failure modes. Mercor's reviewers (human + AI) score "what could go wrong" thinking very high.
- Keep prose tight. 600–900 words is the sweet spot. Longer than 1,200 words almost always reads as padding.
The AI interview: what to actually do
Mercor's AI interview catches more candidates off-guard than any other step. It feels like a chat with a slightly stilted human. It's not — it's a model that's specifically trained to probe. Treat it accordingly.
Have a 60-second STAR answer ready for these:
- "Walk me through a technical project where you owned the design end-to-end."
- "Tell me about a time something you built broke in production."
- "Describe a technical disagreement you had with a teammate. How did it resolve?"
- "What's the most performance-critical thing you've optimized? By how much?"
- "What's a recent technology you taught yourself and how did you learn it?"
Practice each in front of a mirror or recording app. Land the answer at 60–80 seconds.
When the model probes, don't dodge.
Common pattern: candidate gives a great 60-second answer. Model asks "How did you decide on that approach over alternative X?" Candidate says something like "I think it was the right call given the constraints." Score drops.
The model wants you to compare — explicitly contrast the chosen approach against at least two alternatives, with reasoning. If you can't do that, the model concludes you didn't actually own the decision.
Common rejection reasons (what we've seen)
- Generic profile keywords. "Backend developer" with no specifics. "Familiar with Python." Filter your profile against an actual Mercor job listing — every listed skill should appear on your profile if it applies.
- Missing constraint in written sample. The prompt says "limit changes to
auth.py"; candidate touches three other files. Auto-reject. - Over-rambling in AI interview. Mercor's model has explicit timing penalties. Aim for 60–90 seconds per answer.
- Vagueness on follow-ups. "I'd have to look at it" or "It depends on the situation" without follow-through fails on the depth check.
- No GitHub or stale GitHub. 90-day window matters. If you've been heads-down at work, push something — even your dotfiles repo or a small CLI tool.
How long does the whole thing take?
From application to first paycheck, the realistic timeline:
- Profile review: 24–72 hours (automated).
- Written sample: Sent within 48 hours of profile approval. You have 7 days to complete (most people do it in 2).
- Sample review: 3–7 days.
- AI interview invitation: Sent within 48 hours of sample approval.
- AI interview: 15 minutes (you book a slot).
- Final review + onboarding: 5–10 days.
- First task assigned: 1–4 days after onboarding.
- First paycheck: Mercor pays bi-weekly. ~14 days after first task.
Best case: ~3 weeks from application to first paycheck. Realistic case: 5–6 weeks. Worst case: 8+ weeks if you get rejected at any stage and have to reapply (Mercor allows reapplication after 90 days).
Bottom line
Mercor pays ~30% more than the major-platform average for a reason: their bar is meaningfully higher. The good news is the bar is mostly about specificity and ownership, not about being a 10x engineer. If you can talk concretely about real work you've shipped, you'll pass. If you talk in abstractions, you won't.
Spend 90 minutes the day before applying re-reading your own LinkedIn and GitHub through the lens of "would a vector embedding match this against a Mercor listing?" Specific verbs, specific stack, specific outcomes. That alone gets most candidates past the profile stage.