Find your job

From AI training contractor to AI safety researcher.

AI training contracting builds skills relevant to AI safety research roles. Here's the realistic path from senior evaluator to research role at frontier labs.

AI training contracting at senior tier builds skills directly relevant to AI safety research: rubric design, hallucination detection, evaluation methodology, model failure analysis. Some senior contractors transition into research roles at frontier labs. Here's the realistic path.

What "AI safety researcher" actually means in 2026

Three role types use the title:

  • Alignment researcher — works on the foundational problem of building models whose goals align with human values. Most theoretical.
  • Evaluation researcher — designs and runs systematic evaluations of model capabilities and risks. Most practical.
  • Interpretability researcher — works on understanding model internals (mechanistic interpretability, attention analysis, etc.). Most technical.

Senior AI training contractors transition most easily into evaluation research. The skills overlap directly: you've spent months identifying failure modes, calibrating against rubrics, and reasoning about model outputs.

The skills you've built that translate

  • Rubric design intuition. You've worked under platform rubrics for thousands of tasks. You know what makes a rubric productive vs counterproductive.
  • Failure mode taxonomy. You've categorized model failures across hundreds of tasks. This is exactly what evaluation researchers do at scale.
  • Calibration discipline. Senior evaluators have demonstrably good judgment under structured criteria. Research roles need this.
  • Domain depth in eval methodology. You understand what works and what doesn't for measuring model quality.

The skills you'll need to add

Most AI training contractors transitioning to research need to add:

  • Statistical reasoning. Standard error bars, hypothesis testing, A/B-style comparisons.
  • Coding fluency for analysis. Python, Pandas, plotting libraries.
  • Writing. Research blog posts, technical reports, paper-quality writing.
  • ML literacy. Understanding of training, fine-tuning, RLHF mechanics — at a level deeper than what's needed to evaluate.
Research salary vs senior evaluatorJunior researcher = $150-200k FTE. Senior evaluator gross = ~$150k.
Open calculator →

The realistic transition path

Phase 1: Senior tier with documented evaluation thinking (months 1–9)

Reach senior tier on Mercor or Outlier in a specialty track. Maintain 0.93+ scores. Begin documenting your evaluation thinking publicly — short blog posts on what you're learning about model failure patterns.

Phase 2: Public AI safety / evaluation work (months 6–18)

Publish substantive thinking on AI safety topics. LessWrong, Alignment Forum, Substack, or your own blog. Volume matters less than depth — 2–3 substantive pieces per quarter is plenty.

Topics that translate from contracting work: hallucination patterns, rubric design lessons, evaluation methodology critique, specific model failure case studies (without violating NDAs).

Phase 3: Apply to research roles (months 12–24)

Anthropic, OpenAI, Google DeepMind, METR, Apollo Research, MIRI, FAR AI, ARC all hire researchers from non-traditional backgrounds. Your contracting work + public writing is the application package.

Hit rate is significantly higher for evaluation research roles than alignment or interpretability research roles, because your contracting work directly demonstrates relevant skills.

Pay difference: contractor vs researcher

Senior AI training contractor at full-time hours (~$150k gross, ~$110k after self-employment tax).

Junior AI safety researcher (Anthropic, OpenAI, METR): $150k–$220k base + equity + benefits = ~$200k+ total comp at FTE level. Plus structural stability.

The compensation upside is meaningful. So is the lifestyle upgrade — research roles include health insurance, retirement, paid time off, and a defined work environment.

What doesn't work

  • Cold-applying with just a contracting resume. Without public writing demonstrating thinking, applications get filtered.
  • Faking research output by listing tutorials completed. Recruiters can tell.
  • Trying to leap straight to alignment or interpretability research. Those have higher technical bars; evaluation is the realistic entry point.
  • Treating it as a 6-month plan. The transition takes 12–24 months minimum.

Bottom line

Senior AI training contracting is a legitimate stepping stone to AI safety research roles, particularly evaluation research. The required additions are statistical reasoning, coding for analysis, technical writing, and deeper ML literacy. The realistic timeline is 12–24 months from senior contractor to research role. Pay upside is meaningful — both monetarily and in career stability.

Find AI training contractsAll open roles · 9 platforms · filter by rate and hours.
Find your job

Frequently asked questions

Can AI training contractors become AI safety researchers?
Yes, particularly into evaluation research roles. The skills overlap directly. Realistic transition takes 12–24 months and requires adding statistical reasoning, technical writing, and deeper ML literacy.
What pay do AI safety researchers earn?
Junior researcher base pay at frontier labs (Anthropic, OpenAI, Google DeepMind) is $150k–$220k plus equity. Total comp typically $200k+ at FTE level, with health insurance, retirement, and paid time off.
Which AI safety research role is easiest to transition into?
Evaluation research. Senior AI training contractors have built the most relevant skills (rubric design, failure mode taxonomy, calibration discipline). Alignment and interpretability research require additional theoretical or technical depth.
Do I need a PhD to become an AI safety researcher?
No. Frontier labs hire from non-traditional backgrounds. Public AI safety writing, evaluation track record, and statistical reasoning matter more than formal credentials. Some senior researchers came from contracting + self-directed study.