.webp)
.webp)
The demand for domain-specific human experts — professionals who can label, evaluate, grade, and red-team AI outputs — has grown exponentially as AI labs scale their post-training operations. Mercor has positioned itself at the center of this demand. What started as an AI-powered interviewing platform has evolved into one of the most significant talent pipelines serving frontier AI labs. The company’s trajectory reveals important lessons about how the human data market is evolving and what it takes to supply the talent that modern AI development requires.
Mercor’s origin is instructive. The company initially built technology to automate and improve the job interview process using AI. But the founders quickly realized that the same infrastructure — the ability to assess domain expertise, match professionals to specific task requirements, and manage distributed talent at scale — was exactly what AI labs needed to source human data contractors.
The pivot was well-timed. As AI labs began scaling reinforcement learning and RLHF pipelines in 2023 and 2024, the demand for expert contractors exploded. Labs needed people who could not just label data but write domain-specific grading rubrics, design evaluation tasks, produce expert-level solutions, and provide nuanced preference judgments. The talent pool for this work does not exist on traditional crowdsourcing platforms. It exists in professional networks, academic communities, and specialized industries.
Mercor’s AI-powered vetting process gave it a structural advantage. The platform could assess a candidate’s domain expertise, communication skills, and task reliability far more efficiently than manual screening, allowing Mercor to build a qualified expert network at a pace that traditional talent sourcing could not match.
Understanding Mercor’s role requires distinguishing between different types of human data work. Not all annotation is the same, and Mercor’s value proposition is tied to the high end of the spectrum.
At the basic level, annotation involves tasks like image classification, entity tagging, or sentiment labeling. These can be performed by general-purpose annotators with brief training. Mercor does not primarily compete here. Where Mercor excels is in sourcing experts for tasks requiring genuine domain knowledge: RLHF preference judgments in specialized fields, expert evaluation of model outputs, rubric design for complex task grading, and solution authoring for RL environments. These tasks cannot be performed well by generalists, and the difference between low-skill and high-skill annotation directly impacts model quality.
Mercor also produces significant volumes of grading rubrics. These rubrics define how model outputs should be evaluated across different domains and tasks. Currently, most rubrics are written by humans, though some companies are experimenting with model-generated rubrics. The quality of these rubrics is critical because they serve as the reward signal for reinforcement learning. A poorly designed rubric can teach a model to optimize for the wrong objective.
The challenge Mercor addresses is fundamentally one of matching at scale. AI labs know what kind of experts they need. The experts exist somewhere in the world. But connecting the two efficiently, at the scale and speed that modern AI development demands, is an operational problem that most organizations cannot solve internally. Even teams that understand how to recruit human experts for specialized AI tasks often lack the infrastructure to do it at the required volume.
Consider the numbers. A single RLHF campaign might require 200 domain experts working for three months. Those experts need to be sourced, vetted, onboarded, trained on specific task guidelines, and managed through the production process. A lab might run five such campaigns simultaneously across different domains. The logistical complexity is enormous.
Mercor’s platform handles much of this complexity through pre-qualified talent pools, automated matching algorithms that pair experts to task requirements, and workflow tooling that manages the engagement lifecycle. This allows labs to ramp up specialized teams in days rather than weeks.
Mercor competes with other talent marketplaces, most notably Surge and Handshake. Surge is the more established and larger player, with revenue that industry sources estimate to be approaching $1 billion ARR. Mercor’s technology-driven approach to talent matching gives it advantages in sourcing speed and matching precision, though Surge’s scale provides its own competitive moat.
Revenue for these firms is concentrated around western labs. Anthropic, OpenAI, and Google account for a significant portion of the market’s total spending on human data. However, some providers including Surge are reported to also serve international clients, including Chinese labs like Moonshot and Z.ai, where access to RL environments and expert annotation has contributed to capability improvements.
The market is large enough to support multiple providers. AI labs increasingly work with several talent sources simultaneously, both to reduce vendor dependency and to access different pools of expertise. This multi-vendor approach benefits the entire ecosystem, though it means competitive advantage is built on talent quality and operational efficiency rather than exclusivity.
An important distinction in the market is between talent marketplaces like Mercor and full-service managed providers. Talent marketplaces supply the people. Full-service providers supply the people, the workflows, the QC infrastructure, and the project management.
Careerflow operates as a full-service managed provider. Rather than supplying individual contractors for labs to integrate into their own pipelines, Careerflow handles the entire data production process end-to-end: scoping data needs, expert sourcing and vetting, human-powered data production, and multi-layered review before delivering model-ready data. This distinction matters for teams that lack internal annotation infrastructure or prefer to outsource operational complexity.
The choice between a talent marketplace and a managed provider depends on internal capacity. Teams with mature annotation operations, established tooling, and experienced project managers may prefer the flexibility of a marketplace. Teams that want turnkey data delivery without building internal infrastructure may be better served by a managed approach. Understanding this distinction is important when evaluating human data partners.
Several trends are visible in Mercor’s trajectory. First, the market is shifting decisively toward expertise. The era when labs could train models primarily on commodity-labeled data is ending. Post-training — RLHF, DPO, red-teaming, evaluation — requires human judgment from people who genuinely understand the domain.
Second, demand is diversifying beyond coding. While software engineering remains the highest-demand domain, spending on non-technical domains like photography, music, design, healthcare, and finance has increased substantially. Labs are expanding the breadth of their post-training to improve model utility across a wider range of tasks. The economics of this expansion are reshaping how labs allocate their data budgets.
Third, the boundary between talent marketplace and managed service is blurring. Mercor started as a pure talent connector but has increasingly moved toward offering more managed services. This convergence is happening across the market: annotation companies are building talent networks, and talent marketplaces are building annotation workflows.
Mercor’s growth offers an important lesson for AI leaders: the quality of your training data is ultimately a function of the quality of the people producing it. Investing in expert talent — whether through a marketplace like Mercor, a full-service managed provider, or an internal team — is not a cost to minimize. It is a direct investment in model performance.
The companies that treat human data talent as a strategic asset, rather than a line item in their annotation budget, will build better models. Mercor’s rapid growth is evidence that the market is beginning to understand this. The question for each AI team is whether they are acting on that understanding quickly enough.
Sign up now to access Careerflow’s powerful suite of AI tools and take the first step toward landing your dream job.