Human Data Powers AI Safety

Diving deeper into

Jemma White, COO of Prolific, on why humans ensure AI safety

Interview
the next frontier is humanity.
Analyzed 4 sources

This marks a shift from training models to know facts, toward training them to act like acceptable people. For vendors in human data, the scarce input is no longer just a PhD who can judge a hard answer, it is a large, reliable pool of people whose language, culture, temperament, and behavior can be measured and called on quickly for red teaming, product testing, and safety checks.

  • The workflow is becoming more concrete and more productized. Teams set filters for traits, language, country, experience, or behavior, send tasks through API or self serve tools, and get responses in minutes or hours. That favors platforms with deep participant profiles and fast matching, not just big contractor rosters.
  • This is a real competitive turn in the market. Scale grew up on broad labeling and RLHF infrastructure, while Handshake built supply around verified student and expert talent. Prolific is pushing one step further, toward profiling ordinary human differences that matter when a model must sound calm, culturally aware, and trustworthy.
  • The demand base is widening beyond frontier labs. Prolific serves academics, frontier labs, AI B2B companies, and enterprises, and it reports rising use for safety evals, product testing, localization, and multimodal research. That matters because humanity data is needed not only to train base models, but to monitor live AI products after launch.

Over the next few years, the winning human data platforms will look less like labor brokers and more like measurement systems for human judgment. As regulation, global distribution, and always on model monitoring expand, the most valuable asset will be a fast, auditable way to sample real human behavior at scale.