Track This Job
Add this job to your tracking list to:
- Monitor application status and updates
- Change status (Applied, Interview, Offer, etc.)
- Add personal notes and comments
- Set reminders for follow-ups
- Track your entire application journey
Save This Job
Add this job to your saved collection to:
- Access easily from your saved jobs dashboard
- Review job details later without searching again
- Compare with other saved opportunities
- Keep a collection of interesting positions
- Receive notifications about saved jobs before they expire
AI-Powered Job Summary
Get a concise overview of key job requirements, responsibilities, and qualifications in seconds.
Pro Tip: Use this feature to quickly decide if a job matches your skills before reading the full description.
Why This Role Exists
At Mercor, we believe the safest AI is the one that’s already been attacked — by us. That’s why we’re building a pod of AI Red-Teamers: human data experts who probe AI models with adversarial inputs, surface vulnerabilities, and generate the red-team data that makes AI safer for our customers.
This role may include reviewing AI outputs that touch on sensitive topics such as bias, misinformation, or harmful behaviors. All work is text-based, and participation in higher-sensitivity projects is optional and supported by clear guidelines and wellness resources.
What You’ll Do
- Red-team AI models and agents: jailbreaks, prompt injections, misuse cases, exploits
- Generate high-quality human data: annotate failures, classify vulnerabilities, and flag systemic risks
- Apply structure: follow taxonomies, benchmarks, and playbooks to keep testing consistent Document reproducibly: produce reports, datasets, and attack cases customers can act on
- Flex across projects: support different customers, from LLM jailbreaks to socio-technical abuse testing
- You bring prior red-teaming experience (AI adversarial work, cybersecurity, socio-technical probing)
- You’re curious and adversarial: you instinctively push systems to breaking points
- You’re structured: you use frameworks or benchmarks, not just random hacks
- You’re communicative: you explain risks clearly to technical and non-technical stakeholders
- You’re adaptable: thrive on moving across projects and customers
- Adversarial ML: jailbreak datasets, prompt injection, RLHF/DPO attacks, model extraction
- Cybersecurity: penetration testing, exploit development, reverse engineering
- Socio-technical risk: harassment/disinfo probing, abuse analysis
- Creative probing: psychology, acting, writing for unconventional adversarial thinking
- You uncover vulnerabilities automated tests miss
- You deliver reproducible artifacts that strengthen customer AI systems
- Evaluation coverage expands: more scenarios tested, fewer surprises in production
- Mercor customers trust the safety of their AI because you’ve already probed it like an adversary
- Build experience in human data-driven AI red-teaming at the frontier of safety
- Play a direct role in making AI systems more robust, safe, and trustworthy
Key Skills
Ranked by relevanceReady to apply?
Join Mercor and take your career to the next level!
Application takes less than 5 minutes

