Negotiable
Undetermined
Remote
United Kingdom
Summary: The role of AI Red-Teamer involves conducting adversarial AI testing to enhance the security and robustness of AI models. Candidates will engage in red-teaming activities, generating human data, and documenting findings to support various projects. The position offers flexible hours and can be performed remotely. A strong background in AI and cybersecurity is essential for success in this role.
Key Responsibilities:
- Red-team AI models and agents by crafting jailbreaks, prompt injections, misuse cases, and exploit scenarios
- Generate high-quality human data: annotate AI failures, classify vulnerabilities, and flag systemic risks
- Apply structured approaches using taxonomies, benchmarks, and playbooks to maintain consistency in testing
- Document findings comprehensively to produce reproducible reports, datasets, and attack cases
- Flexibly support multiple projects including LLM jailbreaks and socio-technical abuse testing across different customers
Key Skills:
- Prior red-teaming experience, such as AI adversarial work, cybersecurity, or socio-technical probing OR a strong AI background that supports rapid learning
- Expertise in adversarial machine learning, including jailbreak datasets, prompt injection, RLHF/DPO attacks, model extraction
- Cybersecurity skills such as penetration testing, exploit development, reverse engineering
- Experience with socio-technical risk areas like harassment, disinformation, or abuse analysis
- Creative probing using psychology, acting, or writing to develop unconventional adversarial methods
Salary (Rate): £111.00/hr
City: undetermined
Country: United Kingdom
Working Arrangements: remote
IR35 Status: undetermined
Seniority Level: undetermined
Industry: IT
Crossing Hurdles is a referral partner. We connect & refer top candidates to our partners working with the world’s leading AI research labs to help build and train cutting-edge AI models.
Position: AI Red-Teamer — Adversarial AI Testing (Advanced)
Type: Hourly Contract
Compensation: $54–$111/hour
Location: Remote
Duration: Full-time or part-time
Commitment: 10-40 hours/week, flexible and asynchronous
Key Responsibilities (Training support will be provided)
- Red-team AI models and agents by crafting jailbreaks, prompt injections, misuse cases, and exploit scenarios
- Generate high-quality human data: annotate AI failures, classify vulnerabilities, and flag systemic risks
- Apply structured approaches using taxonomies, benchmarks, and playbooks to maintain consistency in testing
- Document findings comprehensively to produce reproducible reports, datasets, and attack cases
- Flexibly support multiple projects including LLM jailbreaks and socio-technical abuse testing across different customers
Required Qualifications
- Prior red-teaming experience, such as AI adversarial work, cybersecurity, or socio-technical probing OR a strong AI background that supports rapid learning
- Expertise in adversarial machine learning, including jailbreak datasets, prompt injection, RLHF/DPO attacks, model extraction
- Cybersecurity skills such as penetration testing, exploit development, reverse engineering
- Experience with socio-technical risk areas like harassment, disinformation, or abuse analysis
- Creative probing using psychology, acting, or writing to develop unconventional adversarial methods
Application process: (Takes 20 min)
- Upload resume
- AI interview based on your resume (15 min)
- Submit form