Security Quality Assurance | $111/hr Remote

Security Quality Assurance | $111/hr Remote

Posted 3 days ago by Crossing Hurdles

Negotiable
Undetermined
Remote
United Kingdom

Summary: The role of AI Red-Teamer involves conducting adversarial AI testing to enhance the security and robustness of AI models. Candidates will engage in red-teaming activities, generating human data, and documenting findings to support various projects. The position offers flexible hours and can be performed remotely. A strong background in AI and cybersecurity is essential for success in this role.

Key Responsibilities:

  • Red-team AI models and agents by crafting jailbreaks, prompt injections, misuse cases, and exploit scenarios
  • Generate high-quality human data: annotate AI failures, classify vulnerabilities, and flag systemic risks
  • Apply structured approaches using taxonomies, benchmarks, and playbooks to maintain consistency in testing
  • Document findings comprehensively to produce reproducible reports, datasets, and attack cases
  • Flexibly support multiple projects including LLM jailbreaks and socio-technical abuse testing across different customers

Key Skills:

  • Prior red-teaming experience, such as AI adversarial work, cybersecurity, or socio-technical probing OR a strong AI background that supports rapid learning
  • Expertise in adversarial machine learning, including jailbreak datasets, prompt injection, RLHF/DPO attacks, model extraction
  • Cybersecurity skills such as penetration testing, exploit development, reverse engineering
  • Experience with socio-technical risk areas like harassment, disinformation, or abuse analysis
  • Creative probing using psychology, acting, or writing to develop unconventional adversarial methods

Salary (Rate): £111.00/hr

City: undetermined

Country: United Kingdom

Working Arrangements: remote

IR35 Status: undetermined

Seniority Level: undetermined

Industry: IT

Detailed Description From Employer:

Crossing Hurdles is a referral partner. We connect & refer top candidates to our partners working with the world’s leading AI research labs to help build and train cutting-edge AI models.

Position: AI Red-Teamer — Adversarial AI Testing (Advanced)

Type: Hourly Contract

Compensation: $54–$111/hour

Location: Remote

Duration: Full-time or part-time

Commitment: 10-40 hours/week, flexible and asynchronous

Key Responsibilities (Training support will be provided)

  • Red-team AI models and agents by crafting jailbreaks, prompt injections, misuse cases, and exploit scenarios
  • Generate high-quality human data: annotate AI failures, classify vulnerabilities, and flag systemic risks
  • Apply structured approaches using taxonomies, benchmarks, and playbooks to maintain consistency in testing
  • Document findings comprehensively to produce reproducible reports, datasets, and attack cases
  • Flexibly support multiple projects including LLM jailbreaks and socio-technical abuse testing across different customers

Required Qualifications

  • Prior red-teaming experience, such as AI adversarial work, cybersecurity, or socio-technical probing OR a strong AI background that supports rapid learning
  • Expertise in adversarial machine learning, including jailbreak datasets, prompt injection, RLHF/DPO attacks, model extraction
  • Cybersecurity skills such as penetration testing, exploit development, reverse engineering
  • Experience with socio-technical risk areas like harassment, disinformation, or abuse analysis
  • Creative probing using psychology, acting, or writing to develop unconventional adversarial methods

Application process: (Takes 20 min)

  • Upload resume
  • AI interview based on your resume (15 min)
  • Submit form