AI Red Team Leader

AI Red Team Leader

Full-Time No home office possible
ActiveFence

ActiveFence is seeking an experienced and detail-oriented Red Teaming Team Lead to oversee complex research and delivery efforts focused on identifying and mitigating risks in Generative AI systems. In this role, you will lead a multidisciplinary team conducting adversarial testing, risk evaluations, and data-driven analyses that strengthen AI model safety and integrity.

You will be responsible for ensuring high-quality project delivery, from methodology design and execution to client communication and final approval of deliverables. This position combines hands-on red teaming expertise with operational leadership, strategic thinking, and client-facing collaboration.

Key Responsibilities

  • Operational and Quality Leadership
  • Oversee the production of datasets, reports, and analyses related to AI safety and red teaming activities.
  • Review and approve deliverables to ensure they meet quality, methodological, and ethical standards.
  • Deliver final outputs to clients following approval and provide actionable insights that address key risks and vulnerabilities.
  • Offer ongoing structured feedback on the quality of deliverables and the efficiency of team workflows, driving continuous improvement.
  • Methodology and Research Development
  • Design and refine red teaming methodologies for new Responsible AI projects.
  • Guide the development of adversarial testing strategies that target potential weaknesses in models across text, image, and multimodal systems.
  • Support research initiatives aimed at identifying and mitigating emerging risks in Generative AI applications.
  • Client Engagement and Collaboration
  • Attend client meetings to address broader methodological or operational questions.
  • Represent the red teaming function in cross-departmental collaboration with other ActiveFence teams.

Requirements

Must Have

  • Proven background in red teaming, AI safety research, or Responsible AI operations.
  • Demonstrated experience managing complex projects or teams in a technical or analytical environment.
  • Strong understanding of adversarial testing methods and model evaluation.
  • Excellent communication skills in English, both written and verbal.
  • Exceptional organizational ability and attention to detail, with experience balancing multiple priorities.
  • Confidence in client-facing environments, including presenting deliverables and addressing high-level questions.

Nice to Have

  • Advanced academic or research background in AI, computational social science, or information integrity.
  • Experience authoring or co-authoring publications, white papers, or reports in the fields of AI Safety, Responsible AI, or AI Ethics.
  • Engagement in professional or academic communities related to Responsible AI, trust and safety, or machine learning security.
  • Participation in industry or academic conferences.
  • Familiarity with developing or reviewing evaluation frameworks, benchmarking tools, or adversarial datasets for model safety testing.
  • Proven ability to mentor researchers and foster professional development within technical teams.
  • A proactive, research-driven mindset and a passion for ensuring safe, transparent, and ethical AI deployment.

Seniority level

  • Mid-Senior level

Employment type

  • Full-time

Job function

  • Project Management, Writing/Editing, and Analyst

Industries

  • Software Development, Technology, Information and Media, and Business Consulting and Services

#J-18808-Ljbffr

ActiveFence

Contact Detail:

ActiveFence Recruiting Team

Land your dream job quicker with Premium

You’re marked as a top applicant with our partner companies
Individual CV and cover letter feedback including tailoring to specific job roles
Be among the first applications for new jobs with our AI application
1:1 support and career advice from our career coaches
Go Premium

Money-back if you don't land a job in 6-months

>