Red Teaming and AI Safety Specialist
In today's rapidly evolving landscape of artificial intelligence, the need for rigorous testing and evaluation has never been more pressing.
The ideal candidate will possess a deep understanding of adversarial techniques, large language models, and the importance of safety protocols in AI development.
- Conduct thorough assessments of AI-generated content to identify potential vulnerabilities and risks.
- Evaluate and stress-test AI prompts across multiple domains to uncover hidden failure modes.
- Develop and apply test cases to assess accuracy, bias, toxicity, hallucinations, and misuse potential in AI-generated responses.
- Collaborate with data scientists, safety researchers, and prompt engineers to report risks and suggest mitigations.
- Perform manual QA and content validation across model versions, ensuring factual consistency, coherence, and guideline adherence.
- Create evaluation frameworks and scoring rubrics for prompt performance and safety compliance.
Requirements :
Proven experience in AI red teaming, LLM safety testing, or adversarial prompt design.Familiarity with prompt engineering, NLP tasks, and ethical considerations in generative AI.Strong background in Quality Assurance, content review, or test case development for AI / ML systems.Understanding of LLM behaviors, failure modes, and model evaluation metrics.Excellent critical thinking, pattern recognition, and analytical writing skills.Ability to work independently, follow detailed evaluation protocols, and meet tight deadlines.