We are seeking a detail-oriented professional with hands-on experience in testing and evaluating AI-generated content to identify vulnerabilities, assess risks, and ensure compliance with safety, ethical, and quality standards.
This role involves conducting Red Teaming exercises to identify adversarial outputs from large language models (LLMs), as well as evaluating and stress-testing AI prompts across multiple domains to uncover potential failure modes.
Key Responsibilities :
- Developing and applying test cases to assess accuracy, bias, toxicity, hallucinations, and misuse potential in AI-generated responses
- Collaborating with data scientists, safety researchers, and prompt engineers to report risks and suggest mitigations
- Performing manual QA and content validation across model versions, ensuring factual consistency, coherence, and guideline adherence
- Creating evaluation frameworks and scoring rubrics for prompt performance and safety compliance
- Documenting findings, edge cases, and vulnerability reports with high clarity and structure