We are seeking analytical professionals to evaluate AI-generated content and ensure compliance with quality standards.
- Conduct Red Teaming exercises to identify vulnerabilities in large language models (LLMs).
- Evaluate AI prompts across multiple domains to uncover potential failure modes.
Key Responsibilities :
Develop test cases for accuracy, bias, toxicity, hallucinations, and misuse potential in AI responses.Collaborate with data scientists to report risks and suggest mitigations.Job Requirements :
Proven experience in AI red teaming or LLM safety testing.Familiarity with NLP tasks and ethical considerations in generative AI.Preferred Qualifications :
Experience in risk assessment or AI policy & governance.