AI Safety Specialist
We are seeking detail-oriented professionals to test and evaluate AI-generated content.
- Conduct thorough evaluations of large language models (LLMs) to identify potential security risks.
- Evaluate AI prompts across multiple domains to uncover possible weaknesses.
- Develop robust test cases to assess accuracy, bias, toxicity, hallucinations, and misuse potential in AI-generated responses.
- Collaborate with data scientists and safety researchers to report findings and suggest mitigations.
- Perform manual quality assurance and content validation ensuring factual consistency, coherence, and adherence to guidelines.
- Create evaluation frameworks and scoring rubrics for prompt performance and safety compliance.
Requirements :
Proven experience in AI safety testing or adversarial prompt design.Familiarity with prompt engineering, NLP tasks, and ethical considerations in generative AI.Strong background in Quality Assurance, content review, or test case development for AI / ML systems.Understanding of LLM behavior's, failure modes, and model evaluation metrics.Excellent critical thinking, pattern recognition, and analytical writing skills.Prior work with teams focused on LLM safety initiatives is preferred.