In this position, you will contribute to the development of innovative AI technology solutions. Innodata's mission is to provide top-tier services for leading companies across various industries.
Our team consists of 5000 professionals working in the US, Canada, UK, Philippines, India, Sri Lanka, Israel, and Germany. We are seeking experts to evaluate AI-generated content, ensuring safety and quality standards.
You will be responsible for :
- Conducting Red Teaming exercises to identify adversarial outputs from large language models (LLMs).
- Evaluating and stress-testing AI prompts across multiple domains to uncover potential failure modes.
- Developing and applying test cases to assess accuracy, bias, toxicity, hallucinations, and misuse potential in AI-generated responses.
- Collaborating with data scientists and safety researchers to report risks and suggest mitigations.
- Performing manual QA and content validation across model versions, ensuring factual consistency, coherence, and guideline adherence.
- Creating evaluation frameworks and scoring rubrics for prompt performance and safety compliance.
- Documenting findings, edge cases, and vulnerability reports with high clarity and structure.
The ideal candidate should have :
Proven experience in AI red teaming, LLM safety testing, or adversarial prompt design.Familiarity with prompt engineering, NLP tasks, and ethical considerations in generative AI.A strong background in Quality Assurance, content review, or test case development for AI / ML systems.An understanding of LLM behaviors, failure modes, and model evaluation metrics.Excellent critical thinking, pattern recognition, and analytical writing skills.Ability to work independently, follow detailed evaluation protocols, and meet tight deadlines.Please complete the Assessment Test at iCAP platform.