Our company is at the forefront of data engineering, serving over 2000 clients and operating in 13 cities worldwide. We leverage AI technology solutions chosen by four out of five of the world's largest tech companies.
By combining advanced machine learning and AI technologies with a global team of experts, we deliver high-quality platforms.
We are seeking highly analytical professionals with hands-on experience in Red Teaming, Prompt Evaluation, and AI / LLM Quality Assurance.
Key Responsibilities :
- Conduct Red Teaming exercises to identify adversarial outputs from large language models.
- Evaluate and stress-test AI prompts across multiple domains.
- Develop test cases to assess accuracy, bias, toxicity, hallucinations, and misuse potential in AI-generated responses.
- Collaborate with data scientists to report risks and suggest mitigations.
- Perform manual QA and content validation across model versions.
- Create evaluation frameworks and scoring rubrics for prompt performance.
Requirements :
Proven experience in AI red teaming or LLM safety testing.Familiarity with prompt engineering and NLP tasks.Strong background in Quality Assurance and test case development.Understanding of LLM behaviors and failure modes.