AI Red Teaming and LLM Quality Assurance Expert
We are seeking analytical professionals with hands-on experience in Red Teaming, Prompt Evaluation, and AI/LLM Quality Assurance to test and evaluate AI-generated content for vulnerabilities, risks, and compliance.
Key Responsibilities:
* Conduct Red Teaming exercises to identify harmful outputs from large language models.
* Evaluate AI prompts across multiple domains to uncover potential failure modes.
* Develop and apply test cases to assess accuracy, bias, toxicity, hallucinations, and misuse potential in AI-generated responses.
* Collaborate with data scientists and safety researchers to report risks and suggest mitigations.
* Perform manual QA and content validation across model versions, ensuring factual consistency and guideline adherence.
Requirements:
* Proven experience in AI red teaming, LLM safety testing, or adversarial prompt design.
* Familiarity with prompt engineering, NLP tasks, and ethical considerations in generative AI.
* Strong background in Quality Assurance, content review, or test case development for AI/ML systems.
* Understanding of LLM behaviour's, failure modes, and model evaluation metrics.
* Critical thinking, pattern recognition, and analytical writing skills.