AI/LLM Quality Assurance Specialist
We seek analytical professionals with hands-on experience in Red Teaming, prompt evaluation, and AI quality assurance.
The ideal candidate will rigorously test and evaluate AI-generated content to identify vulnerabilities, assess risks, and ensure compliance with safety, ethical, and quality standards.
Key Responsibilities:
* Conduct Red Team exercises to identify adversarial outputs from large language models (LLMs).
* Evaluate and stress-test AI prompts across multiple domains to uncover potential failure modes.
* Develop test cases to assess accuracy, bias, toxicity, hallucinations, and misuse potential in AI-generated responses.
* Collaborate with data scientists and researchers to report risks and suggest mitigations.
* Perform manual QA and content validation across model versions, ensuring factual consistency, coherence, and guideline adherence.
Why This Role Matters: Ensuring the reliability and trustworthiness of AI systems is crucial for their adoption in various industries. As an AI/LLM Quality Assurance Specialist, you will play a vital role in shaping the future of AI development.