Job Title: QA/Red Teaming Expert
We seek analytical professionals to test and evaluate AI-generated content, identifying vulnerabilities, assessing risks, and ensuring compliance with safety, ethical, and quality standards.
* Key Responsibilities:
* Conduct Red Teaming exercises to identify adversarial outputs from large language models (LLMs).
* Evaluate and stress-test AI prompts across multiple domains to uncover potential failure modes.
* Develop and apply test cases to assess accuracy, bias, toxicity, hallucinations, and misuse potential in AI-generated responses.
* Collaborate with data scientists, safety researchers, and prompt engineers to report risks and suggest mitigations.
* Perform manual QA and content validation across model versions, ensuring factual consistency, coherence, and guideline adherence.
* Create evaluation frameworks and scoring rubrics for prompt performance and safety compliance.
* Document findings, edge cases, and vulnerability reports with high clarity and structure.
Required Skills and Qualifications:
* Proven Experience:
* Ai red teaming, LLM safety testing, or adversarial prompt design experience.
* Familiarity with prompt engineering, NLP tasks, and ethical considerations in generative AI.
* Strong background in Quality Assurance, content review, or test case development for AI/ML systems.
* Understanding of LLM behavior, failure modes, and model evaluation metrics.
* Desirable Qualifications:
* Prior work with teams like Open AI, Anthropic, Google DeepMind, or other LLM safety initiatives.
* Experience in risk assessment, red team security testing, or AI policy & governance.
About the Role:
This is an opportunity to work on evaluating and improving the safety and quality of AI-generated content. If you have a passion for critical thinking, pattern recognition, and analytical writing, we encourage you to apply.