 
        
        Job Overview
We are seeking highly analytical and detail-oriented professionals to help us rigorously test and evaluate AI-generated content.
 * Conduct rigorous testing exercises to identify vulnerabilities, assess risks, and ensure compliance with safety and quality standards.
 * Evaluate and stress-test AI prompts across multiple domains (e.g., finance, healthcare, security) to uncover potential failure modes.
 * Develop and apply test cases to assess accuracy, bias, toxicity, hallucinations, and misuse potential in AI-generated responses.
 * Collaborate with data scientists and safety researchers to report risks and suggest mitigations.
 * Perform manual QA and content validation across model versions, ensuring factual consistency and coherence.
Key Responsibilities
 * Proven experience in AI red teaming, LLM safety testing, or adversarial prompt design.
 * Familiarity with prompt engineering, NLP tasks, and ethical considerations in generative AI.
 * Strong background in Quality Assurance, content review, or test case development for AI/ML systems.
 * Understanding of LLM behaviors, failure modes, and model evaluation metrics.
Requirements
 * Critical thinking and pattern recognition skills.
 * Ability to work independently and meet deadlines.