 
        
        Job Overview
We are seeking analytical professionals to rigorously test and evaluate AI-generated content.
The ideal candidate will help us identify vulnerabilities, assess risks, and ensure compliance with safety, ethical, and quality standards.
-----------------------------------
Key Responsibilities:
 * Conduct exercises to identify adversarial or unsafe outputs from large language models.
 * Evaluate and stress-test AI prompts across multiple domains to uncover potential failure modes.
 * Develop and apply test cases to assess accuracy, bias, toxicity, hallucinations, and misuse potential in AI-generated responses.
 * Collaborate with data scientists and safety researchers to report risks and suggest mitigations.
 * Perform manual QA and content validation across model versions, ensuring factual consistency and coherence.
 * Create evaluation frameworks and scoring rubrics for prompt performance and safety compliance.
 * Document findings, edge cases, and vulnerability reports with high clarity and structure.
-----------------------------------
Requirements:
 * Proven experience in AI testing, LLM safety testing, or adversarial prompt design.
 * Familiarity with NLP tasks and ethical considerations in generative AI.
 * Strong background in Quality Assurance, content review, or test case development for AI/ML systems.
 * Understanding of LLM behaviors and model evaluation metrics.
 * Excellent critical thinking and analytical writing skills.
 * Ability to work independently and meet deadlines.