 
        
        Key Performance Indicators:
 * Identify vulnerabilities in AI-generated content.
 * Assess risks and ensure compliance with safety, ethical, and quality standards.
We are seeking analytical professionals to test and evaluate AI models. The ideal candidate will have hands-on experience in Red Teaming, Prompt Evaluation, and AI Quality Assurance.
Job Title: Expert in Red Teaming and LLM Safety Testing
Key Responsibilities:
 * Conduct rigorous testing of AI models to identify potential failure modes.
 * Evaluate and stress-test AI prompts across multiple domains.
 * Develop test cases to assess accuracy, bias, toxicity, hallucinations, and misuse potential.
 * Collaborate with data scientists and researchers to report risks and suggest mitigations.
 * Perform manual QA and content validation to ensure factual consistency and coherence.
 * Create evaluation frameworks and scoring rubrics for prompt performance and safety compliance.
Requirements:
 * Proven experience in AI red teaming, LLM safety testing, or adversarial prompt design.
 * Familiarity with prompt engineering, NLP tasks, and ethical considerations in generative AI.
 * Strong background in Quality Assurance, content review, or test case development for AI/ML systems.
 * Understanding of LLM behavior's, failure modes, and model evaluation metrics.
 * Excellent critical thinking, pattern recognition, and analytical writing skills.
Preferred Qualifications:
 * Prior work with teams focused on LLM safety initiatives.
 * Experience in risk assessment, red team security testing, or AI policy & governance.
Background in linguistics, psychology, or computational ethics is a plus.