 
        
        Job Opportunity
We are looking for analytical professionals with experience in Red Teaming, prompt evaluation, and AI/LLM quality assurance.
 * Conduct Red Teaming exercises to identify harmful or unsafe outputs from large language models.
 * Evaluate and stress-test AI prompts across multiple domains to uncover potential failure modes.
 * Develop test cases to assess accuracy, bias, toxicity, hallucinations, and misuse potential in AI-generated responses.
Key Responsibilities:
 * Collaborate with data scientists and safety researchers to report risks and suggest mitigations.
 * Perform manual QA and content validation across model versions, ensuring factual consistency and guideline adherence.
Requirements:
 * Prior experience in AI red teaming, LLM safety testing, or adversarial prompt design.
 * Familiarity with prompt engineering, NLP tasks, and ethical considerations in generative AI.
Preferred Qualifications:
 * Prior work with teams like OpenAI, Anthropic, or Google DeepMind.
 * Experience in risk assessment, red team security testing, or AI policy & governance.
About the Role:
 * This role is ideal for individuals who enjoy solving complex problems and working with diverse teams.
 * The successful candidate will have a strong understanding of AI safety principles and be able to apply them in real-world scenarios.