 
        
        We are seeking analytical professionals with experience in Red Teaming, Prompt Evaluation, and AI Quality Assurance. The ideal candidate will test and evaluate AI-generated content to identify vulnerabilities, assess risks, and ensure compliance with safety standards.
The ideal candidate will:
 * Conduct Red Teaming exercises to identify adversarial outputs from large language models (LLMs).
 * Evaluate and stress-test AI prompts across multiple domains to uncover potential failure modes.
 * Develop and apply test cases to assess accuracy, bias, toxicity, hallucinations, and misuse potential in AI-generated responses.
Requirements:
 * Proven experience in AI red teaming, LLM safety testing, or adversarial prompt design.
 * Familiarity with prompt engineering, NLP tasks, and ethical considerations in generative AI.
 * Strong background in Quality Assurance, content review, or test case development for AI systems.
Preferred Qualifications:
 * Prior work with teams like Open AI, Google DeepMind, or other LLM safety initiatives.
 * Experience in risk assessment, red team security testing, or AI policy & governance.
Key responsibilities include collaboration with data scientists and safety researchers to report risks and suggest mitigations.
Professionals with a strong analytical mindset and hands-on experience in Red Teaming, Prompt Evaluation, or AI Quality Assurance are encouraged to apply.