 
        
        AI Red Teaming Specialist
We are seeking an analytical professional with experience in AI safety testing and prompt evaluation.
Main Responsibilities:
 * Conduct red teaming exercises to identify vulnerabilities in large language models (LLMs).
 * Evaluate and stress-test AI prompts across multiple domains to uncover potential failure modes.
 * Develop test cases to assess accuracy, bias, toxicity, hallucinations, and misuse potential in AI-generated responses.
 * Collaborate with data scientists and safety researchers to report risks and suggest mitigations.
 * Perform manual QA and content validation across model versions to ensure factual consistency and coherence.
 * Create evaluation frameworks and scoring rubrics for prompt performance and safety compliance.
 * Document findings and vulnerability reports with high clarity and structure.
Required Skills and Qualifications:
 * Proven experience in AI red teaming or LLM safety testing.
 * Familiarity with prompt engineering, NLP tasks, and ethical considerations in generative AI.
 * Strong background in quality assurance, content review, or test case development for AI/ML systems.
 * Understanding of LLM behavior, failure modes, and model evaluation metrics.
 * Excellent critical thinking, pattern recognition, and analytical writing skills.
 * Ability to work independently and meet tight deadlines.
Benefits:
 * Prior work with teams like Open AI, Google Deepmind, or other LLM safety initiatives is a plus.
 * Experience in risk assessment, red team security testing, or AI policy & governance is also beneficial.
Others:
 * A background in linguistics, psychology, or computational ethics is a valuable asset.