 
        Job Title: Red Teaming, Prompt Evaluation or AI/LLM Quality Assurance ExpertConduct rigorous Red Teaming exercises to identify adversarial, harmful or unsafe outputs from large language models (LLMs).Evaluate and stress-test AI prompts across multiple domains (e.g. finance, healthcare, security) to uncover potential failure modes.Develop and apply test cases to assess accuracy, bias, toxicity, hallucinations and misuse potential in AI-generated responses.Collaborate with data scientists, safety researchers and prompt engineers to report risks and suggest mitigations.Perform manual QA and content validation across model versions ensuring factual consistency, coherence and guideline adherence.Create evaluation frameworks and scoring rubrics for prompt performance and safety compliance.Document findings, edge cases and vulnerability reports with high clarity and structure.Key Requirements:Proven experience in AI red teaming, LLM safety testing or adversarial prompt design.Familiarity with prompt engineering, NLP tasks and ethical considerations in generative AI.Strong background in Quality Assurance, content review or test case development for AI/ML systems.Understanding of LLM behavior's, failure modes and model evaluation metrics.Excellent critical thinking, pattern recognition and analytical writing skills.Ability to work independently, follow detailed evaluation protocols and meet tight deadlines.