We are seeking skilled professionals to fill a vital role in our organization.
Job Description:
The ideal candidate will test and evaluate AI-generated content to identify vulnerabilities and ensure compliance with safety and quality standards. This involves conducting Red Teaming exercises to identify adversarial outputs from large language models, evaluating AI prompts across multiple domains to uncover potential failure modes, and developing test cases to assess accuracy, bias, toxicity, and misuse potential in AI-generated responses.
Key Responsibilities:
* Conduct thorough analysis of AI-generated content to identify potential risks and areas for improvement.
* Develop and implement effective strategies to mitigate identified risks and improve overall quality.
* Collaborate with data scientists to report findings and suggest mitigations.
* Perform manual QA and content validation to ensure factual consistency and guideline adherence.
Requirements:
* Proven experience in AI red teaming, LLM safety testing, or adversarial prompt design.
* Familiarity with prompt engineering, NLP tasks, and ethical considerations in generative AI.
* Strong background in Quality Assurance, content review, or test case development for AI/ML systems.
The successful candidate will possess a strong analytical mindset, excellent writing skills, and the ability to work independently.