Job Title: AI Red Teaming Specialist
We are seeking highly analytical and detail-oriented professionals to rigorously test and evaluate AI-generated content.
The ideal candidate will help us identify vulnerabilities, assess risks, and ensure compliance with safety, ethical, and quality standards.
Key Responsibilities:
* Conduct red teaming exercises to identify adversarial outputs from large language models.
* Evaluate and stress-test AI prompts across multiple domains to uncover potential failure modes.
* Develop and apply test cases to assess accuracy, bias, toxicity, hallucinations, and misuse potential in AI-generated responses.
* Collaborate with data scientists and safety researchers to report risks and suggest mitigations.
* Perform manual QA and content validation to ensure factual consistency, coherence, and guideline adherence.
* Create evaluation frameworks and scoring rubrics for prompt performance and safety compliance.
* Document findings, edge cases, and vulnerability reports with high clarity and structure.
Requirements:
* Proven experience in AI red teaming, LLM safety testing, or adversarial prompt design.
* Familiarity with prompt engineering, NLP tasks, and ethical considerations in generative AI.
* Strong background in Quality Assurance, content review, or test case development for AI/ML systems.
* Understanding of LLM behavior's, failure modes, and model evaluation metrics.
* Excellent critical thinking, pattern recognition, and analytical writing skills.
* Ability to work independently and meet tight deadlines.
Preferred Qualifications:
* Prior work with teams like Open AI or Google DeepMind.
* Experience in risk assessment, red team security testing, or AI policy & governance.
A background in linguistics, psychology, or computational ethics is a plus.
All details mentioned above are just an example. This does not indicate real company name, exact salary offered etc. All this information should be considered fictional as it was generated randomly according to the input parameters only