Job Opportunity
We seek analytical professionals with experience in AI/LLM Quality Assurance, Red Teaming, and Prompt Evaluation. Ideal candidates rigorously test and evaluate AI-generated content to identify vulnerabilities and ensure compliance.
Main Responsibilities:
* Conduct Red Teaming exercises to identify harmful outputs from large language models.
* Evaluate and stress-test AI prompts across multiple domains.
* Develop and apply test cases to assess accuracy, bias, and misuse potential in AI-generated responses.
* Collaborate with data scientists and safety researchers to report risks and suggest mitigations.
* Perform manual QA and content validation ensuring factual consistency and guideline adherence.
* Create evaluation frameworks and scoring rubrics for prompt performance and safety compliance.
Requirements:
* Proven experience in AI red teaming, LLM safety testing, or adversarial prompt design.
* Familiarity with prompt engineering, NLP tasks, and ethical considerations in generative AI.
* Strong background in Quality Assurance, content review, or test case development for AI/ML systems.
* Understanding of LLM behaviors and failure modes.
PREFERRED QUALIFICATIONS
* Prior work with teams focused on AI safety research.