Job Opportunity
We are seeking analytical professionals with hands-on experience in Red Teaming, Prompt Evaluation, and AI/LLM Quality Assurance.
* Conduct rigorous testing of AI-generated content to identify vulnerabilities, assess risks, and ensure compliance with safety and quality standards.
Key Responsibilities:
1. Perform Red Teaming exercises to identify adversarial outputs from large language models.
2. Evaluate AI prompts across multiple domains to uncover potential failure modes.
3. Develop test cases to assess accuracy, bias, toxicity, hallucinations, and misuse potential in AI-generated responses.
4. Collaborate with data scientists and safety researchers to report risks and suggest mitigations.
5. Perform manual QA and content validation across model versions.
6. Create evaluation frameworks and scoring rubrics for prompt performance and safety compliance.
Requirements:
* Proven experience in AI red teaming or LLM safety testing.
* Familiarity with NLP tasks and ethical considerations in generative AI.
* Strong background in Quality Assurance or test case development for AI/ML systems.
* Understanding of LLM behaviors and failure modes.
* Excellent critical thinking and analytical writing skills.
Benefits:
* The opportunity to work on cutting-edge AI projects.
* A collaborative and dynamic work environment.
Once you have completed the assessments, you will be eligible for onboarding. Please take a few minutes to register and complete your profile on our internal job platform, XConnect. We manage all project onboarding, communication, and documentation through this platform.