Job Opportunity:
We're seeking analytical and detail-oriented professionals to evaluate AI-generated content.
Key Responsibilities:
* Conduct red teaming exercises to identify vulnerabilities in large language models.
* Evaluate and stress-test AI prompts across multiple domains.
* Develop test cases to assess accuracy, bias, toxicity, hallucinations, and misuse potential in AI-generated responses.
* Collaborate with data scientists to report risks and suggest mitigations.
* Perform manual QA and content validation across model versions.
Requirements:
* Proven experience in AI red teaming, LLM safety testing, or adversarial prompt design.
* Familiarity with prompt engineering, NLP tasks, and ethical considerations in generative AI.
* Strong background in Quality Assurance, content review, or test case development for AI/ML systems.
* Understanding of LLM behavior's, failure modes, and model evaluation metrics.
* Excellent critical thinking, pattern recognition, and analytical writing skills.