Job Title:
Linguistic Content Specialist
-----------------------------------
About the Job:
We are seeking a highly analytical and detail-oriented Linguistic Content Specialist to join our team. The ideal candidate will help us rigorously test and evaluate AI-generated content to identify vulnerabilities, assess risks, and ensure compliance with safety, ethical, and quality standards.
Responsibilities:
1. Conduct in-depth analysis of large language models (LLMs) to identify adversarial, harmful, or unsafe outputs.
2. Evaluate and stress-test AI prompts across multiple domains to uncover potential failure modes.
3. Develop and apply test cases to assess accuracy, bias, toxicity, hallucinations, and misuse potential in AI-generated responses.
4. Collaborate with data scientists, safety researchers, and prompt engineers to report risks and suggest mitigations.
5. Perform manual QA and content validation across model versions, ensuring factual consistency, coherence, and guideline adherence.
6. Create evaluation frameworks and scoring rubrics for prompt performance and safety compliance.
7. Document findings, edge cases, and vulnerability reports with high clarity and structure.
Requirements:
1. Proven experience in AI red teaming, LLM safety testing, or adversarial prompt design.
2. Familiarity with prompt engineering, NLP tasks, and ethical considerations in generative AI.
3. Strong background in Quality Assurance, content review, or test case development for AI/ML systems.
4. Understanding of LLM behaviors, failure modes, and model evaluation metrics.
5. Excellent critical thinking, pattern recognition, and analytical writing skills.
6. Ability to work independently, follow detailed evaluation protocols, and meet tight deadlines.
Benefits:
This role offers the opportunity to work on cutting-edge technology and make a significant impact on the development of AI-generated content. You will be part of a dynamic team that values collaboration, innovation, and continuous learning.
Preferred Qualifications:
1. Prior work with teams like OpenAI, Anthropic, Google DeepMind, or other LLM safety initiatives.
2. Experience in risk assessment, red team security testing, or AI policy & governance.
3. Background in linguistics, psychology, or computational ethics is a plus.