At the forefront of innovation, our team specializes in evaluating the performance and suitability of complex AI models. This involves assessing model outputs across multiple formats to ensure accuracy, relevance, and coherence.
* Analysts review outputs generated by large language models (LLMs) and vision models (LVMs) to evaluate their quality against project-specific criteria.
* Evaluate model outputs for correctness, completeness, style, cultural appropriateness, and safety.
* Identify subtle errors or biases in AI responses and provide detailed written feedback.
* Apply domain expertise to resolve ambiguous or unclear outputs and escalate unclear cases as needed.
* Contribute to refining evaluation guidelines and ensuring consistency across the evaluation team.
* Bachelor's degree or equivalent educational qualification in a relevant field.
* 1+ years of experience in data annotation, LLM evaluation, content moderation, or related AI/ML domains.
* Experience working with data annotation tools and software platforms is essential.