AI Model Evaluator Opportunity
iMerit seeks detail-oriented and analytically minded evaluators to assess AI system outputs across different modalities: text, image, video, and multimodal interactions.
Evaluators will assess the accuracy, appropriateness, quality, clarity, and cultural alignment of model outputs against complex guidelines, ensuring that results align with project standards and real-world use cases.
* Assess outputs generated by large language models (LLMs) across multiple modalities.
* Evaluate quality against project-specific criteria such as correctness, coherence, completeness, style, cultural appropriateness, and safety.
* Identify subtle errors, hallucinations, or biases in AI responses.
* Apply domain expertise and logical reasoning to resolve ambiguous or unclear outputs.
* Provide detailed written feedback, tagging, and scoring of outputs to ensure consistency across the evaluation team.
* Escalate unclear cases and contribute to refining evaluation guidelines.
* Collaborate with project managers and quality leads to meet accuracy, reliability, and turnaround benchmarks.
• Analytical mindset and attention to detail.
• Ability to evaluate complex information and make informed decisions.
• Strong understanding of AI systems and their applications.
• Excellent communication and teamwork skills.
• Proficiency in English language.
• Opportunity to work with cutting-edge AI technologies.
• Collaborative and dynamic work environment.
• Professional development and growth opportunities.
• Flexible working hours and remote work options.
• Recognition and rewards for outstanding performance.