AI Output Evaluation Role – Project Hermes

last updated September 8, 2025 0:26 UTC

CrowdGen

HQ: On-site

  • OFF: Any
  • Full-Time
  • All Other Remote

Join Project Hermes 2.0—a global effort aimed at improving the quality of large language model (LLM) outputs across various fields!

We’re seeking contributors for a dynamic evaluation task. In this role, you’ll compare input-output pairs and assess the AI-generated responses for accuracy, clarity, and relevance, following specific project guidelines. Your feedback will play a key role in enhancing model performance and advancing responsible AI development.

Key Responsibilities:

– Analyze input-output pairs and evaluate the quality and accuracy of AI responses
– Assess content based on clarity, tone, relevance, and how well it aligns with the intended meaning
– Submit structured feedback to help improve the model
– Ensure consistent attention to detail and quality in all assigned tasks

Requirements:

– Native or near-native proficiency in English
– Excellent reading comprehension and writing abilities
– Availability to commit at least 20 hours per week
– A bachelor’s degree or relevant experience in content evaluation, writing, linguistics, or a related area
– Strong attention to detail and ability to follow complex guidelines
– Comfortable reviewing content that may include sensitive or potentially harmful material

Spots are limited—early applicants will have the best chance to secure a position and start earning quickly!

This is a project-based opportunity with CrowdGen. Selected candidates will join the CrowdGen Community as Independent Contractors. If chosen, you’ll receive an email with instructions to set up your account, reset your password, and complete the necessary steps to begin your application.

Make a difference in the future of AI—apply now and contribute from the comfort of your home.

Compensation: $3.67 per task (pay-per-unit, in USD).

Apply info ->

To find out more about this job, please visit this link