Part V: Evaluation, Reliability, and Governance
Chapter 21

Human Evaluation

"When in doubt, ask a human. The irony is that we build AI to avoid asking humans, then ask humans to evaluate the AI."

A Reluctant Annotator

When Human Evaluation Is Necessary

Human evaluation is essential for assessing aspects of AI behavior that cannot be automatically evaluated. These include subjective qualities like helpfulness, naturalness, and appropriateness, as well as complex tasks where correctness is context-dependent.

The Human Eval Sweet Spot

Human evaluation is necessary when: the evaluation requires judgment that cannot be automated, the stakes are high enough to justify human time, or automated metrics have not yet proven correlation with human preferences. Do not use human evaluation when automated metrics suffice. It is too slow and expensive.

Subjective Quality Assessment

Some qualities resist automatic measurement and require human judgment to assess. Helpfulness measures whether a response actually solves the user's problem, going beyond grammatical correctness to evaluate practical utility. Naturalness assesses whether the response sounds like something a human would say, capturing the fluency and authenticity of the output. Appropriateness evaluates whether the response is suitable for the specific context and user, recognizing that different situations call for different tones and approaches. Coherence checks whether the response is internally consistent and well-organized, without contradictions or confusing structure. Engagement measures whether a user would find the response satisfying to receive, capturing the overall user experience quality.

High-Stakes Decision Validation

When AI outputs affect important decisions, human evaluation validates that the AI is making appropriate judgments. This is especially critical in healthcare, legal, financial, and safety-critical applications.

The Automation Bias Trap

Humans evaluating AI tend to over-rely on AI outputs, especially when the AI presents information with high confidence. Design human eval protocols to force independent judgment, not validation of AI outputs.

Automated Metric Validation

Even when you use automated metrics, human evaluation validates that those metrics correlate with actual quality. Run human eval periodically to check that your automated metrics are still measuring what matters.

Sampling Strategies

Human evaluation is expensive, so efficient sampling is critical. The goal is to maximize the information gained from each human judgment.

Random Sampling

Random sampling gives every example an equal chance of being evaluated. It provides unbiased estimates of overall performance but may under-sample rare but important failure modes.

Stratified Sampling

Stratified sampling divides examples into strata (e.g., by query type, user segment, or confidence level) and samples proportionally or disproportionately from each stratum. This ensures coverage of important segments even when they are rare.

Targeted Sampling

Targeted sampling focuses human evaluation on examples most likely to reveal problems, concentrating effort where it provides the most insight. Low confidence examples are those where the AI system expressed uncertainty, which often indicates areas where the model is on the boundary between correct and incorrect behavior. Adversarial examples are known difficult cases identified through red teaming efforts that probe system weaknesses. Edge cases are unusual inputs that may cause unexpected behavior, representing the tails of the input distribution. Recent failures are cases where automated metrics detected potential issues, warranting human investigation to understand the root cause.

Active Learning for Eval

Active learning strategies focus human evaluation on examples where the model is most uncertain or where human judgment would most improve future performance estimation.

Active Learning Sampling for Evals

Active learning sampling for evals follows a process that concentrates human effort on the examples most likely to improve your eval framework. For each eval cycle, begin by running all examples through automated metrics to establish baseline measurements. Then score each example by eval uncertainty, considering factors such as model disagreement when using an ensemble approach, distance from the decision boundary indicating proximity to the classification threshold, and previously observed error correlation linking to known failure patterns. Select the top N examples with the highest uncertainty scores for human labeling. After obtaining human labels for the selected examples, update your understanding of where automated metrics fail to identify systematic gaps in your automated evaluation approach.

Inter-Annotator Agreement

When multiple humans evaluate the same examples, their agreement (or disagreement) reveals important information about the evaluation task.

Agreement Metrics

Several metrics exist for measuring agreement among annotators, each with different strengths. Cohen's Kappa measures agreement between two annotators while accounting for agreement that would occur by chance alone, providing a more robust measure than simple agreement percentages. Fleiss' Kappa extends this concept to measure agreement among multiple annotators simultaneously, making it suitable for larger annotation teams. Percent agreement is the simplest measure, representing the fraction of cases where annotators agree, though it does not account for chance agreement and can be inflated when the task has naturally high agreement.

Interpreting Disagreement

Low agreement reveals that the eval task is poorly defined, the criteria are ambiguous, or the task is genuinely subjective. High agreement but low accuracy reveals that everyone agrees on the wrong answer.

Agreement vs. Accuracy

High inter-annotator agreement means annotators agree with each other. It does not mean they are correct. Always validate against ground truth, not just against each other.

Resolving Disagreement

When annotators disagree, there are several approaches to resolving the disagreement and improving your evaluation framework. Begin by having annotators discuss ambiguous cases together, sharing their reasoning to understand why interpretations differ. Use the disagreements as an opportunity to refine criteria, clarifying eval guidelines so future annotations are more consistent. For final labels where a decision is needed, use majority voting to aggregate multiple annotations. For truly ambiguous cases that resist consensus, escalate to expert judgment where specialists can make the final determination based on deep domain knowledge.

Cost-Quality Trade-offs

Human evaluation involves trade-offs between cost, speed, and quality. Understanding these trade-offs helps you design efficient eval protocols.

Annotator Expertise Levels

Human evaluators can be categorized into three expertise levels, each with different cost-quality trade-offs. Expert annotators are domain specialists who provide high-quality labels and are expensive but necessary for complex or high-stakes judgments where nuanced understanding is required. Trained annotators are non-experts who have received specific training on the evaluation task, offering moderate cost and quality that works well for well-defined tasks with clear guidelines. Naive annotators are untrained workers who provide quick but lower-quality labels, offering low cost that makes them useful for rough estimates and exploratory analysis where precision is not critical.

Quality Control Methods

Quality control methods help ensure reliable human evaluation results. Gold questions involve including known-answer examples in the annotation batch to check annotator attention and identify those who are not taking the task seriously. Duplicate labeling has multiple annotators label the same examples so you can compare responses and identify inconsistency. Calibration training trains annotators on carefully selected examples before the main evaluation begins, helping them understand the criteria and expected responses. Performance tracking monitors annotator accuracy over time to identify systematic issues or declining quality that might require retraining or replacement.

Cost Optimization Strategies

The 80/20 of Human Evaluation

You often get 80% of the value from human eval with 20% of the cost by: using trained annotators instead of experts for most cases, focusing expert time on ambiguous or high-stakes cases, and using efficient sampling to minimize the total number of judgments needed.

Designing Human Eval Protocols

A well-designed human eval protocol produces reliable, actionable results efficiently.

Questionnaire Design

Design questions that are unambiguous, clear about what is being judged so annotators know exactly what they are evaluating. Questions should be bounded, focused on specific aspects rather than asking for overall quality impressions which are difficult to attribute to any particular dimension. Questions should be easy to answer using simple scales or discrete choices that do not require complex judgment or lengthy explanations. Finally, questions should be consistent, interpreted the same way by different annotators so that comparisons across raters are meaningful.

Bias Mitigation

Human annotators are subject to various biases that can contaminate evaluation results, so design protocols to minimize their impact. Presentation order effects can be mitigated by randomizing the order of examples so that fatigue or anchoring effects do not systematically bias results. Reference hiding ensures that when comparing systems, annotators do not know which is which, preventing preference for familiar brand names or other identification-based biases. Context control provides consistent context for each judgment so that differences in how questions are interpreted do not introduce noise. Attention checks include specific questions throughout the evaluation to verify that annotators remain engaged and are not rushing through carelessly.

Practical Example: Designing a Helpfulness Eval

A conversational AI team building a customer support chatbot needed to evaluate whether their AI responses were helpful to customers. The challenge they faced was that "helpful" is subjective and ambiguous, making it difficult to define measurable criteria. They considered whether to ask users directly, use expert judges, or infer helpfulness from behavior.

The team decided on a multi-method approach combining user surveys for satisfaction measurement, expert judges for quality assessment, and behavioral signals such as resolution rate and escalation rate for validation against actual outcomes. They designed a structured questionnaire with five-point scales measuring four dimensions: relevance, completeness, clarity, and actionability. The design also included attention checks and reference examples to ensure consistent interpretation.

The result was achieving eighty-seven percent inter-annotator agreement. Notably, they found that clarity scores correlated strongly with user satisfaction, while completeness scores did not show a meaningful relationship. The lesson is to break vague concepts into measurable dimensions, recognizing that different dimensions may have different relationships to the overall outcome you care about.

Research Frontier

New research explores "participatory evaluation" where end users themselves evaluate AI systems during natural usage. This reduces the gap between eval conditions and real-world conditions, though it introduces new challenges in standardization and bias control.