★ OverviewBeginner8 min

Explicit Criteria for Precision

Why vague instructions like 'be conservative' fail to improve precision, and how to replace them with explicit categorical criteria that dramatically reduce false positives in code review, data extraction, and CI/CD pipelines.

Quick Reference

  • Explicit criteria = concrete rules that define exactly what constitutes a finding
  • Vague instructions ('be conservative', 'only report high-confidence findings') do NOT improve precision
  • High false-positive rates erode trust even when accurate categories exist
  • Specific categorical criteria outperform confidence-based filtering every time
  • Define severity levels with concrete code examples, not abstract descriptions
  • Temporarily disable high-false-positive categories rather than adding 'be careful' instructions
  • Each criterion should be independently verifiable by reading the code
  • Exam Scenario 5 (CI/CD): precision is critical when findings block a merge
  • Exam Scenario 6 (Structured Data Extraction): explicit field-level criteria prevent hallucinated data

Why Precision Matters More Than Recall

In production systems, a false positive is often more damaging than a missed finding. When a CI/CD pipeline flags 15 'issues' on every PR and 12 of them are noise, developers stop reading the output entirely. The 3 real issues get lost. This is the precision problem: once trust erodes, even accurate findings get ignored.

Exam trap: 'be conservative' does NOT work

A common exam scenario presents a code review system with high false positives and asks which fix is best. Adding general instructions like 'be conservative' or 'only report high-confidence findings' is always the WRONG answer. Claude cannot reliably self-assess confidence. The correct answer is always to provide explicit categorical criteria.

The core insight is this: Claude is excellent at following specific rules but unreliable at self-assessing its own confidence. When you say 'only report high-confidence findings,' you are asking the model to do something it fundamentally cannot do well. When you say 'flag comments only when the claimed behavior contradicts the actual code behavior,' you are giving it a concrete, verifiable check.