The Eval Culture Maturity Model
Organizations progress through maturity levels in how they approach AI evaluation. Each level is stable—organizations stay at a level until intentional investment moves them to the next. Most are at Level 1-2. Level 4 is rare and valuable.
Level 1: Ad-Hoc Evaluation
Characteristics: Evaluation is reactive, not proactive. Models are evaluated when they fail in production. No standard processes. Eval decisions are made by individuals. No shared metrics or methodologies. High variance in eval quality.
Behaviors: "We don't really have an eval process. Engineers test as they go." "Eval happens when something breaks." "Different teams use different metrics." "We're not sure if our model is improving or getting worse."
Timeline to Level 2: 6-12 months with focused effort. Minimal investment (mostly process, not infrastructure).
Level 2: Process-Oriented Evaluation
Characteristics: Standard eval processes exist. Metrics are defined. Pre-deployment evaluation is standard. Documentation exists. But evaluation is still separate from decision-making. Eval results are sometimes ignored. Eval doesn't influence product roadmap.
Behaviors: "We have a standard eval protocol. All models go through it before deployment." "We track eval metrics in a spreadsheet." "Engineers understand the importance of eval." "Eval results don't always drive decisions." "We know our baseline but struggle to improve."
Timeline to Level 3: 12-24 months. Requires deep organizational change (more than just process).
Level 3: Integrated Evaluation
Characteristics: Eval is embedded in decision-making. Models can't ship without passing eval thresholds. Eval results directly influence product roadmap. Cross-functional teams (engineering, product, eval specialists) are involved in eval design. Continuous eval is standard. Feedback loops from production are captured and used to improve evaluation.
Behaviors: "Eval results are the primary metric we track." "Engineers can't deploy without eval approval." "Eval specialists are involved in product planning." "We have continuous eval in production." "Model improvements are driven by eval insights." "Product roadmap is influenced by eval bottlenecks."
Timeline to Level 4: 12-18 months. Requires significant cultural shift—eval becomes valued equally to shipping.
Level 4: Eval-Driven Excellence
Characteristics: Eval is core to organizational identity. Evaluation is proactive, exploratory, and strategic. The organization runs eval challenges internally. External evaluation is published. Eval findings shape long-term strategy. Eval-driven insights are competitive advantage. New evaluation methodologies are developed and tested within the organization before deployment.
Behaviors: "Eval is how we think about AI." "We run internal eval competitions regularly." "We publish our eval methodology." "Product strategy is set by eval insights." "We're developing new eval approaches." "Evaluation is a valued career path." "The best engineers want to work on eval."
Culture Assessment Survey: Diagnostic Checklist
Use this 30-question survey to diagnose where your organization sits. Rate each 1-5 (1=never, 5=always):
Strategy and Leadership (8 questions)
1. Executive leadership understands AI evaluation importance 2. Evaluation is included in strategic planning 3. Eval budget is protected (not first to be cut) 4. We have a dedicated eval team or function 5. Eval is part of organizational metrics/OKRs 6. Product roadmap is influenced by eval findings 7. Long-term strategy considers eval capabilities 8. We invest in eval infrastructure and tools
Processes and Standards (8 questions)
9. We have documented eval protocols 10. All models go through standard eval before deployment 11. Eval metrics are standardized across teams 12. We track eval metrics consistently over time 13. Eval checkpoints are defined in development workflow 14. We have a process for responding to eval failures 15. Eval results are documented and archived 16. We have clear evaluation criteria by use case
Capability and Expertise (6 questions)
17. We have people with eval expertise 18. Engineers understand how to design evals 19. We invest in eval training 20. Eval specialists are involved in product decisions 21. We stay current with eval research 22. We have access to evaluation tools and infrastructure
Culture and Mindset (8 questions)
23. Engineers value evaluation as much as shipping 24. We celebrate good eval practices 25. Failure is seen as information, not blame 26. People feel safe reporting negative eval results 27. Evaluation is seen as risk mitigation, not delay 28. Teams want to improve eval capability 29. Evaluation findings influence hiring/promotion 30. People understand eval basics (not just specialists)
Scoring: Sum all responses. 30-75 = Level 1. 76-120 = Level 2. 121-135 = Level 3. 136-150 = Level 4. Use this to identify gaps. If you score 120 on processes but 80 on culture, you have a cultural problem.
Culture Change Timeline: Realistic Expectations
Culture doesn't change fast. Each level transition takes time:
L1 → L2 (6-12 months). Relatively easy. You're documenting existing work and standardizing. Minimal resistance. Quick wins visible early. Focus: document processes, train people, select standard metrics.
L2 → L3 (12-24 months). Harder. You're now asking engineers to care about eval results, not just shipping speed. There's cultural resistance ("eval is slowing us down"). You need leadership support and visible wins. Focus: integrate eval into decision gates, celebrate eval wins, align incentives.
L3 → L4 (12-18 months). Hardest. You're asking people to rethink identity. "We're an eval-driven company" requires deep cultural change. Requires significant leadership modeling. Focus: storytelling about eval wins, making eval a career path, publishing research, running internal challenges.
Total: 30-54 months (2.5-4.5 years) from L1 to L4. This is realistic. Companies trying to move too fast (12 months L1 → L4) fail because culture doesn't move that fast.
Leadership Behaviors That Drive Eval Culture
Culture is set from the top. Here are specific behaviors senior leaders take to signal eval matters:
Public Commitment
(1) In all-hands meetings, talk about eval. "This quarter, we improved our hallucination detection by 15%. This eval capability is why our product is better than competitors." This takes 3-5 minutes but signals priority. (2) Ask about eval in reviews. "How did you improve our evaluation capability this year?" makes people realize eval is career-path relevant. (3) Mention eval in recruiting. "One of the things we're known for is our rigorous evaluation process." Attracts evaluation-minded people.
Resource Allocation
(1) Hire for eval. Senior eval specialist, junior eval engineers, eval infrastructure engineers. Eval team should be 5-10% of total engineering (vs. current 1-2% in most companies). (2) Fund eval infrastructure. Tools, data, compute for evaluation. $50K-500K annually depending on scale. (3) Protect eval budget. During cuts, don't cut eval first. This signals priority.
Decision-Making Integration
(1) Require eval in launch criteria. "We don't ship X until eval shows Y." This is a forcing function—makes eval relevant. (2) Ask "How are we measuring this?" in product meetings. Before shipping a feature, discuss how you'll evaluate it. (3) Make eval part of success metrics. Product success is measured by: user retention (product), latency (engineering), eval score (eval). Equal weight.
Storytelling
(1) Tell war stories about eval catches bugs. "Our fairness eval caught a bug that would have harmed 10,000 users. Thank you, eval team." (2) Share eval research internally. Publish papers on your eval methodology. Share with employees first. (3) Celebrate eval achievements. Internal award: "Best Eval Insight of the Year." Make visible.
Modeling
(1) Leaders participate in eval. SVP of Product reviews eval results monthly. Chief Scientist designs eval protocols. Signal: "This is important enough for me." (2) Admit uncertainty. "We're not sure if this improvement is real. Let's design an eval to find out." Models intellectual humility. (3) Change decisions based on eval. Ship something, eval shows it's worse than expected, revert. Publicly say "eval was right, we were wrong." Shows eval actually drives decisions.
Psychological Safety and Eval Culture
Psychological safety is essential for eval culture. Without it, people hide negative results. If your eval catches a problem, you need to be able to talk about it without fear of blame.
The safety problem: "My eval shows our model is worse. If I report this, I look incompetent. If I hide it, the company deploys broken AI." People often choose to hide. Results: eval becomes meaningless (people game it), products ship broken, evaluation stops being trusted.
Building safety: (1) No blame for negative results. Bad eval results are information. Celebrate them (you caught a problem!). (2) Share failures openly. Leaders talk about their eval failures. "I thought this approach would work. The eval said no. I was wrong." (3) Reward honesty. "Who caught a problem that would have shipped?" Give bonus/recognition. (4) Punish hiding. If someone hides negative eval, that's serious. Hiding is worse than failure. (5) Make failures non-fatal. Design so failures are caught early (pre-deployment eval) not late (production eval).
Eval Culture in Remote and Distributed Teams
Remote makes cultural change harder. You can't rely on hallway conversations. Culture is deliberate, not emergent.
Strategies: (1) Regular eval ceremonies. Weekly 30-min "Eval Standup." Anyone can attend. Share recent evals, blockers, wins. Creates community. (2) Async communication. Document evals in shared repos. Eval notebooks that anyone can read. This scales to distributed teams. (3) Synchronous learning. Monthly "Eval Talk" (30-45 min): someone shares deep eval work. Watch together. (4) Eval partnerships. Pair senior and junior on evals. Distributed pairs use video calls. (5) Annual in-person eval summit. Once a year, bring eval team to one location. Team building + deep work. Signals eval is important.
Celebrating Eval Wins: Recognizing and Rewarding Good Practice
What gets measured gets managed. What gets celebrated gets repeated.
Recognition program ideas: (1) "Eval Catch of the Month." Monthly award for best eval that caught a bug. $500 bonus + public recognition. (2) "Eval Innovation." New methodology or tool that improves evaluation. Conference talk opportunity + internal prestige. (3) "Eval Culture Champion." Person who most improved eval practices. Public award. (4) "Eval Dashboard." Public leaderboard. Teams competing to improve eval coverage, reduce eval time. Friendly competition. (5) "Eval Story." Tell stories of evals that mattered. In all-hands, blog posts, internal newsletter.
Case Study: Three Companies at Different Maturity Levels
TechStartup Inc. (Level 1)
50 people, 2 year old, building a customer service chatbot.
Eval practice: No formal evaluation. One person (Sarah, an engineer) is "the QA person." She manually tests models before deployment. No documentation. No metrics. Her tests are sometimes rigorous, sometimes cursory depending on her bandwidth. Customers sometimes report chatbot issues. Nobody knows whether the model is improving over time.
Culture: Eval is seen as overhead. "Can't we just ship and see what happens?" Leadership doesn't ask about eval metrics. Eval is done (rarely) but not valued. New engineers don't learn about eval because nobody teaches it. Sarah is overworked and burnt out.
Path forward: (1) Hire an eval specialist. (2) Document what Sarah does. (3) Define 3 key metrics: accuracy on known test cases, response time, user satisfaction rating. (4) Create a 1-week eval process before production deployment. (5) Track metrics monthly. (6) In 6 months, move to Level 2.
ScaleUp Corp. (Level 2)
500 people, evaluating 10+ LLM applications (Q&A, summarization, code generation).
Eval practice: Documented eval protocols. Each use case has defined metrics and baselines. All models go through pre-deployment eval. Eval results are tracked in a dashboard. Engineers understand the process. BUT: eval results don't always drive decisions. Sometimes models ship despite poor eval because there's schedule pressure. Eval team has no influence on product roadmap. Production issues happen, but the connection to eval gaps isn't made.
Culture: Eval is respected but not beloved. "We do eval because it's the right thing." Engineers tolerate it. But no one is excited. There's no "eval career path." No one chose to work at ScaleUp specifically for eval.
Path forward: (1) Integrate eval into product launch criteria. Model can't ship unless it meets eval threshold. (2) Eval specialist sits in product planning meetings. (3) Create "eval champions" on each team—engineer trained to design and run evals. (4) Published eval research on your blog. (5) In 18 months, move to Level 3.
AI Leader Corp. (Level 3)
2,000 people, evaluating 50+ models and systems. Industry known for rigorous eval.
Eval practice: Sophisticated multi-metric evaluation frameworks. Continuous evaluation in production. Feedback loops from production inform eval design. Cross-functional eval teams (engineers, product, data scientists, domain experts). Eval blockers are treated like production bugs. Eval is part of formal design review process.
Culture: "We're an eval-first company." New engineers are trained in eval as part of onboarding. Eval specialists are well-paid and respected. Internal talks about eval happen monthly. Senior leaders participate in eval design. Product roadmap explicitly influenced by eval gaps. Mistakes in eval are failures; mistakes in shipping are learning experiences. Eval people are seen as "the ones who catch bugs before users do."
Path forward: (1) Start publishing eval research. (2) Run internal eval competitions/challenges. (3) Develop novel eval methodologies. (4) Make "Chief Evaluation Officer" a named role. (5) In 15 months, move to Level 4 if you continue investing.
Transformation Roadmap: L1 to L4 Transition Plan
| Timeline | Level 1 → 2 | Level 2 → 3 | Level 3 → 4 |
|---|---|---|---|
| Duration | 6-12 months | 12-24 months | 12-18 months |
| Key Focus | Standardize processes, define metrics | Integrate into decisions, align incentives | Make eval core identity, develop research |
| Leadership | Acknowledge eval importance | Require eval in launch gates | Model eval thinking in all decisions |
| Hiring | 1 eval specialist | 3-5 eval team members | 10+ eval people, spread across org |
| Investment | $50K-150K (mostly people) | $200K-500K (tools + people) | $500K-2M (infrastructure + research) |
| Cultural Wins | Eval is understood | Eval is valued | Eval is loved |
Anti-Patterns: What Kills Eval Culture
Anti-pattern 1: Eval theater. You have an eval process, but it doesn't matter. Models ship regardless of eval results. People see through this. Eval becomes meaningless. Fix: Make eval results matter. Don't ship if eval fails.
Anti-pattern 2: Blame culture. Eval catches a problem. The person who built the model is blamed/fired. People learn to hide problems. Fix: Frame failures as information. No blame for honest failures.
Anti-pattern 3: Eval specialization. Only "eval specialists" do eval. Engineers aren't trained. When specialists leave, eval collapses. Fix: Spread eval knowledge. Every engineer should understand eval basics.
Anti-pattern 4: Eval metrics obsession. You optimize for eval metrics, not real performance. Models get better on evals but worse in production. Fix: Use leading indicators + production monitoring. Evals predict, but don't define, success.
Anti-pattern 5: Process without purpose. Heavy eval processes that slow shipping but don't improve quality. People resent eval. Fix: Make processes lean. Eval should catch important problems, not create bureaucracy.
(1) You can't skip levels. Companies trying to jump L1 → L3 fail. Culture doesn't jump. (2) You can slide backward. If leadership changes and deprioritizes eval, you slide from L3 → L2. It's not one-way. (3) Different teams can be different levels. Product team might be L3, research team L1. This is OK (until you integrate). (4) Maturity isn't about tooling. Spreadsheets can be L3, fancy platforms can be L1. It's about culture and integration, not tools.
Eval Culture and Maturity Summary
- Four levels: L1 (Ad-hoc), L2 (Process), L3 (Integrated), L4 (Excellence)
- Current state: 62% at L1, 28% at L2, 8% at L3, <2% at L4
- Assessment: 30-question survey diagnoses current state
- Timeline: 6-12 months per level (30-54 months L1 → L4)
- Leadership: Public commitment, resource allocation, decision integration, storytelling, modeling
- Psychology: Safety essential—people must feel safe reporting negative results
- Remote: Requires deliberate ceremonies and documentation, not emergent culture
- Celebration: What gets celebrated gets repeated—recognize eval wins publicly
- Anti-patterns: Theater, blame, specialization, metric obsession, bureaucracy—avoid these
- Investment: L1→L2: $50K-150K. L2→L3: $200K-500K. L3→L4: $500K-2M annually
Ready to Build Eval Culture?
Start with the assessment survey. Identify gaps. Find one thing to improve. Build from there. Culture change is slow but compounding.
Exam Coming Soon