Case Study: Elena's Quality Dashboard — Tracking AI-Assisted Consulting Deliverables
The Starting Point
Elena had been using AI tools in her consulting practice for about nine months when she made a discovery that changed how she thought about measurement.
She was reviewing a strategic analysis deliverable she'd produced for a healthcare client — one of her best pieces of work, she thought. She'd used AI extensively to synthesize literature, structure the analysis framework, and draft the narrative. She'd spent substantial time reviewing and editing. The client had been satisfied.
Looking at it again months later, she noticed something she'd missed at the time: the analysis was comprehensive, well-organized, and clear, but it was also somewhat generic. It could have applied, with minor modifications, to almost any healthcare organization facing the same strategic question. What it lacked was the specific institutional insight — the understanding of this client's particular constraints, culture, and competitive position — that her best non-AI-assisted work had.
She hadn't noticed this at the time because the AI-assisted version was so much better than a first draft could have been without AI. But compared to her best work from before AI adoption, it was subtly shallower.
This discovery raised an uncomfortable question: Was AI making her work better or worse? And how would she know, without measuring systematically?
Designing the Quality Framework
Elena's consulting practice has clear deliverable types: strategic analysis reports, organizational assessments, competitive landscape analyses, operational improvement recommendations, and stakeholder communication documents. Each has different quality dimensions.
She started with strategic analysis reports — her flagship deliverable and the one where quality mattered most to client outcomes.
Working through what "excellent" meant for a strategic analysis report, she identified five dimensions:
1. Diagnostic accuracy: Does the analysis correctly identify the key issues and their root causes? (Not just the symptoms, but the underlying drivers)
2. Evidence quality: Are the claims supported by current, relevant, properly-sourced evidence? Is evidence evaluated critically rather than accepted at face value?
3. Analytical rigor: Does the analysis follow from the evidence? Are alternative explanations considered and addressed? Are the limitations of the analysis acknowledged?
4. Institutional specificity: Does the analysis reflect genuine understanding of this particular client's situation, culture, and constraints? Or could it apply, with minimal modification, to any organization in the sector?
5. Actionability: Are the recommendations concrete, prioritized, and feasible for this specific organization to implement?
For each dimension, she defined a 1-5 scale with specific anchor descriptions. Score 5 on institutional specificity meant: "Every major recommendation reflects analysis of this organization's specific situation and couldn't be directly applied to a peer organization without significant rethinking." Score 1 meant: "The analysis and recommendations are generic sector advice that any consultant could produce without deep engagement with this specific client."
The dimension she'd missed in her post-hoc review was institutional specificity. Her AI-assisted analysis had scored well on everything else but low on this.
Implementing the Rating System
Elena built a simple tracking system: a spreadsheet with one row per deliverable, columns for each quality dimension, the AI assistance level (high/medium/low), and client feedback (when available).
She rated each deliverable within 48 hours of delivery, while the work was fresh. The time cost: about 15 minutes per deliverable to complete the rating thoughtfully.
For two months, she collected data before analyzing patterns.
What the Data Showed
After two months and 17 rated deliverables, Elena ran her first analysis. The patterns were instructive:
AI-assisted work scored higher on evidence quality and communication quality. AI helped her find more sources, synthesize them more comprehensively, and organize the communication more clearly. These benefits were consistent across deliverable types.
AI-assisted work scored roughly equivalent on diagnostic accuracy. This was reassuring — AI wasn't impairing her ability to correctly identify the core issues. Her domain expertise, applied through careful prompt construction and critical review of AI analysis, was maintaining this dimension.
AI-assisted work scored lower on institutional specificity, particularly for deliverables with high AI assistance. The discovery that had prompted the measurement exercise was confirmed by the data. High AI assistance, without specific intervention to address this dimension, produced analyses that were more generic than her best non-AI-assisted work.
AI-assisted work scored equivalent on actionability. This was surprising — Elena had expected AI to produce more generic recommendations, but her practice of explicitly asking AI to generate recommendations for "an organization with [specific constraints]" appeared to maintain actionability.
The pattern was clearest at the extremes. Deliverables where AI assistance was low (she'd used AI mainly for research and used her own analysis and writing) were highest rated overall. Deliverables where AI assistance was high (AI had been involved in analysis and significant drafting) showed the most variance — some were excellent, others were noticeably weaker.
Finding the Explanation
The high variance in high-AI-assistance deliverables prompted Elena to look more carefully at what distinguished the good ones from the weaker ones.
She identified two differentiating factors:
Factor 1: The quality of her brief to AI. The strongest AI-assisted deliverables were those where she had provided AI with substantial client-specific context in her initial prompt — specific information about the client's organizational history, competitive challenges, cultural factors, and strategic constraints. When this context was rich, AI incorporated it into the analysis and produced work that felt specific. When this context was thin (usually because she was in a hurry), the analysis was more generic.
Factor 2: The analytical challenge step. She noticed that the strongest AI-assisted analyses had an additional step she'd taken informally: asking AI to identify the weakest points in the analysis and where the recommendations might fail for this specific organization. This challenge step forced AI to engage with the specifics of the client situation in a way that initial analysis prompts often didn't.
The weak analyses — the ones that felt generic — were those where she'd taken AI's first analytical output and refined it without first challenging it.
The Workflow Intervention
Based on this analysis, Elena redesigned her high-AI-assistance workflow:
Before the analysis prompt: She now spends 20-30 minutes writing a detailed client context brief — a structured summary of everything specific to this client that should inform the analysis. This brief becomes the foundation of her AI context, ensuring that AI is working with rich institutional information rather than generic sector data.
The challenge step: After receiving AI's initial analysis, she now runs a structured challenge prompt: "Review this analysis and identify: (1) where the recommendations could apply to any organization in this sector without modification, (2) where our analysis of this specific client's situation should lead to different conclusions than a generic sector analysis, and (3) what we're missing about this organization's specific context."
The challenge step produces a list of gaps. She then researches and addresses them before finalizing the analysis.
Time cost: The client context brief adds 20-30 minutes per engagement. The challenge step adds about 15-20 minutes. Total additional time: 35-50 minutes per deliverable.
Quality impact: After implementing the new workflow on her next eight deliverables, Elena's institutional specificity scores improved from an average of 3.1 to 4.0. The other quality dimensions remained stable or slightly improved.
Her "all-in" time on these deliverables was 30-40 minutes longer, but the quality improvement was significant. And given that her fee per deliverable was in the thousands of dollars, the time investment was trivially small relative to the value of the quality improvement.
Connecting Quality to Client Feedback
Two months after implementing the new workflow, Elena had enough client feedback to check whether her quality ratings correlated with client perception.
Her clients don't fill out formal surveys — she has ongoing relationship conversations instead. She went back through her notes from client conversations following high-rated and low-rated deliverables.
The correlation was clear, though not perfect. Her three highest-rated deliverables (by her own rubric) had all been described by clients as "exactly what we needed" or "the clearest thinking on this problem we've seen." Her two lowest-rated deliverables had both prompted what she recognized in retrospect as mild disappointment — "this is helpful but felt somewhat familiar" from one client, and "can we discuss whether the recommendations can be adapted more specifically to our situation?" from another.
The two low-rated deliverables were both from the early period before she'd implemented the new workflow. Both were high-AI-assistance work without the client context brief or challenge step.
The correlation didn't prove causation, but it was meaningful directional signal: her quality rubric appeared to be measuring something that actually affected client experience.
The Insight About Expert Practitioners and AI
Elena's measurement practice led her to an insight that she's shared with other consultants in her network: for expert-level practitioners, the primary risk of AI assistance is not capability loss — it's depth loss.
AI can help an expert practitioner do more, faster. What it can't do automatically is substitute for the deep engagement with a specific client situation that produces the most differentiated and valuable consulting work. That depth comes from hard-won domain expertise, careful attention to what makes this situation different from similar ones, and the willingness to push past the first adequate answer to find the genuinely right one.
AI, when used without careful design, tends toward the adequate. It produces the synthesis of what is generally known, the analysis that applies broadly, the recommendation that is defensible for most similar situations. For a consultant whose value proposition is exceptional, specific insight — not adequate, general insight — this tendency toward the adequate is the primary quality risk.
The measurement practice helped Elena see this clearly. And seeing it clearly, she was able to design around it.
The Ongoing Practice
Elena now reviews her quality ratings quarterly and looks for drift — trends in which dimensions are improving or declining.
She's added two additional quality checks to her workflow: a monthly "worst deliverable" review (pulling her lowest-rated piece from the month and understanding what made it weaker) and a quarterly "best deliverable" review (understanding what made her highest-rated piece work so well and whether she can systematically replicate it).
Her current quality metrics, eight months after implementing measurement, are substantially higher than at baseline. More importantly, she understands why they're higher — and she understands the specific risks that remain.
That understanding is the most valuable product of the measurement practice. It's not just knowing the numbers. It's knowing what the numbers mean, what drives them, and what to do about them.
Elena's long-term practice — how she continues to develop her AI-augmented consulting approach over two years — is the subject of Chapter 41's first case study.