Supply Chain Diagnostic Assessment
An independent, structured assessment that examines supply chain operations end-to-end, digging into data, processes, and workflows to identify the root causes of inefficiencies — not just the symptoms. Typically the first deliverable in any supply chain consulting engagement.
Standard Assessment Structure
Section titled “Standard Assessment Structure”Duration: 2–4 days on-site (initial diagnostic phase). Full assessment including analysis: 2–6 weeks depending on scope.
Three data collection methods run in parallel:
- Personnel interviews — cross-functional, multi-level. Functions covered: Finance, Sales, Operations, Supply Chain, Engineering, Customer Service
- Data analysis — financial records, WMS/ERP extracts, operational KPIs
- Facility tours — direct observation of current operations
Validation principle: Qualitative observations from interviews must be cross-referenced against quantitative data. When they diverge, the divergence is itself a finding.
The Six Assessment Pillars (SCOR-Aligned)
Section titled “The Six Assessment Pillars (SCOR-Aligned)”| Pillar | What’s assessed |
|---|---|
| Plan | Demand forecasting accuracy; S&OP process maturity; inventory policy; supply/demand balancing |
| Source | Procurement strategy; supplier evaluation and performance; make-vs-buy; contract structure |
| Make / Operations | Production flow; lean maturity; quality systems; capacity utilization |
| Deliver / Warehousing | Inventory management; pick/pack/ship operations; multi-channel fulfillment |
| Deliver / Transportation | Mode selection; carrier management; routing optimization; freight cost |
| Return | Returns processing; reverse logistics; closed-loop sustainability |
SCOR Model — The Diagnostic Framework
Section titled “SCOR Model — The Diagnostic Framework”The Supply Chain Operations Reference (SCOR) model is the cross-industry standard diagnostic tool for supply chain management. Published by ASCM (formerly APICS).
6 process domains:
- Plan — Balance aggregate demand and supply
- Source — Procure goods and services
- Make — Transform product to finished state
- Deliver — Fulfill demand (order management, transportation, installation)
- Return — Handle reverse flows
- Enable — Manage rules, performance, data, resources, contracts, risk, compliance
3-level hierarchy:
- Level 1: Scope (what processes are in play)
- Level 2: Configuration (type of supply chain — make-to-stock, make-to-order, engineer-to-order)
- Level 3: Process elements with performance attributes and best practices
5 performance attributes:
| Attribute | What it measures | Key metrics |
|---|---|---|
| Reliability | Supply chain delivers the right product to the right place, right time, right condition, right quantity | Perfect Order Fulfillment |
| Responsiveness | Speed of supply chain | Order Fulfillment Cycle Time |
| Agility | Ability to respond to external influences | Upside/Downside Flexibility |
| Cost | Cost of operating the supply chain | Total Cost to Serve |
| Asset Management Efficiency | Effective use of assets | Inventory Days of Supply, Cash-to-Cash Cycle |
How consultants use SCOR:
- Current state mapping: Document processes at Level 2–3 detail
- Benchmarking: Compare performance against SCOR industry benchmarks (parity, advantage, superior)
- Gap identification: Flag gaps between current performance and benchmark targets
- Improvement development: Select and sequence best practices from the SCOR best practice library
- Implementation: Pilot, measure, scale
Benchmarking in Diagnostics
Section titled “Benchmarking in Diagnostics”Benchmarks give the gap its number. Without them, “inventory turns are low” is a qualitative observation. With benchmarks, “inventory turns are 6× vs. industry median of 10× and best-in-class 18×” is an actionable finding.
Benchmark sources used in supply chain consulting:
- APQC Process Classification Framework — process cost benchmarks
- WERC (Warehousing Education and Research Council) — DC operations benchmarks
- Gartner Supply Chain Top 25 — strategic performance benchmarks
- SCOR Digital Standard — industry-specific benchmarks across all 5 performance attributes
Benchmark framing: Express gaps in financial terms wherever possible. “Our forecast accuracy is 68% vs. best-in-class 92%” is less persuasive than “Our 68% forecast accuracy generates $4.2M in excess inventory at current carrying costs.”
Diagnostic Report Deliverables
Section titled “Diagnostic Report Deliverables”| Deliverable | Description |
|---|---|
| Current state process maps | End-to-end product and information flow |
| Performance data summary | KPI baseline across all 5 SCOR attributes |
| Gap analysis | Current vs. benchmark, quantified in financial terms |
| Root cause analysis | Why the gap exists (not just what the gap is) |
| Prioritized opportunity register | Improvements ranked by impact, effort, and risk |
| ROI calculations | Financial case for top opportunities |
| Transformation roadmap (high-level) | Phased improvement sequence |
Current State vs. Root Cause Distinction
Section titled “Current State vs. Root Cause Distinction”The most common diagnostic failure: treating a gap as a root cause rather than a symptom.
Example:
- Symptom: “Inventory turns are 6×”
- Wrong root cause: “We have too much inventory”
- Actual root cause: Forecast error at SKU level is 47% (MAPE), driving safety stock to 3× what a 20% MAPE would require, because S&OP has no statistical safety stock policy — planners set stock by gut feel
The root cause identifies what process, system, or policy must change. The symptom only tells you there is a problem.
Citwell’s As-Is / To-Be Framework
Section titled “Citwell’s As-Is / To-Be Framework”Citwell’s approach adds three diagnostic dimensions beyond the six pillars:
- Processes — how work is sequenced and executed
- Flows — how product and information moves between nodes
- Organization — how teams are structured and how decisions are made
The iterative sharing approach — distributing summary notes throughout the project rather than waiting for a final report — builds team buy-in progressively and surfaces client-side objections early.
Recommendations should be “achievable and adapted” to the client’s organizational context and internal capacity — not generic framework output applied without modification.
Pro content
Subscribe to read the rest
This article is part of our Pro library — practitioner-level guidance, frameworks, and decision tools written from real projects.
$9/mo Basic · $13/mo Pro · cancel anytime