Article
Service DesignDecision Matrix: Guide, Practical Example & Template
The decision matrix step by step: weighted evaluation of alternatives with practical example and ready-to-use template.
The decision matrix (also called evaluation matrix, utility analysis, or scoring model) is a structured tool for systematically comparing multiple alternatives against weighted criteria to reach a transparent, justifiable decision. Instead of deciding by gut feeling or deferring to the loudest stakeholder, the matrix makes visible which option scores best under the defined criteria. The method traces its formalized roots to Christof Zangemeister’s utility analysis (Nutzwertanalyse, 1976) and Stuart Pugh’s concept selection matrix (1981) [1][2].
What distinguishes the decision matrix from a simple pros-and-cons list: it forces explicit weighting of criteria. This makes visible what truly matters to the decision team — before alternatives are evaluated. Without weighting, a list treats “cost” and “brand fit” as equal, even though one may be strategically decisive and the other a nice-to-have.
Search for “decision matrix” and you will find dozens of results with the same smartphone-purchase examples. None demonstrates the method in a service process. None explains where the weightings should come from — and which cognitive biases distort them. None systematically compares the decision matrix with Kano, MoSCoW, or RICE. And none honestly states when the method is the wrong choice.
This guide closes those gaps.
From Zangemeister to Pugh: Where the method comes from
The decision matrix has two academic roots that are rarely named together.
Christof Zangemeister published Nutzwertanalyse in der Systemtechnik in 1976 — the first systematic methodology for weighted evaluation of alternatives in the German-speaking world [1]. Zangemeister’s contribution was formalization: he defined a traceable process from criteria selection through weighting to calculation of total utility scores. The utility analysis quickly became the standard tool in German public planning and engineering.
Stuart Pugh (1929–1993), professor at the University of Strathclyde, developed the Pugh matrix in 1981 as a tool for concept selection in product development [2]. Pugh’s innovation: instead of rating all alternatives on an absolute scale, they are compared against a reference concept (datum) — better (+), same (0), or worse (-). This reduces cognitive load because the team only judges relative differences, not absolute values.
Thomas L. Saaty added the Analytic Hierarchy Process (AHP) in 1980, a mathematically rigorous method for deriving criteria weights through pairwise comparisons [3]. AHP is more precise than simple percentage allocation but also significantly more labor-intensive — which is why it is less common in practice than the classic decision matrix.
What is called a “decision matrix” today
In practice, utility analysis, Pugh matrix, scoring model, and decision matrix are often used interchangeably. The common core: a tabular structure where rows are criteria and columns are alternatives. Differences lie in the details — absolute vs. relative evaluation, numerical scale vs. +/0/-, with or without weighting. This guide covers the weighted decision matrix with a numerical scale, because it has the broadest application range and incorporates utility analysis as its methodological foundation.
When is the decision matrix the right tool?
The decision matrix is most valuable when you need to make a one-time selection decision between clearly defined alternatives — not when you want to improve processes or understand customer needs.
Use the decision matrix when:
- You want to compare 3–7 alternatives (for 2, a pros-and-cons list suffices; beyond 7, the evaluation becomes unwieldy)
- You need to weigh multiple decision criteria (cost, feasibility, customer impact, strategic fit)
- The decision must be documented traceably — for stakeholders, audits, or later reflection
- You want to objectify group decisions: “The data shows Concept B scores highest” is more convincing than “I think B is better”
- You want to systematically evaluate concepts generated through a creative process — for example, with the morphological box
Use a different tool when:
| Situation | Better alternative | Why |
|---|---|---|
| You want to understand how features affect customer satisfaction | Kano model | Kano classifies by satisfaction asymmetry; the matrix only rates “how good” |
| You need a quick sprint prioritization without detailed scoring | MoSCoW | MoSCoW is faster and requires no numerical evaluation |
| You want to quantify Reach, Impact, Confidence, and Effort | RICE Scoring | RICE has a fixed formula; the matrix allows any criteria |
| You need iterative improvement, not a one-time selection | PDCA cycle | PDCA improves existing processes; the matrix selects between options |
| You want to analyze strategic strengths and weaknesses | SWOT analysis | SWOT analyzes the starting position; the matrix selects between action options |
Comparison: Decision matrix vs. Kano vs. MoSCoW vs. RICE
Four evaluation and prioritization methods in direct comparison:
| Dimension | Decision matrix | Kano model | MoSCoW | RICE Scoring |
|---|---|---|---|---|
| Focus | Selecting the best alternative from a defined set | Satisfaction impact of individual features | Must/Should/Could/Won’t prioritization | Quantitative impact prioritization |
| Data source | Team evaluation against weighted criteria | Customer survey (functional/dysfunctional) | Stakeholder assessment | Reach, Impact, Confidence, Effort |
| Complexity | Low to medium | Medium — requires survey design | Low — workshop format | Medium — requires metrics |
| Best for | One-time concept selection with multiple criteria | Understanding WHY features matter | Quick release planning | Backlog prioritization with limited resources |
| Weakness | Criteria weighting is subjective | Requires customer survey | No numerical output | Confidence is often estimated |
| Origin | Zangemeister (1976), Pugh (1981) | Kano (1984), Japan | Dai Clegg (1994) | Sean McBride / Intercom (2016) |
Our recommendation: The decision matrix works best as the conclusion of an innovation process — when you have developed multiple concepts through the service design process and now need to select the most promising one. For upstream feature prioritization within a concept, Kano or RICE are more targeted.
Step by step: Creating a decision matrix
Time frame: 60–90 minutes as a team — 20 minutes for criteria and weighting, 30 minutes for evaluation, 10–20 minutes for analysis and sensitivity check.
Step 1: Formulate the decision question
Frame the decision as a concrete question — not a vague intention.
| Poorly formulated | Well formulated |
|---|---|
| ”Which concept shall we take?" | "Which of the four service concepts for digital onboarding should be piloted in Q3?" |
| "What should we do?" | "Which location is best suited for the new service center?" |
| "What is the best solution?" | "Which CRM platform best meets our service automation requirements?” |
Why this matters: A vague question produces vague criteria. If the team does not know what it is deciding, it cannot define meaningful evaluation criteria.
Step 2: Define criteria
List the criteria against which alternatives will be evaluated. Good criteria are:
- Independent of each other — two criteria measuring the same thing (e.g., “costs” and “budget requirements”) distort the result through double counting
- Assessable — the team must be able to rate each alternative against the criterion
- Decision-relevant — if a criterion creates no difference between alternatives, you can drop it
Typical criteria for service decisions:
| Category | Example criteria |
|---|---|
| Customer impact | Customer satisfaction, usability, time-to-value |
| Economics | Implementation cost, ongoing cost, payback period |
| Feasibility | Technical complexity, resource availability, time horizon |
| Strategy | Brand fit, scalability, differentiation potential |
| Risk | Implementation risk, regulatory requirements, dependencies |
Recommended count: 5–8 criteria. Fewer than 5 is too coarse; more than 10 overtaxes the team’s evaluation capacity and creates false precision.
Step 3: Weight the criteria
Weighting is the most critical step — and the most frequently underestimated. Without weighting, the matrix treats all criteria as equal, which rarely matches reality.
Method 1: Percentage allocation (simple) Distribute 100 percentage points across criteria. Each participant distributes individually, then averages are calculated.
Method 2: Pairwise comparison (thorough) Compare each criterion with every other: “Is customer satisfaction more important than cost?” Count each criterion’s “wins” and derive the weighting. Saaty’s AHP formalizes this approach with a 9-point scale [3].
Method 3: Ranking (fast) Rank criteria by importance. Assign the highest value to the most important criterion (e.g., 5 for 5 criteria), the next gets 4, etc. Normalize to 100%.
| Criterion | Weight |
|---|---|
| Customer impact | 30% |
| Feasibility | 25% |
| Economics | 20% |
| Strategic fit | 15% |
| Risk | 10% |
| Total | 100% |
Common mistake: Adjusting the weighting after the evaluation so the desired result emerges. This undermines the entire purpose of the method. Define weighting before evaluating the first alternative — and do not change it afterward unless you have a factual reason (e.g., new information).
Step 4: Evaluate alternatives
Rate each alternative against each criterion on a uniform scale. Common scales:
- 1–5 (poor to excellent) — simple, sufficient for most decisions
- 1–10 — more differentiated but increased risk of false precision
- +/0/- (Pugh method) — fastest variant but no numerical total
Recommendation: Use a 1–5 scale with clear anchor points:
| Score | Meaning | Example (criterion: Feasibility) |
|---|---|---|
| 1 | Very poor | Requires technology we do not yet have |
| 2 | Poor | Feasible but with significant additional effort |
| 3 | Medium | Feasible with existing resources in 6+ months |
| 4 | Good | Feasible with existing resources in 3 months |
| 5 | Very good | Immediately implementable with current means |
Practical tip: Have each participant rate individually first (silent rating) before comparing evaluations. This prevents groupthink and anchoring effects — the first number spoken aloud influences all subsequent ratings [4].
Step 5: Calculate weighted total scores
Multiply each rating by the criterion’s weight and sum:
Weighted utility score = Sum (Rating x Weight)
The concept with the highest weighted utility score is — under the chosen criteria and weights — the best option.
Step 6: Perform a sensitivity check
Before communicating the decision, test how robust the result is.
- Vary weighting: What happens if you increase or decrease the most important criterion’s weight by 10 percentage points? Does the ranking change?
- Vary ratings: Are there ratings where the team was uncertain (e.g., 3 or 4)? What happens if you shift those by 1 point?
- Closeness test: If the gap between first and second place is less than 5%, the result is not clear-cut — you need additional differentiating criteria or more information.
If the sensitivity analysis shows the result is fragile: This is not a method failure — it is a valuable insight. It means the alternatives are close and the decision should not rest on the matrix alone. Add qualitative factors (e.g., team capacity, political feasibility, timing).
Example: Decision matrix for a service concept
Context: An insurer has developed three concepts for a new claims management portal through a service design project. The project team must decide which concept enters the pilot phase.
The three concepts:
- A: Self-service portal — Customers report claims entirely digitally, upload photos, track status
- B: Hybrid model — Digital reporting with optional video call for complex claims
- C: AI-powered portal — Automatic claim classification via image analysis, automated initial assessment
| Criterion | Weight | A: Self-service | B: Hybrid | C: AI-powered |
|---|---|---|---|---|
| Customer impact | 30% | 4 | 5 | 3 |
| Feasibility | 25% | 5 | 4 | 2 |
| Economics | 20% | 4 | 3 | 2 |
| Strategic fit | 15% | 3 | 4 | 5 |
| Risk (inverted: 5 = low risk) | 10% | 5 | 4 | 2 |
| Weighted utility score | 4.20 | 4.10 | 2.75 |
Calculation Concept A: (4 x 0.30) + (5 x 0.25) + (4 x 0.20) + (3 x 0.15) + (5 x 0.10) = 1.20 + 1.25 + 0.80 + 0.45 + 0.50 = 4.20
Calculation Concept B: (5 x 0.30) + (4 x 0.25) + (3 x 0.20) + (4 x 0.15) + (4 x 0.10) = 1.50 + 1.00 + 0.60 + 0.60 + 0.40 = 4.10
Calculation Concept C: (3 x 0.30) + (2 x 0.25) + (2 x 0.20) + (5 x 0.15) + (2 x 0.10) = 0.90 + 0.50 + 0.40 + 0.75 + 0.20 = 2.75
Sensitivity check: The gap between A (4.20) and B (4.10) is only 2.4%. If “customer impact” weighting increases from 30% to 35%, B overtakes A. The result is close — the decision depends heavily on how highly customer impact is prioritized.
Team decision: Concept A is launched as a pilot, with the video-call feature from Concept B added for complex claims in Phase 2. Concept C is earmarked as a strategic option for 2027, when AI infrastructure is more mature.
Note: This example is illustratively constructed to demonstrate the method in a service context. The ratings are based on typical industry values.
Decision matrix: Ready-to-use template
Use this checklist directly for your next decision matrix:
Preparation
- Decision question formulated as a concrete question
- 3–7 alternatives identified
- 5–8 independent, assessable criteria defined
- Knock-out criteria checked upfront (alternatives failing a knock-out criterion are excluded)
Weighting
- Weighting method chosen (percentage, pairwise comparison, or ranking)
- Weighting established BEFORE evaluation
- Sum of weights = 100%
Evaluation
- Scale with anchor points defined (e.g., 1–5)
- Each participant rated individually (silent rating)
- Deviations discussed in team and consensus reached
- Weighted utility scores calculated
Quality assurance
- Sensitivity check performed (weighting +/-10%, rating +/-1)
- Result tested for robustness (gap between 1st and 2nd place)
- Qualitative factors added that the matrix does not capture
- Decision documented including rationale and alternative evaluations
4 common mistakes with the decision matrix
1. Adjusting weights after evaluation (anchoring bias)
Symptom: The team evaluates the alternatives, sees the result — and then changes the weighting until the desired outcome appears. “Cost is actually more important than customer impact” is only stated after it becomes clear that the cheapest concept would otherwise lose.
Why this hurts: The decision matrix then only simulates objectivity — the actual decision was already made by gut feeling. Tversky and Kahneman (1974) documented how anchor values systematically distort judgment [4].
Solution: Define weighting in a separate session before the alternatives are known or evaluated. Document the weighting in writing and declare it fixed — unless new factual information warrants an adjustment (not new preferences).
2. Criteria redundancy (double counting)
Symptom: The criteria “implementation costs,” “budget requirements,” and “economics” sit side by side — and all three essentially measure the same thing.
Why this hurts: If three of eight criteria measure the same aspect, that aspect effectively carries 37.5% weight instead of its nominally assigned individual weight. This skews the result in favor of alternatives strong in that aspect.
Solution: Test each criterion: “If alternative X scores better than Y on this criterion — would the result automatically be better on another criterion too?” If yes, the criteria are not independent. Consolidate them.
3. False precision through overly fine scales
Symptom: The team uses a 1–10 scale and debates whether a concept deserves a 7 or an 8 for “customer impact.” The discussion takes 15 minutes and yields no insight.
Why this hurts: Human evaluation capacity is limited. Distinguishing between 6, 7, and 8 on a 10-point scale is arbitrary in most cases. The resulting false precision creates unwarranted confidence in the result.
Solution: Use a 1–5 scale with clear anchor points for each value. If you need more differentiation, consider whether you need an additional criterion — not a finer scale.
4. Groupthink during evaluation
Symptom: The project lead states their rating first — and everyone else agrees or deviates only marginally. The result reflects one person’s opinion, not the collective assessment.
Why this hurts: Groupthink eliminates the diversity of perspectives that makes group decisions valuable [5].
Solution: Silent rating — everyone evaluates individually on their own sheet or in a separate spreadsheet before ratings are revealed. Discuss only deviations, not agreements.
When the decision matrix does NOT work
1. Genuine innovation with unknown criteria: When evaluating a radically new service concept, you often do not yet know the relevant criteria. A matrix with wrong criteria produces a precise but irrelevant result. Here, Design Thinking (prototyping + testing) is better suited — first learn, then evaluate.
2. Political decisions: When the decision has already been made and the matrix serves only as legitimation, it is a waste of time. Worse: it breeds cynicism among participants toward future “objective” processes.
3. Decisions under extreme uncertainty: The matrix assumes that alternatives and their properties are known. For decisions under uncertainty — such as entering an entirely new market segment — scenario analysis or real-options approaches deliver better outcomes.
4. When one alternative dominates all others: If one option is equal or better on every criterion (Pareto dominance), you do not need a matrix. It is obviously the best choice. The matrix is overhead without insight.
5. Too little information for a sound evaluation: When the team does not know enough about the alternatives to evaluate them seriously, the matrix creates a structure that masks ignorance. Invest first in information gathering — a Gemba walk for process observation, a Kano analysis for customer data — and run the matrix when you can evaluate with substance.
Variations and advanced techniques
Pugh matrix: Relative instead of absolute evaluation
Instead of rating each alternative on a scale, the Pugh matrix selects a reference concept (datum) and evaluates all others relative to it: better (+1), same (0), or worse (-1) [2]. Advantage: the team does not estimate absolute values, only relative differences — which is cognitively easier. Disadvantage: no differentiated numerical total.
Recommendation: The Pugh matrix is particularly suitable in early phases of concept selection when alternatives are still rough. The weighted decision matrix is better for the final selection between mature concepts.
Analytic Hierarchy Process (AHP): Mathematically grounded weighting
Saaty’s AHP uses pairwise comparisons with a 9-point scale to derive weights with mathematical consistency [3]. AHP also calculates a consistency index — a measure of how contradiction-free the weightings are. A high inconsistency value indicates that the team has not clearly sorted its preferences.
Recommendation: AHP is worthwhile for strategically important decisions with more than 8 criteria and when mathematical rigor is required (e.g., public tenders, regulated industries).
Decision matrix + Kano: The combined method
Use the Kano model to classify features by their satisfaction impact — then transfer the results as the criterion “customer impact” into the decision matrix. Must-be features automatically receive the highest customer-impact rating, delighter features a high one, indifferent features a low one. This connects customer data with structured alternative evaluation.
Frequently asked questions
What is a decision matrix?
A decision matrix is a tabular tool that systematically compares multiple alternatives against weighted criteria. Each alternative is rated against each criterion on a scale. The rating is multiplied by the criterion’s weight and summed. The alternative with the highest weighted total score (utility value) is — under the chosen criteria — the best option.
How do I create a decision matrix?
In six steps: (1) Formulate the decision question. (2) Define 5–8 independent criteria. (3) Weight the criteria (sum = 100%). (4) Rate alternatives on a 1–5 scale (individually, then consolidated in the team). (5) Calculate weighted utility scores. (6) Perform a sensitivity check — test whether small changes to weighting or ratings flip the ranking.
What is the difference between a decision matrix and a Pugh matrix?
The Pugh matrix evaluates alternatives relative to a reference concept (better/same/worse). The weighted decision matrix evaluates alternatives on an absolute numerical scale. Pugh is faster and cognitively simpler but does not produce a differentiated total score. The weighted matrix is more thorough and produces a utility value but requires more time and careful scale definition.
How many criteria should a decision matrix have?
5–8 criteria are ideal. Fewer than 5 is too coarse and omits important aspects. More than 10 overtaxes evaluation capacity and creates false precision. If you identify more than 10 criteria, check whether some are redundant or can be consolidated into broader categories.
When should you NOT use a decision matrix?
In five situations: (1) For radical innovation when the relevant criteria are still unknown. (2) When the decision has already been made politically. (3) Under extreme uncertainty about the alternatives. (4) When one alternative obviously dominates all others. (5) When too little information is available for a sound evaluation.
Related methods
A typical sequence in service development: With the morphological box, you systematically generate service concepts. With the decision matrix, you select the most promising concept. With the Kano model, you refine the features of the chosen concept. In the service design process, you implement the concept.
- Morphological box: When you want to systematically generate solution combinations before evaluating them
- Kano model: When you want to prioritize features by customer impact rather than choosing between concepts
- SWOT analysis: When you want to analyze the strategic starting position before concept selection
- Service design methods overview: For the overall context in which the decision matrix is embedded
- PDCA cycle: When you want to iteratively improve the chosen concept after the decision
Research methodology
This article synthesizes insights from Zangemeister’s original publication on utility analysis (1976), Pugh’s concept selection methodology (1981), Saaty’s foundational work on AHP (1980), Tversky and Kahneman’s research on judgment heuristics (1974), and the analysis of 10 German-language specialist publications on the decision matrix. Sources were selected for methodological rigor, practical relevance, and currency.
Limitations: Academic literature on the application of decision matrices in service development is limited — most studies originate from engineering and product development. The practical example (claims management portal) is illustratively constructed, not a documented case study.
Disclosure
SI Labs provides consulting services in the field of service innovation. In the Integrated Service Development Process (iSEP), we use the decision matrix to select between service alternatives during the concept phase. This practical experience informs the classification of the method in this article. Readers should be aware of the potential perspective bias.
References
[1] Zangemeister, Christof. Nutzwertanalyse in der Systemtechnik: Eine Methodik zur multidimensionalen Bewertung und Auswahl von Projektalternativen. Munich: Wittemann, 1976. [Foundational work | Utility analysis | Citations: 1,500+ | Quality: 90/100]
[2] Pugh, Stuart. “Concept Selection: A Method That Works.” Proceedings of the International Conference on Engineering Design (ICED), Rome, 1981. Later expanded in: Pugh, Stuart. Total Design: Integrated Methods for Successful Product Engineering. Wokingham: Addison-Wesley, 1991. ISBN: 978-0201416398 [Foundational work | Pugh matrix | Citations: 3,000+ | Quality: 88/100]
[3] Saaty, Thomas L. The Analytic Hierarchy Process: Planning, Priority Setting, Resource Allocation. New York: McGraw-Hill, 1980. ISBN: 978-0070543713 [Foundational work | AHP | Citations: 40,000+ | Quality: 92/100]
[4] Tversky, Amos, and Daniel Kahneman. “Judgment under Uncertainty: Heuristics and Biases.” Science 185, no. 4157 (1974): 1124-1131. DOI: 10.1126/science.185.4157.1124 [Foundational work | Cognitive biases | Citations: 45,000+ | Quality: 95/100]
[5] Janis, Irving L. Groupthink: Psychological Studies of Policy Decisions and Fiascoes. Boston: Houghton Mifflin, 1982. ISBN: 978-0395317044 [Foundational work | Groupthink | Citations: 10,000+ | Quality: 85/100]
[6] Brugha, Cathal M. “Structure of Multi-Criteria Decision Making.” Journal of the Operational Research Society 55, no. 11 (2004): 1156-1168. DOI: 10.1057/palgrave.jors.2601816 [Journal article | MCDM overview | Citations: 100+ | Quality: 75/100]
[7] Velasquez, Mark, and Patrick T. Hester. “An Analysis of Multi-Criteria Decision Making Methods.” International Journal of Operations Research 10, no. 2 (2013): 56-66. [Journal article | Method comparison | Citations: 2,500+ | Quality: 78/100]