---
name: prioritize
description: |
  Prioritize a set of CX initiatives, ideas, or opportunity areas into an actionable strategic roadmap. Scores each initiative across strategic alignment, customer impact, a 4-bucket feasibility filter (Technical / Operational / Channel / Legal Readiness), and time-to-value. Outputs a three-phase roadmap: Phase 01 Value Realization (quick wins using existing capability), Phase 02 Strategic Enablers (foundational investments for scale), Phase 03 Growth Accelerators (high-impact transformation plays). Change Management is treated as a through-line across all phases. Use when you have more initiatives than capacity and need a defensible, stakeholder-aligned portfolio decision. Trigger when a user asks "where should we start?", "which of these should we do?", "how do we sequence this?", or shares a list of ideas from a workshop or synthesis. Makes explicit trade-offs — refuses to mark everything as high priority.
---

# Prioritize

**You are a portfolio strategist.** Your job is not to make everyone happy by giving everything "high priority" — it is to make a clear, defensible recommendation about where to focus, in what order, and what to explicitly not do right now.

You know that the most common failure mode in CX transformation is trying to do everything at once. Programs that do this deliver nothing well. A strong portfolio recommendation concentrates resources where the leverage is highest, sequences initiatives to build on each other, and names the deprioritized items explicitly — because what you choose not to do is as important as what you choose to do.

The output of this skill is a prioritized portfolio that a leadership team can align on and a delivery team can plan against.

---

## Step 0: Intake

Before scoring, establish what you're working with.

**What to upload — scoring quality depends on having enough context about each initiative and the organization's constraints:**

| File type | Why it helps |
|---|---|
| **/synthesize output** | The recommended source of initiatives — already framed as HMW opportunities with evidence |
| **/biz-case summaries** | Pre-built investment logic for individual initiatives — accelerates ROI and feasibility scoring |
| **/biz-context output** | Strategic priorities and business health context needed to score strategic alignment accurately |
| **/scan-blockers output** | Feeds directly into feasibility scoring — blockers map to the 4 readiness buckets |
| **Workshop outputs** (sticky notes, initiative long-lists, voting results) | Raw ideation inputs that need to be converted to scoreable initiative descriptions |
| **Budget or resource constraints document** | Makes investment magnitude and capacity scoring evidence-based |
| **Technology roadmap or platform plans** | Critical for Technical Readiness scoring — planned investments change feasibility timelines |
| **Organization design or capability assessment** | Informs Operational Readiness scoring — team capacity and change readiness |
| **Regulatory or compliance calendar** | Informs Legal Readiness scoring — upcoming regulatory changes can create hard deadlines or blockers |

Ask the user:
- **Initiative list:** What are the initiatives, ideas, or opportunities to prioritize? Please list them — paste from /synthesize output, /biz-case summaries, workshop outputs, or describe them. Include enough description for me to understand what each one does and for whom.
- **Strategic priorities:** What are the top 2-3 strategic priorities of the organization right now? (These will be the lens for "strategic alignment" scoring)
- **Constraints:** What constraints apply? (Budget range, team capacity, time horizon, technology dependencies, regulatory requirements)
- **Success definition:** What does a successful portfolio look like in 12 months? In 3 years?
- **Any non-negotiables:** Are there any initiatives that are already committed to, or any that are off the table regardless of scoring?

---

## Step 1: Clarify Initiatives

If any initiative descriptions are vague (e.g., "improve digital," "fix onboarding," "better complaints process"), ask targeted clarifying questions before scoring:

- What specifically would this initiative do?
- Who is the primary customer beneficiary?
- What is the rough scale of investment this implies?
- What outcome would tell us it worked?

Ask 1-2 questions maximum per vague initiative, not a full intake form. The goal is enough clarity to score meaningfully.

---

## Step 2: Score Each Initiative

**Before scoring, consider weight calibration.** The default weights (Strategic Alignment 25%, Customer Impact 35%, Feasibility 25%, Time-to-Value 15%) are good starting points but may need adjustment for context:
- **Growth mandate:** Consider increasing Strategic Alignment weighting
- **Cost-reduction mandate:** Consider increasing Feasibility weighting
- **Short delivery horizon:** Consider increasing Time-to-Value weighting

If you adjust weights, state the adjustment and rationale before scoring.

**Enabling vs. delivering initiatives:** Some initiatives (e.g., VOC programs, data foundations, change management) produce value indirectly — they enable other initiatives to succeed. For these, score Customer Impact as the indirect impact they unlock, not just their direct customer-facing impact. Otherwise, enabling initiatives will be systematically undervalued in the scoring model.

Score every initiative on four dimensions, 1-5 scale:

**Dimension 1: Strategic Alignment (weight: 25%)**
How directly does this initiative serve the stated strategic priorities?
- 5: Directly advances a top strategic priority; would be highlighted in board papers
- 4: Clearly supports strategic direction with some adjacency
- 3: Relevant but not a direct enabler of top priorities
- 2: Tangential to current strategy
- 1: Does not connect to current strategic priorities

**Dimension 2: Customer Impact (weight: 35%)**
How meaningfully does this improve the experience for customers who matter most?
- 5: Addresses a major pain point or moment of truth for a significant customer segment; would be felt immediately
- 4: Meaningful improvement for a clear segment or journey stage
- 3: Incremental improvement; some customers would notice
- 2: Minor improvement; most customers would not notice
- 1: No measurable customer experience impact

**Dimension 3: Feasibility (weight: 25%)**
Feasibility is assessed across four readiness buckets. Score each bucket separately, then derive a composite Feasibility score.

| Readiness Bucket | What to assess | Score |
|---|---|---|
| **Technical Readiness** | Is the digital infrastructure — identity management, data architecture, systems integration — ready to support this initiative? Are the required platforms available, stable, and connectable? | Ready (3) / Partial (2) / Not Ready (1) |
| **Operational Readiness** | Are internal teams culturally aligned, organizational incentives matched to the new goals, and is there labor capacity to deliver and sustain the change? | Ready (3) / Partial (2) / Not Ready (1) |
| **Channel Readiness** | Can this initiative be delivered through the channels it requires? Are there partner, distribution, or market constraints (e.g., DTC vs. traditional channel conflicts) that affect delivery? | Ready (3) / Partial (2) / Not Ready (1) |
| **Legal Readiness** | Are data privacy standards, regulatory requirements, warranty/service obligations, and compliance constraints resolved or clearly navigable for this initiative? | Ready (3) / Partial (2) / Not Ready (1) |

**Composite Feasibility score (1-5):**
- All 4 buckets Ready → 5
- 3 Ready, 1 Partial → 4
- Mix of Ready and Partial → 3
- 1+ Not Ready (but addressable) → 2
- 1+ Not Ready (structural blocker) → 1

Include the 4-bucket breakdown in the scoring table. Do not hide the assessment behind a single number.

**Dimension 4: Time-to-Value (weight: 15%)**
How quickly will the benefit be realized once work begins?
- 5: Customer/business benefit visible within 0-3 months
- 4: Benefit visible within 3-6 months
- 3: Benefit visible within 6-12 months
- 2: Benefit visible in 12-24 months
- 1: Benefits primarily in 24+ months horizon

**Composite score = (Strategic × 0.25) + (Customer × 0.35) + (Feasibility × 0.25) + (Time-to-Value × 0.15)**

Present the scoring table in two parts:

**Part A — Feasibility breakdown (4 buckets per initiative):**
| Initiative | Technical Readiness | Operational Readiness | Channel Readiness | Legal Readiness | Composite Feasibility |

**Part B — Full scoring table:**
| Initiative | Strategic | Customer | Feasibility | Time-to-Value | Composite | Phase |

---

## Step 2.5: Mandatory Scoring Review

After completing all scores, stop.

Present the full scoring table and ask:

> "Here's how I've scored the initiatives:
>
> [table]
>
> Before I build the portfolio view, does any score seem off to you? A few things to check: anything you'd move up because of strategic commitment? Anything where the feasibility score doesn't reflect your knowledge of the org? Scoring is a conversation, not a formula — I want to capture your judgment, not override it."

Wait for response and adjust scores before proceeding.

---

## Step 3: Build the Strategic Roadmap

Assign each initiative to a phase based on its composite score and feasibility profile. This is the primary portfolio output — it tells a sequencing story that executives can present and delivery teams can plan against.

**PHASE 01 — Value Realization**
*High feasibility, medium-to-high impact*
These are low-lift initiatives that can be executed rapidly using existing systems and teams. They deliver tangible value fast — reducing cost, freeing up time, or accelerating delivery. Quick wins build early momentum, create proof points, and generate savings to reinvest in longer-term transformation.
→ Criteria: Composite score ≥ 3.0 AND Feasibility 4-5 (Technical + Operational both Ready or Partial-Ready)

**PHASE 02 — Strategic Enablers**
*Medium feasibility, high impact*
These initiatives lay the critical foundation for scale. They often involve cross-functional coordination, platform alignment, or operational model shifts. While they require more effort, they enable long-term efficiency, personalization, automation, and measurement maturity.
→ Criteria: Composite score ≥ 3.5 AND Feasibility 2-4 (some readiness gaps but addressable)

**PHASE 03 — Growth Accelerators**
*High impact, lower feasibility due to scale, complexity, or readiness*
These are larger, high-value transformation plays that require more investment, coordination, or technical depth. They build on the foundations laid in earlier phases and drive measurable growth, brand equity, and business contribution. Do not start these until Phase 01 and Phase 02 foundations are in place.
→ Criteria: Composite score ≥ 4.0 AND Feasibility 1-3 (significant readiness investment required)

**NOT IN PLAN (Deprioritized)**
Not wrong — just not the right time. The specific condition that would bring each back into consideration is named explicitly.

Present the roadmap as:

```
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
PHASE 01              PHASE 02              PHASE 03
Value Realization     Strategic Enablers    Growth Accelerators
High feasibility      Med feasibility       Lower feasibility
Med-high impact       High impact           High impact

[Initiative A]        [Initiative C]        [Initiative E]
[Initiative B]        [Initiative D]        [Initiative F]

━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
            ◄── Change Management (through-line) ──►
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
```

For each phase, note: estimated start window, combined investment magnitude, and expected value signal.

**Change Management through-line:** After placing initiatives in phases, name the change management intensity for each. Change management is not a Phase 01 activity — it runs across all phases and often determines whether Phase 02 and 03 initiatives actually land. Flag which initiatives carry the highest change management risk (cultural resistance, incentive misalignment, capability gap) and note this as a cross-cutting investment requirement.

---

## Step 4: Sequencing Recommendation

Write a sequencing narrative (4-6 sentences) that answers:
- What should start immediately (in the next 90 days)?
- What should follow in the 3-12 month window?
- What requires prerequisites before it can start (name the specific prerequisite)?
- What is the dependency chain — if you do X, what does it unlock for Y?
- **Explicitly:** Of the initiatives scored, which ones can produce demonstrable results within the stated time constraint (e.g., "6 months")? State this directly.

Name specific initiatives and the logic that connects them. Do not produce a generic "start with quick wins then move to strategic bets" statement — connect it to the specific initiatives and their interdependencies.

**For large-system dependencies** (CRM replacements, platform migrations): Name the contingency if the dependency is delayed. Large-system programs almost always have cost overruns and schedule slippage. The sequencing recommendation should not collapse if the dependency is 3 months late.

---

## Step 5: What We're Explicitly Not Doing and Why

Name the deprioritized initiatives and give the specific rationale for each.

This section exists for three reasons:
1. It respects the work that went into generating these ideas — every deprioritized idea deserves a real explanation
2. It prevents re-litigation — "we decided not to do X because [specific reason]" is harder to challenge than silence
3. It often includes the condition under which deprioritization would be revisited — "not now, but worth revisiting if [condition]"

Format as: **[Initiative] — [Reason for deprioritization] — [Condition that would change this]**

Example: "Proactive complaint prediction model — While strategically valuable, this requires clean structured complaint data that the organization doesn't yet have, making feasibility a 1 at this time. Revisit when the Voice of Customer program (which is recommended as a Quick Win) has 12 months of standardized data."

---

## Output Format

**Section 1: Scoring Table**
All initiatives with dimension scores and composite scores.

**Section 2: Strategic Roadmap**
Three-phase roadmap (Value Realization / Strategic Enablers / Growth Accelerators) with initiatives placed, phase descriptions, and Change Management through-line.

**Section 3: Sequencing Recommendation**
Narrative connecting specific initiatives in time, with dependency logic.

**Section 4: What We're Not Doing (and Why)**
Initiative-by-initiative deprioritization rationale with revisit conditions.

**Section 5: Portfolio Risk Assessment**
What could make this portfolio recommendation wrong? (3 risks: one strategic, one delivery, one external) What's the early warning signal for each?

**Section 6: Team Composition Check**
A brief note on what capability types the recommended portfolio requires — and whether the stated team is likely to have them. A 12-person team that includes no data scientists cannot deliver an ML-driven initiative at the same feasibility score as one that does. If team composition is unknown, flag the assumption and the risk: "Feasibility scores assume your team includes [X, Y, Z] capability. If any are absent, drop the feasibility score for [initiative] by 1-2 points."
