Production Readiness Assessment
Learning Objectives
Conduct comprehensive production readiness assessments
Evaluate readiness across all required dimensions
Make defensible go/no-go recommendations
Identify gaps requiring remediation before launch
Resist pressure to launch before genuine readiness
PRODUCTION READINESS PILLARS
PILLAR 1: TECHNICAL READINESS
├── Platform stability proven
├── Security validated
├── Performance confirmed
├── Integration complete
├── Resilience tested
└── Weight: 25%
PILLAR 2: OPERATIONAL READINESS
├── Support infrastructure ready
├── Processes documented
├── Staff trained
├── Monitoring operational
├── Incident response tested
└── Weight: 25%
PILLAR 3: LEGAL/REGULATORY READINESS
├── Legal framework enacted
├── Regulatory approvals obtained
├── Compliance mechanisms operational
├── Privacy framework implemented
├── Consumer protection ready
└── Weight: 20%
PILLAR 4: ECOSYSTEM READINESS
├── Bank partners operational
├── Merchant network adequate
├── User acquisition ready
├── Support ecosystem prepared
├── Communications ready
└── Weight: 20%
PILLAR 5: ORGANIZATIONAL READINESS
├── Team scaled for production
├── Governance functioning
├── Budget secured
├── Leadership committed
├── Contingency plans ready
└── Weight: 10%
SCORING:
Each pillar: 0-100 points
Weighted total: Must exceed 80
No pillar below 60
Any "critical" item missing = not ready
```
TECHNICAL READINESS CHECKLIST
PLATFORM STABILITY:
□ 99.9%+ uptime in pilot (last 90 days)
□ No P1 incidents in last 30 days
□ All P2 incidents resolved within SLA
□ Known issues documented and acceptable
□ No workarounds required for core functionality
Score: ___ / 20
SECURITY:
□ External penetration test passed
□ All critical/high findings remediated
□ Security monitoring operational
□ Incident response tested
□ Cryptographic review complete
□ Access controls audited
Score: ___ / 20
PERFORMANCE:
□ Load tested at 10x expected peak
□ Latency within SLA at peak
□ No degradation over 72-hour endurance test
□ Auto-scaling validated
□ Capacity headroom confirmed
Score: ___ / 20
INTEGRATION:
□ All bank integrations production-ready
□ Payment network connections tested
□ Third-party services contracted and tested
□ API stability confirmed
□ Data feeds operational
Score: ___ / 20
RESILIENCE:
□ DR tested successfully
□ RTO/RPO requirements met
□ Failover automatic and tested
□ Backup/restore validated
□ Chaos engineering conducted
Score: ___ / 20
- Security penetration test passed
- DR tested successfully
- No P1 incidents in 30 days
OPERATIONAL READINESS CHECKLIST
SUPPORT INFRASTRUCTURE:
□ Tier 1/2/3 support staffed
□ Support tools deployed
□ Knowledge base complete
□ Escalation paths defined
□ SLAs established
Score: ___ / 20
PROCESSES:
□ All operational processes documented
□ Runbooks for common issues
□ Change management operational
□ Release process tested
□ Incident management proven
Score: ___ / 20
TRAINING:
□ Support staff trained and certified
□ Operations team trained
□ Bank partner staff trained
□ Merchant support trained
□ Ongoing training program established
Score: ___ / 20
MONITORING:
□ Real-time dashboards operational
□ Alerting configured and tested
□ On-call rotation established
□ Performance monitoring active
□ Business metrics tracking ready
Score: ___ / 20
INCIDENT RESPONSE:
□ Incident classification defined
□ Response procedures documented
□ War room capability ready
□ Communication templates prepared
□ Post-incident review process established
Score: ___ / 20
- Support staffed for expected volume
- Incident response tested
- Monitoring operational
SUPPORT CAPACITY MODEL
EXPECTED SUPPORT VOLUME:
Phase 1 (Launch): High support ratio
├── Users: 10,000-50,000
├── Ticket ratio: 10-15% of users/month
├── Expected tickets: 1,000-7,500/month
├── Support staff: 15-30 agents
└── Scaling: Weekly reassessment
Phase 2 (Growth): Moderate support ratio
├── Users: 50,000-200,000
├── Ticket ratio: 5-8% of users/month
├── Expected tickets: 2,500-16,000/month
├── Support staff: 25-60 agents
└── Scaling: Monthly reassessment
Phase 3 (Maturity): Low support ratio
├── Users: 200,000+
├── Ticket ratio: 2-4% of users/month
├── Expected tickets: 4,000-8,000/month
├── Support staff: 40-80 agents
└── Scaling: Quarterly reassessment
CAPACITY REQUIREMENTS:
├── Agent capacity: ~150 tickets/agent/month
├── Tier 2 ratio: 1:5 (Tier 2 : Tier 1)
├── Tier 3 ratio: 1:10 (Tier 3 : Tier 1)
├── Management ratio: 1:10 (Manager : Agents)
└── Buffer: 25% above calculated need
---
LEGAL/REGULATORY READINESS CHECKLIST
LEGAL FRAMEWORK:
□ Primary legislation enacted
□ Secondary regulations published
□ Central bank authority explicit
□ Legal tender status defined
□ Liability framework established
Score: ___ / 25
REGULATORY APPROVALS:
□ Central bank internal approval
□ Finance ministry approval (if required)
□ Data protection authority clearance
□ Consumer protection compliance confirmed
□ AML/CFT framework approved
Score: ___ / 25
COMPLIANCE MECHANISMS:
□ AML monitoring operational
□ Reporting systems tested
□ Sanctions screening active
□ Suspicious activity detection working
□ Regulatory reporting ready
Score: ___ / 25
CONSUMER PROTECTION:
□ Terms of service finalized
□ Privacy policy published
□ Complaint mechanism operational
□ Dispute resolution process ready
□ Error resolution procedures tested
Score: ___ / 25
- Primary legislation enacted
- Central bank authority explicit
- AML monitoring operational
ECOSYSTEM READINESS CHECKLIST
BANK PARTNERS:
□ Minimum 3 banks live and tested
□ Bank integration stable (>99.9% uptime)
□ Bank staff trained
□ Customer acquisition processes ready
□ Bank marketing materials approved
Score: ___ / 25
MERCHANT NETWORK:
□ Target merchant count achieved (minimum 500)
□ Merchant sectors diversified
□ Merchant integration tested
□ Merchant support ready
□ Merchant incentive program operational
Score: ___ / 25
USER ACQUISITION:
□ Marketing plan approved
□ Launch communications ready
□ Media briefings prepared
□ Influencer/ambassador program ready
□ User education materials complete
Score: ___ / 25
COMMUNICATIONS:
□ Launch announcement drafted
□ FAQ prepared
□ Crisis communication templates ready
□ Social media strategy defined
□ Press kit complete
Score: ___ / 25
- Minimum 3 banks live
- Minimum 500 merchants
- Crisis communication ready
GO/NO-GO DECISION MATRIX
SCORING SUMMARY:
├── Technical Readiness: ___ / 100 × 0.25 = ___
├── Operational Readiness: ___ / 100 × 0.25 = ___
├── Legal/Regulatory: ___ / 100 × 0.20 = ___
├── Ecosystem: ___ / 100 × 0.20 = ___
├── Organizational: ___ / 100 × 0.10 = ___
└── TOTAL WEIGHTED SCORE: ___ / 100
CRITICAL ITEMS CHECK:
□ All critical items satisfied? YES / NO
DECISION THRESHOLDS:
SCORE 85+ AND ALL CRITICAL ITEMS:
├── Recommendation: GO
├── Confidence: High
└── Action: Proceed to launch
SCORE 75-84 AND ALL CRITICAL ITEMS:
├── Recommendation: CONDITIONAL GO
├── Confidence: Medium
├── Conditions: Specific items to address
└── Action: Address conditions, reassess in 2 weeks
SCORE 65-74 OR MISSING 1-2 CRITICAL ITEMS:
├── Recommendation: DELAY
├── Confidence: Low
├── Gap analysis required
└── Action: Remediation plan, reassess in 4-8 weeks
SCORE <65 OR MISSING 3+ CRITICAL ITEMS:
├── Recommendation: NOT READY
├── Confidence: N/A
├── Major gaps present
└── Action: Comprehensive remediation, reassess in 12+ weeks
ANY PILLAR <60:
├── Recommendation: DELAY
├── Regardless of total score
└── Action: Address pillar deficit
```
HONEST READINESS ASSESSMENT PRINCIPLES
PRINCIPLE 1: EVIDENCE OVER OPINION
├── Every score backed by objective evidence
├── "We think it works" ≠ "We tested and it works"
├── Require documentation for claims
└── Third-party validation for critical items
PRINCIPLE 2: RECENCY MATTERS
├── Tests from 6 months ago may not be valid
├── Retest critical items within 30 days of launch
├── Require current state, not historical state
└── Conditions change; reassess regularly
PRINCIPLE 3: WORST CASE CONSIDERATION
├── "What if this fails on day 1?"
├── Plan for maximum load, not average
├── Assume problems will occur
└── Assess recovery capability, not just prevention
PRINCIPLE 4: INDEPENDENCE
├── Assessment team separate from project team
├── No penalty for identifying gaps
├── Incentives aligned with honest assessment
└── Senior leadership receives unfiltered report
PRINCIPLE 5: DELAY IS LEGITIMATE
├── "We're not ready" is a valid conclusion
├── Sunk cost is not a reason to proceed
├── Reputation damage from failed launch exceeds delay cost
└── Better to delay than to fail publicly
```
LAUNCH PRESSURE TACTICS AND RESPONSES
PRESSURE 1: "WE ANNOUNCED A DATE"
├── Response: Announcements can be updated
├── Reality: Failed launch damages more than delay
├── Action: Prepare revised communication
└── Message: "We're taking extra time to ensure quality"
PRESSURE 2: "COMPETITION IS LAUNCHING"
├── Response: Their launch is irrelevant to our readiness
├── Reality: Rushed launches usually fail (see Nigeria)
├── Action: Document competitor status honestly
└── Message: "Quality over speed"
PRESSURE 3: "STAKEHOLDERS ARE IMPATIENT"
├── Response: Stakeholders prefer working CBDC to failed CBDC
├── Reality: Stakeholder impatience < stakeholder disappointment
├── Action: Brief stakeholders on realistic status
└── Message: "Additional preparation time protects everyone"
PRESSURE 4: "WE'VE SPENT SO MUCH ALREADY"
├── Response: Sunk cost fallacy—spent money is gone
├── Reality: More spending on failed launch wastes more
├── Action: Present cost of failure vs. cost of delay
└── Message: "Investment protection requires readiness"
PRESSURE 5: "THE TEAM IS READY TO MOVE ON"
├── Response: Team morale ≠ system readiness
├── Reality: Team will suffer more from failed launch
├── Action: Acknowledge team effort, maintain standards
└── Message: "Your hard work deserves successful launch"
```
READINESS ESCALATION FRAMEWORK
LEVEL 1: PROGRAM DIRECTOR
├── Authority: Minor gap remediation
├── Delay authority: Up to 2 weeks
├── Escalate: Significant gaps or pressure
└── Document: Weekly status report
LEVEL 2: STEERING COMMITTEE
├── Authority: Moderate gap decisions
├── Delay authority: Up to 8 weeks
├── Escalate: Major gaps or external pressure
└── Document: Formal assessment report
LEVEL 3: GOVERNOR/BOARD
├── Authority: Go/no-go decision
├── Delay authority: Any duration
├── Escalate: Fundamental viability questions
└── Document: Board-level recommendation
ESCALATION TRIGGERS:
├── Total score <75
├── Any pillar <60
├── Missing critical items
├── External pressure to override assessment
├── Team concerns about readiness
└── Significant new risks identified
```
✅ Rushed launches fail: Every major CBDC failure launched before achieving genuine readiness.
✅ Structured assessment prevents blind spots: Comprehensive checklists catch issues that informal review misses.
✅ Independence matters: Assessment teams with project team reporting bias miss problems.
⚠️ What score threshold is "enough": 80% readiness might be sufficient or insufficient depending on context.
⚠️ Whether all dimensions equally important: Weightings are judgment calls.
🔴 Pressure-driven launches: Political or timeline pressure that overrides honest assessment.
🔴 "Good enough" thinking: Accepting marginal readiness because delay is inconvenient.
🔴 Confirmation bias: Seeing readiness because you want to see readiness.
Assignment: Conduct a mock production readiness assessment for a CBDC program.
- Complete all five pillar checklists with realistic scores
- Document evidence for each score
- Identify all critical item status
- Calculate weighted total score
- Make and justify go/no-go recommendation
- If not ready: gap analysis and remediation plan
Time investment: 3-4 hours
Q1: What weighted score threshold indicates "GO" for production?
A) 65+ B) 75+ C) 85+ with all critical items D) 95+
Answer: C
Q2: If one pillar scores 55 but total weighted score is 82, what is the recommendation?
A) GO B) Conditional GO C) DELAY D) Not enough information
Answer: C - Any pillar below 60 requires delay regardless of total score.
Q3: How should you respond to "we announced a launch date" pressure?
A) Launch anyway B) Ignore the announcement C) Explain failed launch damages more than delay D) Resign
Answer: C
Q4: Who has final go/no-go authority?
A) Program Director B) Steering Committee C) Governor/Board D) Technology team
Answer: C
Q5: What is the minimum number of banks required for ecosystem readiness?
A) 1 B) 3 C) 10 D) All banks in country
Answer: B
End of Lesson 14
Key Takeaways
Five pillars must all be ready
: Technical, operational, legal, ecosystem, and organizational—weakness in any pillar threatens success.
Critical items are non-negotiable
: Missing any critical item means not ready, regardless of total score.
Evidence over opinion
: Every readiness claim must be backed by objective evidence, not team confidence.
Delay is legitimate
: Better to delay and succeed than launch and fail. Reputation damage from failure exceeds delay cost.
Resist pressure systematically
: Have frameworks for responding to common pressure tactics. ---