Building Your Monitoring Dashboard - Practical Implementation | XRP Network Metrics | XRP Academy - XRP Academy
3 free lessons remaining this month

Free preview access resets monthly

Upgrade for Unlimited
Skip to main content
intermediate60 min

Building Your Monitoring Dashboard - Practical Implementation

Learning Objectives

Design a dashboard architecture balancing comprehensiveness with manageability

Set up data collection workflows using free tools and APIs

Create visualization templates for clear metric communication

Establish alert systems that flag significant changes

Integrate monitoring into sustainable daily/weekly/monthly routines

You've learned to analyze transactions, addresses, DEX activity, fees, ecosystem health, ODL, and growth trajectories. Now what?

THE IMPLEMENTATION GAP:

COMMON FAILURE:
├── Complete course
├── Feel informed
├── Return to random Twitter analysis
├── Never actually track metrics systematically
└── Knowledge unused

BETTER OUTCOME:
├── Complete course
├── Build monitoring system
├── Track metrics consistently
├── Update views based on data
└── Knowledge applied

This lesson bridges the gap. No more theory—practical implementation that you'll actually use.


Building sustainable monitoring:

DASHBOARD DESIGN PRINCIPLES:

1. MANAGEABILITY OVER COMPREHENSIVENESS

1. AUTOMATION WHERE POSSIBLE

1. CLEAR ORGANIZATION

1. ACTIONABLE ALERTS

1. UPDATE CADENCE MATCHING METRIC NATURE

Choosing your core dashboard metrics:

RECOMMENDED CORE METRICS (15-20 total):

ACTIVITY PILLAR (4 metrics):
├── Daily transactions (spam-filtered)
├── Daily payment volume (USD)
├── Transaction velocity
├── Fee burn rate

ADOPTION PILLAR (4 metrics):
├── Monthly Active Addresses (MAA)
├── New accounts (weekly average)
├── 30-day retention rate
├── DAU/MAU ratio

LIQUIDITY PILLAR (4 metrics):
├── AMM TVL
├── DEX depth (XRP/USD at 2%)
├── Average spread (major pairs)
├── DEX volume (weekly)

ECOSYSTEM PILLAR (4 metrics):
├── Trust line growth (monthly)
├── RLUSD metrics (supply, trust lines)
├── Developer activity (GitHub proxy)
├── ODL volume estimate (with range)

GROWTH INDICATORS (2-3 metrics):
├── MAA growth rate (MoM)
├── New account quality (retention of new)
├── Key leading indicator of your choice

Options for building your dashboard:

TOOL OPTIONS BY TECHNICAL LEVEL:

NON-TECHNICAL (Spreadsheet-based):
├── Google Sheets / Excel
├── Manual + some import formulas
├── Free, familiar, flexible
├── Recommended for most users
└── This lesson focuses here

SEMI-TECHNICAL (Low-code):
├── Notion databases
├── Airtable + integrations
├── Google Sheets + Apps Script
├── More automation possible
└── Moderate setup effort

TECHNICAL (Code-based):
├── Python + APIs + database
├── Custom web dashboard
├── Full automation possible
├── Significant setup effort
└── Best for programmatic analysis

RECOMMENDATION:
├── Start with spreadsheet
├── Proves the concept works
├── Upgrade later if needed
├── Don't over-engineer initially

For spreadsheet-based tracking:

MANUAL DATA COLLECTION:

DAILY (5-10 minutes):
├── Check XRPScan dashboard
├── Record: Transactions, volume, fee burn
├── Note: Any obvious anomalies
└── Source: XRPScan network stats

WEEKLY (15-20 minutes):
├── All daily metrics averaged
├── Check MAA/WAA (XRPScan or Bithomp)
├── New accounts (weekly total)
├── DEX volume and depth snapshot
├── AMM TVL snapshot
└── Sources: Multiple explorers

MONTHLY (30-45 minutes):
├── Full metric refresh
├── Trust line count
├── RLUSD metrics
├── Developer activity check (GitHub)
├── ODL estimate update
├── Comparative metrics (Stellar)
└── Sources: Comprehensive review

Reducing manual effort:

GOOGLE SHEETS IMPORT OPTIONS:

IMPORTHTML (Limited for crypto):
├── =IMPORTHTML(url, "table", index)
├── Works for some sites
├── Often blocked by crypto sites

IMPORTDATA (CSV feeds):
├── =IMPORTDATA(url)
├── If source provides CSV
├── Limited availability for XRPL

IMPORTXML (Web scraping):
├── =IMPORTXML(url, xpath)
├── Can extract specific data
├── Requires xpath knowledge
├── May break if site changes

API INTEGRATION (Apps Script):
├── More reliable
├── Requires coding
├── Can query XRPL APIs directly
└── Best for automation

PRACTICAL APPROACH:
├── Manual for most metrics (reliable)
├── Automate 2-3 high-frequency metrics if able
├── Use explorer data download features
├── Accept some manual effort

Where to get each metric:

DATA SOURCE REFERENCE:

ACTIVITY METRICS:
├── Transactions: XRPScan network stats
├── Volume: XRPScan or Bithomp
├── Fee burn: XRPScan or calculate from txs
└── Velocity: Calculate (volume/supply)

ADOPTION METRICS:
├── MAA/WAA/DAA: XRPScan accounts section
├── New accounts: XRPScan or Bithomp
├── Retention: Requires tracking cohorts (manual)
├── DAU/MAU: Calculate from DAA/MAA

LIQUIDITY METRICS:
├── AMM TVL: XRPLMeta or direct AMM queries
├── DEX depth: XRPScan or book_offers API
├── Spreads: Calculate from bid/ask
├── DEX volume: XRPScan DEX stats

ECOSYSTEM METRICS:
├── Trust lines: Bithomp or XRPScan
├── RLUSD: Issuer tracking on explorers
├── Developer: GitHub manually or API
├── ODL: Utility Scan or manual estimation

COMPARATIVE:
├── Stellar: StellarExpert
├── Market context: CoinGecko
└── Broader crypto: Various aggregators

Organizing your dashboard:

SPREADSHEET STRUCTURE:

TAB 1: DASHBOARD (Summary View)
├── Current values for all metrics
├── Trend indicators (↑ ↓ →)
├── Alert flags for anomalies
├── Last updated timestamp
└── One-page overview

TAB 2: ACTIVITY DATA
├── Daily/weekly transaction data
├── Volume data
├── Fee data
├── Historical time series
└── Charts

TAB 3: ADOPTION DATA
├── Active address time series
├── New account tracking
├── Retention cohort tables
├── Distribution data
└── Charts

TAB 4: LIQUIDITY DATA
├── DEX metrics time series
├── AMM TVL tracking
├── Spread history
├── Charts

TAB 5: ECOSYSTEM DATA
├── Trust line tracking
├── RLUSD metrics
├── Developer activity log
├── ODL estimates
└── Charts

TAB 6: GROWTH & SCENARIOS
├── Growth rate calculations
├── Scenario probabilities
├── Leading indicator tracking
├── Projection updates

TAB 7: COMPARATIVE
├── XRPL vs Stellar metrics
├── Market context
└── Relative position tracking

The at-a-glance view:

DASHBOARD TAB LAYOUT:

ROW 1-2: Header and Last Updated

ROW 3-8: ACTIVITY PILLAR
| Metric | Current | Prev Week | Trend | Alert |
|--------|---------|-----------|-------|-------|
| Txs (filtered) | 1.2M | 1.1M | ↑9% | |
| Volume (USD) | $85M | $78M | ↑9% | |
| Velocity | 0.8 | 0.78 | → | |
| Fee burn | 8,500 | 7,200 | ↑18% | |

ROW 9-14: ADOPTION PILLAR
| Metric | Current | Prev Week | Trend | Alert |
|--------|---------|-----------|-------|-------|
| MAA | 245K | 238K | ↑3% | |
| New accounts | 52K | 48K | ↑8% | |
| Retention | 32% | 31% | ↑ | |
| DAU/MAU | 0.21 | 0.20 | → | |

[Continue for other pillars...]

ROW 30-35: ALERTS & NOTES
├── Active alerts
├── Key observations
├── Action items
└── Next review date

Key spreadsheet formulas:

USEFUL FORMULAS:

TREND INDICATOR:
=IF(B2>B3, "↑"&TEXT((B2-B3)/B3,"0%"), 
   IF(B2<B3, "↓"&TEXT((B3-B2)/B3,"0%"), "→"))

GROWTH RATE (MoM):
=(Current - Previous) / Previous

CAGR:
=(End/Start)^(1/Years) - 1

7-DAY AVERAGE:
=AVERAGE(B2:B8)

ALERT FLAG:
=IF(ABS((B2-AVERAGE(B2:B31))/STDEV(B2:B31))>2, "⚠️", "")

VELOCITY:
=Volume / CirculatingSupply

DAU/MAU RATIO:
=DAA / MAA

CONDITIONAL FORMATTING:
├── Green: >10% improvement
├── Yellow: -10% to +10%
├── Red: >10% decline
└── Apply to trend columns

What deserves an alert:

ALERT THRESHOLD FRAMEWORK:

TIER 1: MONITOR (Yellow)
├── >1.5 standard deviations from 30-day mean
├── Or: >25% change week-over-week
├── Action: Note for review

TIER 2: INVESTIGATE (Orange)
├── >2 standard deviations from mean
├── Or: >50% change week-over-week
├── Action: Same-day investigation

TIER 3: URGENT (Red)
├── >3 standard deviations from mean
├── Or: >100% change (double or half)
├── Action: Immediate investigation

ALERT EXAMPLES:
├── MAA drops 30% in a week: Tier 2 (investigate)
├── Transactions spike 200%: Tier 3 (likely spam)
├── AMM TVL up 20%: Tier 1 (monitor, probably good)
├── ODL corridor goes silent: Tier 2 (investigate)

Building alerts into your system:

SPREADSHEET ALERT IMPLEMENTATION:

ALERT FORMULA EXAMPLE:
=IF(ABS((B2-AVERAGE(B2:B31))/STDEV(B2:B31))>3, "🔴",
   IF(ABS((B2-AVERAGE(B2:B31))/STDEV(B2:B31))>2, "🟠",
      IF(ABS((B2-AVERAGE(B2:B31))/STDEV(B2:B31))>1.5, "🟡", "")))

ALERT DASHBOARD SECTION:
├── List all current alerts
├── Show metric, value, threshold, trigger time
├── Clear when investigated and resolved
├── Track alert history

EMAIL NOTIFICATIONS (Advanced):
├── Google Sheets + Apps Script
├── Trigger on data update
├── Send email if alert conditions met
├── Requires Apps Script coding

What to do when alerted:

ALERT RESPONSE WORKFLOW:

STEP 1: VERIFY (5 minutes)
├── Is the data accurate?
├── Check source directly
├── Rule out data error
└── Confirm alert is real

STEP 2: CONTEXTUALIZE (10 minutes)
├── What happened in crypto market?
├── Any news or announcements?
├── Seasonal or cyclical factors?
├── Is this affecting other metrics?

STEP 3: INVESTIGATE (15-30 minutes)
├── Drill into the metric
├── What's driving the change?
├── Is it sustainable or temporary?
├── What does it mean for thesis?

STEP 4: DOCUMENT (5 minutes)
├── Record findings
├── Update scenario probabilities if needed
├── Note any follow-up needed
├── Clear or escalate alert

STEP 5: ACT (If needed)
├── Adjust thesis if warranted
├── Update projections
├── Communicate findings
├── Schedule follow-up monitoring

Quick daily check:

DAILY CHECK (5-10 minutes):

TIMING: Morning or evening, consistent

TASKS:
□ Quick scan of dashboard
□ Note any alerts
□ Record daily metrics if tracking
□ Check crypto news briefly
□ Flag anything for deeper review

WHAT TO LOOK FOR:
├── Any Tier 2+ alerts
├── Major news affecting XRP/XRPL
├── Market-wide movements
├── Anything unusual

OUTPUT:
├── Dashboard updated (if daily tracking)
├── Alerts noted
├── Issues flagged for weekly review
└── Total time: 5-10 minutes

Deeper weekly analysis:

WEEKLY REVIEW (30-45 minutes):

TIMING: Same day each week (e.g., Sunday evening)

TASKS:
□ Update all weekly metrics
□ Calculate week-over-week changes
□ Review any alerts from week
□ Investigate flagged items
□ Check comparative metrics
□ Update trend assessments
□ Note key observations

STRUCTURE:
├── 10 min: Data collection
├── 10 min: Dashboard update
├── 10 min: Alert/anomaly review
├── 10 min: Notes and planning
└── Total: 30-45 minutes

OUTPUT:
├── Dashboard fully updated
├── Weekly summary notes
├── Action items for month
├── Issues for deeper monthly review

Comprehensive monthly analysis:

MONTHLY REVIEW (2-3 hours):

TIMING: First weekend of month

TASKS:
□ Full metric refresh (all metrics)
□ Monthly calculations (growth rates, ratios)
□ Cohort retention update
□ Ecosystem metrics refresh
□ ODL estimate update
□ Comparative analysis update
□ Scenario probability review
□ Growth projection update
□ Monthly summary report

STRUCTURE:
├── 45 min: Complete data collection
├── 30 min: Calculations and updates
├── 30 min: Trend and scenario analysis
├── 30 min: Report writing
├── 15 min: Next month planning
└── Total: 2-3 hours

OUTPUT:
├── Complete dashboard update
├── Monthly analysis report
├── Updated scenario probabilities
├── Revised projections if needed
├── Priorities for coming month

Strategic reviews:

QUARTERLY REVIEW (Half day):
├── Full course deliverable refresh
├── Comprehensive competitive analysis
├── Thesis reassessment
├── Dashboard methodology review
├── Long-term trend analysis
└── Strategy adjustment if needed

ANNUAL REVIEW (Full day):
├── Year-over-year analysis
├── Complete baseline refresh
├── Dashboard redesign if needed
├── Course re-review for updates
├── Full thesis validation/revision
└── Planning for next year

Making monitoring sustainable:

SUSTAINABILITY STRATEGIES:

START SMALL:
├── Begin with 10 metrics, not 20
├── Weekly updates initially
├── Add complexity gradually
├── Build habit before expanding

AUTOMATE WHAT YOU CAN:
├── Even partial automation helps
├── Focus automation on tedious metrics
├── Accept manual for complex analysis
└── Reduce friction where possible

ACCEPT IMPERFECTION:
├── Missed weeks happen
├── Don't abandon system over gaps
├── Some data is better than none
├── Restart after breaks

SET REALISTIC EXPECTATIONS:
├── This is a tool, not a full-time job
├── Match effort to investment size
├── Adjust cadence to your life
├── Quality over quantity

Evolving your system:

IMPROVEMENT CYCLE:

MONTHLY: Small tweaks
├── Adjust thresholds if too many/few alerts
├── Add/remove metrics based on value
├── Refine data collection process

QUARTERLY: Medium changes
├── Reassess metric selection
├── Update comparison networks
├── Improve visualizations
├── Add new leading indicators

ANNUALLY: Major review
├── Full dashboard redesign if needed
├── Update methodology for network changes
├── Incorporate new course learnings
├── Reset baselines and benchmarks

SIGNALS TO IMPROVE:
├── Spending too much time on data collection
├── Missing important developments
├── Alerts not actionable
├── System not informing decisions

✅ Systematic monitoring improves analysis quality

✅ Spreadsheet-based dashboards work for most users

✅ Alert systems catch significant changes

✅ Routine integration enables sustainability

⚠️ Optimal metric selection depends on individual needs

⚠️ Alert thresholds require tuning over time

⚠️ Automation feasibility varies by technical skill

⚠️ Time investment varies by depth desired

📌 Over-engineering initial dashboard (analysis paralysis)

📌 Letting perfect be enemy of good (never starting)

📌 Alert fatigue from too many notifications

📌 Abandoning system after first missed week

A simple dashboard you actually use beats a sophisticated system you abandon. Start with a basic spreadsheet tracking 10-15 metrics on a weekly basis. Build the habit first, then optimize. The goal isn't a beautiful dashboard—it's informed decision-making. If your system helps you understand XRPL better and make better investment decisions, it's working.


Assignment: Build and populate a complete monitoring dashboard using the frameworks from this course.

Requirements:

Part 1: Dashboard Construction (40%)

  • Create spreadsheet with all recommended tabs
  • Include 15-20 metrics across all four pillars
  • Build summary dashboard with trend indicators
  • Implement alert formulas/conditional formatting
  • Populate with current data

Part 2: Automation/Import (20%)

  • Implement at least 2-3 semi-automated data imports
  • OR document manual collection workflow in detail
  • Test and verify data collection process
  • Document data sources for each metric

Part 3: Alert System (20%)

  • Implement alert thresholds for key metrics
  • Test alerts with sample data
  • Document alert response protocol
  • Create alert log/history tracker

Part 4: Routine Integration (20%)

  • Create daily/weekly/monthly checklists

  • Schedule first month of monitoring sessions

  • Document time estimates for each routine

  • Identify potential sustainability issues and mitigation

  • Dashboard completeness and organization (30%)

  • Data accuracy and sources (25%)

  • Alert system functionality (20%)

  • Routine practicality (15%)

  • Documentation quality (10%)

Time investment: 5-6 hours initial setup
Value: This is your actual working system. Unlike other deliverables, you'll use this indefinitely for ongoing XRPL monitoring.


1. Dashboard Design:

Why is "manageability over comprehensiveness" the first design principle for dashboards?

A) Comprehensive dashboards are technically impossible
B) Fewer metrics are always more accurate
C) Sustainable monitoring requires manageable scope; abandoned dashboards provide no value
D) XRPL doesn't have enough metrics to track comprehensively

Correct Answer: C

Explanation: The best system is one you'll actually use. A 15-metric dashboard updated consistently beats a 100-metric dashboard abandoned after a month. Manageability ensures sustainability. Comprehensiveness without maintenance is worthless.


2. Alert Thresholds:

A metric shows 2.5 standard deviations from its 30-day mean. What alert tier should this trigger?

A) No alert—still within normal variation
B) Tier 1 (Monitor)—yellow flag
C) Tier 2 (Investigate)—orange flag requiring same-day investigation
D) Tier 3 (Urgent)—immediate action required

Correct Answer: C

Explanation: The framework defines Tier 2 alerts at >2 standard deviations. 2.5 standard deviations indicates a statistically significant deviation warranting investigation. Not urgent (Tier 3 is >3 SD), but more than just monitoring.


3. Routine Integration:

What is the recommended approach if you miss a scheduled weekly review?

A) Abandon the dashboard—consistency is required for valid tracking
B) Back-fill all missed data before resuming
C) Resume with current data, accept the gap, don't let one miss derail the system
D) Double the next week's analysis to make up for it

Correct Answer: C

Explanation: Missed weeks happen. Sustainability requires accepting imperfection. Resume with current data rather than abandoning (A) or creating unsustainable catch-up work (B, D). Some data is better than none. The habit matters more than perfect completeness.


4. Data Collection:

For most non-technical users, what is the recommended approach to dashboard automation?

A) Build custom API integrations for all metrics
B) Hire a developer to automate everything
C) Use manual collection for most metrics, automate only 2-3 high-frequency items if possible
D) Only use metrics that can be fully automated

Correct Answer: C

Explanation: Manual collection is sustainable for most metrics at weekly/monthly cadence. Full automation requires technical skills and maintenance. Automate high-value items where feasible, but accept manual for most. This balances effort with reliability.


5. Alert Response:

You receive a Tier 2 alert that MAA dropped 35% week-over-week. What is the correct first step?

A) Immediately revise your investment thesis
B) Verify the data accuracy by checking the source directly
C) Ignore it—MAA is a lagging indicator
D) Wait for next week's data to confirm the trend

Correct Answer: B

Explanation: Step 1 in the alert response protocol is verification. Before reacting, confirm the data is accurate (not a collection error or source issue). Then contextualize and investigate. Immediate thesis revision (A) is premature. Ignoring (C) defeats the purpose of alerts. Waiting (D) might be appropriate after verification but isn't the first step.


  • Google Sheets documentation
  • Excel dashboard templates
  • Apps Script for automation
  • Edward Tufte's principles
  • Spreadsheet charting best practices
  • Atomic Habits (James Clear) for routine sustainability

For Next Lesson:
Lesson 15 is the Capstone—building a comprehensive Network Health Report that synthesizes everything you've learned.


End of Lesson 14

Total words: ~6,200
Estimated completion time: 60 minutes reading + 5-6 hours for deliverable (initial setup)

Key Takeaways

1

Start simple

: A 15-metric spreadsheet updated weekly is better than a complex system you won't maintain. Begin minimal, expand later.

2

Match cadence to metric nature

: Daily metrics need frequent updates; monthly metrics don't. Design your routine around metric volatility.

3

Alerts focus attention

: Use statistical thresholds (standard deviations) to flag significant changes. Investigate alerts, don't just collect metrics.

4

Integration is everything

: Block time in your calendar. Make monitoring a habit. Missed consistency destroys the value of tracking.

5

Evolve continuously

: Adjust thresholds, add metrics, refine processes. Your dashboard should improve over time, not stay static. ---