When the user wants to define GTM metrics, build a metrics dashboard, measure pipeline efficiency, or track AI product performance. Also use when the user mentions 'GTM metrics,' 'revenue latency,' 'pipeline metrics,' 'TTFV,' 'time-to-first-value,' 'data health,' 'attribution,' 'conversion rate,' 'CAC,' 'LTV,' 'NRR,' 'GTM dashboard,' 'magic number,' 'pipeline velocity,' or 'funnel metrics.' This skill covers GTM measurement from metric selection through dashboard design, including AI-specific cost metrics, attribution models, and weekly review cadences. Do NOT use for technical implementation, code review, or software architecture.
71
63%
Does it follow best practices?
Impact
Pending
No eval scenarios have been run
Passed
No known issues
Optimize this skill with Tessl
npx tessl skill review --optimize ./packages/skills-catalog/skills/(gtm)/gtm-metrics/SKILL.mdYou are an expert in GTM measurement, dashboard architecture, and performance analytics for AI-native products. You understand the critical differences between traditional SaaS metrics and AI product metrics, including usage-based consumption tracking, AI cost-of-revenue dynamics, and outcome-based pricing measurement. You help founders and revenue leaders select the right metrics, build actionable dashboards, design attribution models, and run weekly review cadences that drive decisions. You know that the median B2B SaaS growth rate has settled to 26% in 2025-2026 while CAC has risen 14% to $2.00 per new ARR dollar, making measurement discipline the difference between efficient growth and cash burn.
Gather this context before building any metrics framework, dashboard, or measurement plan:
| Metric | Definition | How to Calculate | Target |
|---|---|---|---|
| ARR / MRR | Recurring revenue | Sum of active subscription revenue | Growth rate benchmarks below |
| Net New ARR | New minus churned | New ARR + Expansion - Churned ARR | Positive every quarter |
| Revenue Latency | Days from first signal to closed deal | Median days first-touch to closed-won | <30d SMB, <90d mid-market, <180d enterprise |
| Expansion Revenue % | New ARR from existing customers | Expansion ARR / Total New ARR | >40% at scale ($50M+ ARR companies ~60%) |
| Metric | How to Calculate | Target |
|---|---|---|
| CAC | Total S&M spend / New customers | Varies by segment |
| CAC Payback | CAC / (ARR per customer * Gross Margin) | <8 months (median 8.6; top performers 5-7) |
| Magic Number | Net New ARR (qtr) / S&M Spend (prior qtr) | >0.75 efficient, >1.0 excellent, <0.5 red flag |
| LTV:CAC Ratio | (ARPA * Margin * Lifetime) / CAC | >3:1 healthy, >5:1 may be under-investing |
| Burn Multiple | Net Burn / Net New ARR | <2x good, <1x excellent, >3x concerning |
| Metric | How to Calculate | Target |
|---|---|---|
| Pipeline Coverage | Pipeline value / Period quota | 3-4x sales-led, 2-3x PLG |
| Pipeline Velocity | (Qualified Opps * Deal Size * Win Rate) / Cycle Length | Increasing QoQ |
| Pipeline per Rep | Total pipeline / Quota-carrying reps | Track trend, not absolute |
| Slippage Rate | Deals moved out / Total deals in forecast | <15% weekly |
| Metric | How to Calculate | Target |
|---|---|---|
| NRR | (Start MRR + Expansion - Contraction - Churn) / Start MRR | >106% median; >120% best-in-class |
| GRR | (Start MRR - Contraction - Churn) / Start MRR | >90%; >94% at scale |
| Logo Churn | Customers lost / Customers at start | <2% monthly SMB, <1% mid-market |
| TTFV | Median time from signup to first value event | <15 min self-serve, <1 day sales-led |
| ARR Band | Median NRR | Top Quartile | Notes |
|---|---|---|---|
| $1-3M | ~90% | 94% | Focus on finding high-retention segments |
| $3-15M | ~95% | 99% | Expansion motions starting |
| $15-30M | ~100% | 105%+ | Expansion should offset churn |
| $50-100M | ~110% | 120%+ | Expansion revenue exceeds new logos |
| $100M+ | ~115% | 130%+ | Aggressive expansion expected |
| ARR Band | Median Growth | Top Quartile |
|---|---|---|
| <$1M | 100%+ | 200%+ |
| $1-5M | 80-100% | 150%+ |
| $5-20M | 50-80% | 100%+ |
| $20-50M | 30-50% | 70%+ |
| $100M+ | 20-30% | 40%+ |
Visitor --> Signup (3-5%) --> Activation (30-40%) --> Conversion (5-8%) --> Expansion (NRR 110-120%)PLG-specific metrics: PQL conversion rate, time-to-activation (<15 min target), feature adoption breadth (core features used in first 14 days), viral coefficient (>0.3 target).
Signal --> Outreach (3-5% reply) --> Meeting (50%) --> Demo (60%) --> Pilot (40%) --> Close (30%)Sales-led specific: ACV trend, sales cycle length (median days), win rate by segment, pipeline created per rep per month, quota attainment distribution.
Signal --> AI Qualification (10-15%) --> Human Meeting (50%) --> Close (35%)Agent-led specific: cost per meeting booked, cost per qualified lead, AI outreach ROI (revenue from AI pipeline / AI cost), send-to-reply ratio, human-to-AI leverage ratio.
AI products carry cost structures that traditional SaaS metrics miss. These supplementary metrics are essential for AI-native businesses.
| Metric | How to Calculate | Target |
|---|---|---|
| AI Cost of Revenue | Inference + compute cost / Revenue | <20% of revenue |
| Cost per AI Action | Total AI compute / Actions generated | Decreasing over time |
| ROAI | AI-attributed revenue / (Inference + compute overhead) | >10x for high performers |
| Gross Margin after AI | (Revenue - COGS - AI compute) / Revenue | >70% (vs. ~80% pure SaaS) |
42% of SaaS companies use consumption-based pricing in 2025 (up from 29% in 2023). When pricing is usage-based, supplement ARR metrics with:
| Metric | Why It Matters |
|---|---|
| Committed vs. Consumed ARR | Gap indicates pricing misalignment or under-adoption |
| Usage Growth Rate | Leading indicator of expansion revenue |
| Overage Frequency | Signals pricing tier design quality |
| Unit Economics per Consumption Unit | Revenue minus cost per unit; must be positive and improving |
| NRR by Cohort (usage-based only) | Separates usage-driven expansion from seat expansion |
| SaaS Metric | AI Difference | Additional AI Metric |
|---|---|---|
| Gross margin (~80%) | AI inference lowers to 60-75% | Track AI cost of revenue separately |
| DAU/MAU | Usage is task-driven, not session-driven | Task completion rate, actions per session |
| Feature adoption | AI features are singular and deep | Outcome success rate per AI action |
| Time-on-platform | Less time can mean more value | Time-saved-per-task |
| Per-seat revenue | Consumption pricing varies by user | Revenue per consumption unit |
Bad CRM data makes every other metric unreliable. Quantify data trustworthiness before trusting pipeline reports.
Data Health Score = (Completeness * 0.35) + (Accuracy * 0.30) + (Recency * 0.20) + (Consistency * 0.15)| Component | Weight | What It Measures |
|---|---|---|
| Completeness | 35% | % of required fields populated per record |
| Accuracy | 30% | % of data points verified against enrichment sources |
| Recency | 20% | % of records updated within 90 days |
| Consistency | 15% | % of records matching format standards |
| Score | Grade | Action |
|---|---|---|
| 90-100% | A | Maintain current enrichment cadence |
| 80-89% | B | Schedule enrichment refresh for lowest-scoring segments |
| 70-79% | C | Pipeline metrics may be unreliable; run enrichment sprint |
| Below 70% | F | Stop trusting pipeline reports; full data cleanup required |
B2B data decays at 2.1% monthly on average. Required enrichment refresh cadence: contact email/phone every 90 days, firmographics every 90 days, intent signals weekly or real-time, ICP scores recalculated on any underlying data refresh.
Attribution answers "what caused the deal?" Getting it right determines where you invest next.
| Model | How It Works | Best For | Limitation |
|---|---|---|---|
| First-touch | 100% to first interaction | Top-of-funnel channel effectiveness | Ignores nurture and closing touches |
| Last-touch | 100% to final interaction | Bottom-of-funnel conversion analysis | Ignores awareness investment |
| Linear | Equal credit to all touchpoints | Simple fairness | Treats blog visit same as demo request |
| U-shaped | 40% first, 40% last, 20% middle | B2B with clear awareness-to-conversion journey | Undervalues mid-funnel |
| W-shaped | 30/30/30/10 (first/lead/opp/rest) | B2B with defined marketing-to-sales handoff | Requires clear CRM stage definitions |
| Time-decay | Increasing credit toward conversion | Long sales cycles | Undervalues early brand investment |
| AI-driven | ML determines credit dynamically | Orgs with 500+ conversions | Black box; requires data maturity |
| Stage | Model | Why |
|---|---|---|
| Pre-revenue / <$1M | First-touch | Know which channels generate any pipeline |
| $1-5M | U-shaped | Credits awareness and conversion, most actionable |
| $5-20M | W-shaped | Marketing-to-sales handoff stages worth measuring |
| $20M+ | Time-decay or AI-driven | Enough data; long cycles justify recency weighting |
| PLG (any stage) | Product-touch | Attribute to in-product actions, not just marketing |
Set lookback to match your sales cycle: 90 days for SMB, 180 days for mid-market, 365 days for enterprise. Run parallel first-touch and multi-touch models for 2 quarters to calibrate. Review quarterly.
| Challenge | Mitigation |
|---|---|
| AI SDR touches invisible to buyers | Tag AI-generated touches with source=AI-SDR in CRM |
| Multi-channel AI sequences | Track channel and sequence membership, not just "AI outreach" |
| Influence vs. creation confusion | Separate "source" from "influence" attribution |
| Dark social (Slack, Discord, DMs) | Ask "how did you hear about us?" in demo forms |
Tier 1: Board (5-7 metrics, monthly) - ARR + Net New ARR waterfall, NRR, CAC Payback, Burn Multiple, Pipeline Coverage, Magic Number, Cash Runway.
Tier 2: Executive (10-12 metrics, weekly) - Pipeline created, pipeline by stage, win rate by segment, deal size trend, sales cycle length, quota attainment by rep, NRR by cohort, CAC by channel, TTFV, data health score, slippage rate.
Tier 3: Operator (15-25 metrics, daily) - Activity (emails, calls, meetings booked), pipeline (new opps, stage movements), response (speed-to-lead, follow-up rate), conversion (stage-by-stage rates), quality (ICP fit distribution), AI ops (AI messages, AI reply rate, cost per meeting).
| Tool | Best For | Cost |
|---|---|---|
| HubSpot Dashboards | Teams already on HubSpot | Included |
| Metabase | SQL-native, self-hosted | Free |
| Looker | Enterprise-grade, governed | $$$ |
| Mode | SQL + Python + viz | $$ |
| Google Sheets | Solo founders, pre-revenue | Free |
| Anti-Pattern | Fix |
|---|---|
| 50+ metrics on one screen | Limit to tier-appropriate count |
| Vanity metrics without context | Every metric needs benchmark, target, or trend line |
| Manual data entry | All metrics from system-of-record APIs |
| No dashboard owner | Named owner + review schedule required |
| Snapshot without trend | Always show trailing 4-week or 13-week trend |
Maintain a 60/40 balance: 60% leading indicators (what is about to happen) and 40% lagging indicators (what already happened).
| Indicator | Predicts | If Declining |
|---|---|---|
| Pipeline created this week | Revenue 1-2 quarters out | Increase top-of-funnel investment |
| Meeting conversion rate | Win rate next quarter | Audit qualification and demo quality |
| Speed-to-lead | Inbound conversion rate | Fix routing (<5 min target) |
| Product activation rate | Free-to-paid conversion | Audit onboarding flow |
| Sequence reply rate | Meeting volume next month | Refresh messaging and targeting |
| Feature adoption depth | NRR next quarter | Proactive CS intervention |
| Champion engagement frequency | Deal probability | Deal is at risk if champion goes quiet |
Revenue, win rate, CAC/payback, NRR/GRR, LTV:CAC, burn multiple, quota attainment distribution. Review monthly or quarterly.
Revenue (lagging)
<-- Win Rate
<-- Demo Quality Score (leading)
<-- ICP Fit Score of Pipeline (leading)
<-- Pipeline Volume
<-- Meetings Booked (leading)
<-- Outreach Volume + Reply Rate (leading)
<-- Deal Size
<-- Multi-threading Depth (leading)The single most important GTM operating ritual. Every metric from system-of-record data. No hand-edited slides.
| Time | Section | Content |
|---|---|---|
| 0-5 min | Scorecard | Walk 5-7 metrics: green/yellow/red vs. targets |
| 5-15 min | Pipeline | New pipeline, stage movements, slippage, forecast changes |
| 15-20 min | Leading indicators | Inbound volume, outreach metrics, meeting conversion |
| 20-25 min | Deals at risk | Stalled deals, blockers, help requests |
| 25-35 min | Actions | 2-3 specific actions with owners and deadlines |
| 35-45 min | Deep-dive | One strategic topic per week (rotating) |
| Metric | This Week | Last Week | 4-Wk Avg | Target | Status | Owner |
|---|---|---|---|---|---|---|
| Pipeline created | $X | $X | $X | $X | G/Y/R | Name |
| Meetings booked | X | X | X | X | Name | |
| Win rate (30d) | X% | X% | X% | X% | Name | |
| Cycle length | Xd | Xd | Xd | Xd | Name | |
| Slippage rate | X% | X% | X% | <15% | Name | |
| Speed-to-lead | Xm | Xm | Xm | <5m | Name |
NRR/retention analysis (cohort curves, churn reasons, expansion pipeline), CAC/efficiency review (CAC by channel, payback trend, Magic Number), data health audit (CRM completeness, enrichment gaps), competitive update (pricing, positioning, feature changes).
ICP refresh (win/loss analysis, drift detection, scoring update), funnel benchmarking (stage conversions vs. industry), attribution model review (channel ROI, budget allocation), GTM motion evaluation (sales-led vs. PLG vs. agent performance).
Product-Qualified Leads replace MQLs in product-led and hybrid motions. Score on product usage instead of content downloads.
PQL Score = (Usage Signals * 0.50) + (Fit Signals * 0.30) + (Intent Signals * 0.20)| Score | Tier | Action |
|---|---|---|
| 80-100 | Hot | Route to AE, respond within 4 hours |
| 60-79 | Warm | Sales-assist sequence (email + SDR) |
| 40-59 | Nurture | In-app messaging + drip emails |
| Below 40 | Self-serve | No sales touch; optimize product experience |
PQL-to-customer conversion: 5-15% (vs. 1-3% MQL-to-customer). Signal strength is higher because product usage requires effort that content downloads do not.
| Concept | Key Number or Rule |
|---|---|
| CAC Payback benchmark | Median 8.6 months; top performers 5-7 |
| Magic Number threshold | >0.75 efficient, >1.0 excellent, <0.5 red flag |
| Pipeline coverage | 3-4x sales-led, 2-3x PLG |
| NRR median (2025) | 106% across B2B SaaS |
| NRR best-in-class | >120% (130%+ at $100M+ ARR) |
| B2B SaaS median growth | 26% in 2025 |
| CAC trend | Up 14% to $2.00 per new ARR dollar |
| TTFV target | <15 min self-serve, <1 day sales-led |
| Revenue latency | <30d SMB, <90d mid-market |
| Data health target | >85%; below 70% is unreliable |
| Data decay rate | 2.1% monthly |
| Leading/lagging balance | 60% leading, 40% lagging |
| Weekly review | 30-45 min, every week, no exceptions |
| Attribution lookback | 90d SMB, 180d mid-market, 365d enterprise |
| PQL conversion | 5-15% (vs. 1-3% MQL) |
| Usage-based adoption | 42% of SaaS companies in 2025 |
| AI gross margin target | >70% (vs. ~80% pure SaaS) |
| Expansion at scale | >40% of new ARR from existing customers |
| Slippage target | <15% weekly |
| Speed-to-lead | <5 minutes |
| Skill | When to Cross-Reference |
|---|---|
| ai-pricing | Measuring pricing model impact on revenue metrics; usage-based pricing instrumentation |
| expansion-retention | NRR improvement plans, churn analysis, expansion playbooks |
| sales-motion-design | Redesigning stages, qualification, or handoffs based on funnel data |
| ai-cold-outreach | Outreach performance: reply rate, cost per meeting, AI SDR ROI |
| solo-founder-gtm | Prioritizing 3-5 metrics before building a full dashboard |
| gtm-engineering | Metric collection infrastructure, CRM automation, data pipelines |
| ai-sdr | AI SDR performance measurement, agent-led funnel metrics |
| lead-enrichment | Data health remediation, enrichment workflows |
| positioning-icp | When attribution reveals positioning or ICP changes needed |
| content-to-pipeline | Content attribution, connecting top-of-funnel to pipeline |
906a57d
If you maintain this skill, you can claim it as your own. Once claimed, you can manage eval scenarios, bundle related skills, attach documentation or rules, and ensure cross-agent compatibility.