A health score that just tells you who is red is half the value. The full value is using the score to predict churn risk on a rolling 30/60/90-day horizon and intervening before the customer asks to cancel.
Most SaaS teams stop at the dashboard. They build the weights, color-code accounts, send a Slack alert when something drops to red, and wait for the CSM to react. That works only when red is already late — when the customer has already drafted the cancellation email or scheduled an internal call to evaluate alternatives. The intervention misses by weeks.
The predictive version of the same score catches the customer 60 to 90 days earlier, when the signals are quieter but the trajectory is clear. The math is not machine learning. It is a few weighted signals, a forward-looking trend window, and a triage cadence that maps every score band to a specific human action. This guide walks through exactly how to build that — the signals, the model, the playbook, and the failures that look like progress but quietly bleed retention.
Predict vs detect — the difference that matters
A detection score answers a present-tense question: is this customer healthy right now? You take the latest snapshot of usage, support, and billing, run them through weights, and output a number.
A prediction score answers a future-tense question: where is this customer likely to be in 30, 60, and 90 days? You take the same signals, but you measure their trend over time and project forward. The difference is small in math and enormous in outcome.
A detection score sees a customer with 200 active users today and labels them green. A prediction score sees that same customer had 240 active users 60 days ago, 220 thirty days ago, and 200 today — a 17% engagement drop on a clean linear slope — and flags churn risk for 60-90 days out, while the relationship is still warm enough to save.
By the time a detection score turns red, the customer has often already made the internal decision to leave. Predictive scoring buys you the window where intervention still works.
The 5 signals that predict SaaS churn 60+ days early
You do not need 30 inputs. You need 5 that move ahead of churn instead of with it. These are the signals that operators across B2B SaaS consistently report leading cancellation by 1-3 months.
1. Engagement decline (not engagement level)
Active user count and login frequency are the obvious metrics, and the obvious mistake is using them as levels rather than trends. A customer with 50 weekly active users could be healthy or terminally ill depending on whether that 50 is up from 30 or down from 80.
Measure the slope: weekly active users this period vs. the rolling 8-week median. A drop of more than 15% on that ratio for two consecutive weeks is the most reliable single early-warning signal in B2B SaaS. It typically appears 45-90 days before a cancellation email.
Segment by license tier. A customer paying for 100 seats with 60 active is leaking. A customer paying for 100 seats with 95 active and dropping is leaking. The seat-utilization ratio matters more than raw login counts because it directly maps to the renewal conversation: if they are not using what they bought, they will buy less or nothing.
2. Support ticket sentiment and volume shape
Most teams track ticket volume and CSAT. Both are weak predictors on their own. A customer with no tickets could be perfectly happy or quietly disengaged. A customer with many tickets could be deeply integrated and demanding, or actively suffering.
The predictive variant is the shape of ticket activity. Look for two patterns:
- Sudden silence after sustained engagement. A customer who logged 3-5 tickets per month for six months and goes to zero for 30 days is rarely satisfied — they have stopped trying. - Escalation in tone or seniority. When the person filing tickets shifts from a power user to a director or VP, you are watching the renewal conversation start without you. Tag tickets by sender role and watch for the executive escalation curve.
If you have an LLM-powered helpdesk, sentiment classification on inbound tickets adds a layer: a quiet but increasing share of negative-sentiment tickets is a 60-day warning even when CSAT scores stay flat.
3. Feature breadth contraction
Customers who churn rarely just stop using the product. They first stop using the edges of the product — the integrations, the advanced reports, the workflows their team set up six months ago. The core feature usage often holds until the very end.
This is why feature breadth is more predictive than feature depth. Count the number of distinct features a customer touches in a rolling 30-day window. When that count contracts by 30% or more relative to their 90-day baseline, you are watching a customer simplify their relationship with you in preparation for leaving it.
The nuance: define features at the right grain. Counting button clicks is noise. Counting meaningful product surfaces — integrations connected and actively syncing, reports run, automations executed, mobile sessions, API calls — is signal.
4. Executive sponsor turnover
The single biggest non-product churn driver in B2B SaaS is the person who championed the purchase leaving the company. New leadership reviews vendor contracts, runs build-vs-buy debates, and wants to put their stamp on the stack. If your champion left and you did not know within a week, you are already behind.
This signal does not come from your product. It comes from LinkedIn job-change data, email bounces, and the moment a CSM cannot get a meeting with the original sponsor anymore. Operators who track this systematically — through CRM enrichment or a job-change monitoring tool — get a 90-120 day head start on the renewal risk that follows a sponsor change.
The playbook when the signal fires: identify the new decision-maker within 14 days, run a re-discovery session with their priorities, and produce a fresh ROI summary that ties your product to the new leader's stated goals — not the old champion's.
5. Billing health (late payments and downgrade signals)
Finance teams know what customer success teams often do not: a 30-day-late invoice is a churn signal even when the customer assures you the AP delay is purely administrative. The pattern that matters is change: a customer who has paid on time for 18 months and suddenly pays 21 days late is signaling friction somewhere — internal budget review, executive scrutiny, a cancellation conversation that has not surfaced yet.
Other billing signals that predict churn:
- Removing seats below initial purchase quantity - Declining the auto-renewal opt-in or pushing renewal to monthly - Asking for an itemized usage report (the precursor to a downsize negotiation) - A new procurement contact appearing on a previously direct relationship
None of these guarantee churn. All of them shift the probability meaningfully. They are also some of the most underweighted inputs in standard health-score models — billing data lives in finance, the score lives in customer success, and the two systems often do not talk.
Building a churn risk model without an ML team
You do not need a data scientist. You need a weighted average that updates daily and a calibration loop that adjusts the weights every quarter based on what actually predicted churn in your last cohort.
Start with a simple linear model. Each signal gets a weight that sums to 1.0. Each signal is normalized to a 0-100 scale where 100 is best and 0 is worst. The output is a single 0-100 risk score, where lower means higher churn risk.
Starter weights for B2B SaaS
| Signal | Weight | Predictive Window | Source System |
|---|---|---|---|
| Engagement trend (WAU vs 8wk median) | 30% | 45-90 days | Product analytics |
| Support ticket shape and sentiment | 20% | 30-60 days | Helpdesk / CRM |
| Feature breadth contraction | 20% | 60-90 days | Product analytics |
| Executive sponsor stability | 15% | 60-120 days | CRM enrichment |
| Billing health and seat trend | 15% | 30-60 days | Billing / finance |
Two adjustments matter once you have the basic model running.
Segment by ARR tier. A $500/month customer and a $50,000/month customer should not run on the same weights. For SMB accounts, engagement decline carries more weight (no champion structure to lean on). For enterprise accounts, executive sponsor and billing signals weigh more (the contract is the relationship). A reasonable split: SMB uses 40/20/15/10/15, enterprise uses 20/15/15/30/20.
Calibrate quarterly. Pull the last quarter's churned accounts. Look at where their score landed 30, 60, and 90 days before cancellation. If your model was already showing them as risk by day 90, your weights are roughly correct. If most churns came from accounts that scored healthy at the 60-day mark, increase the weight of whichever signal showed a divergence first. Two or three calibration cycles will get you to a model that catches 70%+ of churn 60+ days out.
The intervention playbook
A score that does not trigger a defined human response is not a system. It is a dashboard. The intervention playbook below ties every score band to a 30/60/90-day rolling action — same cadence, different intensity by tier.
Red (0-40 score): CSM rescue mode
- Within 48 hours: CSM owns the account, executive briefing prepped, and a rescue plan drafted with one specific desired outcome (renewal at current ARR, downsize to retain, or planned offboarding).
- Day 7: Live call with the customer's primary user and the executive sponsor (or their replacement). Do not ask how they are. Show them the data — usage decline, ticket pattern, breadth contraction — and ask what changed.
- Day 14: Joint success plan signed: 3 measurable goals, 60-day check-in, named champion on customer side.
- Day 30: Score check. Movement of 10+ points means the plan is working. Flat or declining means escalate to executive sponsor on your side or accept the loss and plan a clean offboarding.
- Day 60-90: If score has recovered to yellow, transition back to standard cadence. If not, hold a renewal-or-graceful-exit conversation early — surprise cancellations damage future references, planned exits do not.
Yellow (41-70 score): success-plan check-in
- Day 0: Automated alert to CSM, no exec involvement yet.
- Day 7: CSM sends a personalized check-in referencing the specific signal that moved (do not say the score moved — name the behavior). Example: 'I noticed your team's usage of the reporting workspace has dropped meaningfully this month. Is something blocking the report you usually run on Mondays?'
- Day 14: If no response, second touch — short loom video walking through the dropped feature or a relevant new release.
- Day 30: If still no response and score still yellow, escalate to a 30-minute discovery call with the goal of rediscovering the customer's current priorities.
- Day 60: Score check. Yellow that is trending up needs only standard cadence. Yellow trending down moves to red playbook.
Green (71-100 score): monitoring, not silence
- Monthly automated value report (usage, outcomes, savings), no human touch required.
- Quarterly executive business review — every green account, no exceptions. Most lost expansion happens because the customer never had a structured forum to surface what else they wanted.
- Active referral and case study asks at month 6, 12, and renewal. Green accounts are your pipeline for both renewals and new logos.
Run the score, the dashboard, and the playbook in one platform
Deelo combines CRM, helpdesk, analytics, and automation in one workspace, so engagement data, ticket sentiment, executive contacts, and billing health feed the same health score and trigger the same intervention workflows. Try the platform free.
Start Free — No Credit CardMeasuring what works (so you can tune the playbook)
Most CS teams measure churn rate and stop. That tells you whether last quarter's interventions worked in aggregate; it does not tell you which interventions did the work. Three measurements matter.
- Cohort recovery rate. Of accounts that hit red in a given month, what percentage were green or recovering at the 90-day mark? This isolates the playbook from the score. A solid CS function in B2B SaaS will see recovery rates in the 25-45% range; below 15% means the playbook is too generic or too late.
- Save velocity. Average days from red trigger to score recovery. Watch this metric, not just the rate. A team that recovers 30% in 15 days is healthier than a team that recovers 35% over 90 days because the second team is fighting fires instead of preventing them.
- A/B intervention. When you have 30+ red accounts a quarter, split them. Half get the executive-led rescue, half get the CSM-led rescue. Half get a structured success plan, half get an open conversation. Measure recovery rate by arm. The gains compound — a team that runs two A/B tests a quarter will outperform a team that copies a playbook from a webinar.
Common churn-score failures
Over-weighting NPS
NPS is a satisfaction snapshot from the people willing to answer surveys. It is biased toward champions and biased away from the operational users whose disengagement actually drives churn. Use NPS for product feedback. Do not let it carry more than 10% weight in a churn-prediction model — and arguably do not include it at all.
Under-weighting product engagement
Customer success teams often anchor scores on what they can see directly — emails, calls, NPS, QBR attendance. Product engagement lives in a different system, and when integration is hard, it gets reduced to a token weight or dropped. The result: scores that move with relationship quality but not with whether the product is being used. Engagement trend is the single most predictive signal in B2B SaaS — give it the weight it earns, even if it requires a data engineer to wire up the pipeline.
Ignoring billing health
Late invoices, seat removals, and procurement involvement are some of the loudest churn signals in any portfolio, and they are routinely missing from the score because the data sits in finance. Connect the billing system to the CRM (or use a platform where they share a database from day one), and let the model see the signals that are already obvious to your CFO.
Building the score in retrospect rather than rolling forward
If your dashboard shows the current score but not the trend, you have built a thermometer, not a forecast. The score must be stored as a daily time series, not overwritten. The 60-day slope of the score is more predictive than the score itself. Without history, you cannot run the cohort recovery measurement above, and you cannot calibrate the model — you are flying with the same weights you set on day one.
Confusing red with churned
A red account is not a lost account; it is an account that needs a defined response in the next 48 hours. Teams that treat red as a cancellation queue stop investing in the playbook because the outcome looks predetermined. Teams that treat red as a 90-day rescue window post recovery rates above 30% — meaning roughly one in three near-cancellations gets saved, which over a year is meaningful ARR protected.
How Deelo automates this
Most SaaS teams hit one structural problem: the signals live in five different tools. Engagement is in product analytics. Tickets are in the helpdesk. Sponsor changes are in the CRM. Billing data is in Stripe or finance. Email response patterns are in the email tool. Stitching them into a single rolling score requires either a data engineer or a months-long Zapier graph that breaks every time a tool updates its API.
Deelo collapses the systems. CRM, helpdesk, analytics, invoicing, and email run on the same database, which means a customer health score can reference engagement, ticket sentiment, executive contact stability, and billing health without an integration layer.
In practice that looks like:
- A custom field on the company record stores the daily health score, written by a scheduled workflow that reads from CRM, helpdesk, analytics, and invoicing. - A daily automation compares today's score to the 30-day average; a drop of 10+ points emits an event. - That event drives the playbook: a CSM task is created, a Slack alert fires, an executive briefing email is queued for accounts above an ARR threshold, and the account is added to a rescue cohort for cohort-level reporting. - The Analytics app charts score-by-cohort, recovery rate, and save velocity without anyone exporting a CSV.
The point is not the tooling. It is that a predictive churn system needs the data unified and the actions automated. When the signals and the playbook live in the same workspace, customer success teams stop debating data quality and start running the loop that actually retains revenue.
Build a predictive customer health score in Deelo
CRM, helpdesk, analytics, and automation in one platform. Try Deelo free — no credit card, full access, your churn-prediction system live in days, not quarters.
Start Free — No Credit CardCustomer health score and churn prediction FAQ
- How accurate is a weighted-average churn prediction model compared to ML?
- A well-calibrated weighted-average model typically catches 60-75% of churn at the 60-day mark in B2B SaaS, which is enough to drive a meaningful save rate. ML models can push that into the 75-85% range with significantly more engineering investment, but the marginal lift rarely justifies the build cost for teams under $50M ARR. The bigger leverage is not the model — it is the intervention playbook attached to it. A 60% predictive model paired with a disciplined CSM cadence outperforms a 80% model that triggers no consistent action.
- How often should the score update — daily, weekly, or real-time?
- Daily. Real-time is overkill — humans cannot intervene faster than a daily cadence anyway, and it creates noise from intra-day fluctuations. Weekly is too slow because you lose the leading-edge signal on engagement drops, which often appear inside a 7-day window before a cancellation conversation starts. Run the calculation overnight, deliver alerts in the morning, give CSMs a clean queue to work from at the start of each day.
- Should the customer success team or RevOps own the health score?
- RevOps owns the model and the data pipeline. Customer success owns the playbook and the intervention. The single most common failure mode is when CS owns both — the score becomes whatever makes the dashboard look good, weights get tuned to flatter outcomes rather than predict them, and calibration stops happening. The single most common failure on the other side is when RevOps owns both — the score is mathematically clean but never connects to a human action, so it changes nothing. Split the ownership: model and metrics with the data team, response and rescue with CS.
- What if we are too small to have a customer success team?
- The score still works, but the playbook becomes founder-led. For teams under $1M ARR, run the model on your top 20 accounts manually each week and let the rest run on automated email touchpoints. The 80/20 economics are dramatic: above $5K MRR, every red customer is worth a founder phone call. Below that threshold, automate the yellow and red touchpoints with templated outreach and review the cohort weekly. The score still tells you where to focus the limited human time you have.
- How long until a predictive score actually starts saving revenue?
- Build phase is 2-4 weeks if your data is reasonably unified, 6-12 weeks if you are pulling from disparate tools. First reliable predictions appear after 60-90 days of data accumulation — you need enough history to calculate trends. Save rate becomes measurable after the first churn cohort runs through a full 90-day intervention window. Realistically, expect 4-6 months from build start to defensible 'this saved us X dollars in ARR' reporting. Teams that try to declare victory in month one have not yet earned the right to interpret the data.
Related pages
Explore More
Related Articles
Best Personal Injury Case Management Software in 2026
A head-to-head comparison of the top personal injury case management platforms in 2026. Lien tracking, medical record management, demand letters, contingency math, and settlement distribution compared across Clio, MyCase, Filevine, CASEpeer, PracticePanther, Smokeball, and Deelo.
12 min read
How-ToHow to Start a Plastic Surgery Practice: Complete 2026 Guide
A step-by-step guide to launching a plastic surgery practice in 2026. Licensing, credentialing, facility setup, liability insurance, patient pipeline, operations software, and first-year revenue targets.
14 min read
Best OfBest Podcast Management Software in 2026
The top podcast management platforms compared for 2026. Descript, Captivate, Buzzsprout, Transistor, Riverside, and Deelo — features, pricing, and the angle each takes for professional podcasters.
11 min read
ComparisonDeelo vs ServiceTitan: The Honest 2026 Comparison
A genuinely fair side-by-side comparison of Deelo and ServiceTitan for field service businesses. Pricing, features, strengths, weaknesses, and who each platform is really built for.
12 min read