← Reference Deep-Dives Reference Deep-Dive · 9 min read

Why NPS Fails in Due Diligence (And What to Measure Instead)

By Kevin, Founder & CEO

Net Promoter Score has achieved something remarkable in the business world: near-universal adoption paired with widespread misapplication. In its intended context — longitudinal benchmarking within a single company over time — NPS provides a useful, if blunt, signal. In the context of commercial due diligence — where a single measurement must inform a nine-figure investment decision — NPS is not just insufficient. It is actively misleading.

This is not a theoretical concern. Deal teams that anchor on NPS during diligence routinely misread the health of the customer base because NPS collapses complex, multidimensional customer sentiment into a single number that obscures more than it reveals. A target company presenting an NPS of 55 could be a healthy business with genuinely loyal customers, or it could be a business where high switching costs trap dissatisfied customers who inflate the score while quietly preparing to leave.

Understanding why NPS fails in diligence — and what to measure instead — is essential for any PE team that takes customer evidence seriously.

The Four Failure Modes of NPS in Due Diligence


Failure Mode 1: No Trend Visibility

NPS as typically encountered in diligence is a point-in-time number. Management presents it in the CIM or during a management meeting: “Our NPS is 52.” This tells the deal team absolutely nothing about direction. Is 52 an improvement from 35 two years ago, suggesting positive momentum? Or a decline from 68 eighteen months ago, suggesting deterioration? The number alone carries no trajectory information.

Even when management provides historical NPS data, the methodology behind each measurement is rarely consistent. One survey may have been sent to the full customer base; another may have targeted only recent buyers. Response rates fluctuate, and the customers who respond to NPS surveys are systematically different from those who ignore them — typically more engaged, more opinionated, and more extreme in either direction. Without controlled methodology, NPS trend data is comparing measurements taken with different instruments and calling it a trend.

Failure Mode 2: No Segmentation

A company-wide NPS of 50 could reflect uniform satisfaction across the customer base, or it could conceal a bimodal distribution where enterprise customers score 75 and SMB customers score 20. These are fundamentally different businesses with fundamentally different investment implications.

Segment-level variance is the norm, not the exception. In most B2B software companies, satisfaction varies dramatically by customer size (enterprise vs. mid-market vs. SMB), vertical industry (the product may be excellent for healthcare and mediocre for financial services), tenure (new customers in the honeymoon phase score differently than three-year veterans), and use case (customers using the core product may be satisfied while those depending on a secondary module are frustrated).

A deal thesis built on aggregate NPS ignores these dynamics entirely. An acquisition predicated on expanding into mid-market financial services — because the aggregate NPS looks healthy — might discover post-close that the specific segment it plans to grow in is the one dragging the average down.

Failure Mode 3: No Causal Explanation

NPS tells you what customers scored. It does not tell you why. A promoter (score 9-10) might be enthusiastic because the product solves a critical problem with no viable alternative, because a recent support interaction was exceptional, because they have a personal relationship with the account manager, or because they signed a favorable contract and feel good about the economics. Each of these reasons has different implications for retention durability, competitive vulnerability, and scalability.

Detractors (score 0-6) are similarly opaque. A score of 4 might reflect a fundamental product gap, a temporary implementation problem, organizational politics that the vendor cannot influence, or a pricing dispute that could be resolved with a contract adjustment. The score is identical; the appropriate response ranges from urgent product investment to a simple commercial conversation.

Without understanding causation, NPS is a thermometer that tells you the patient has a fever but not whether it is a cold or a critical infection. In diligence, where the treatment plan (post-close value creation strategy) depends entirely on accurate diagnosis, a thermometer reading is dangerously inadequate.

Failure Mode 4: Switching Cost Inflation

This is the most insidious failure mode because it makes unhealthy businesses look healthy. Customers who are trapped by switching costs — deep integrations, data migration complexity, organizational retraining requirements, multi-year contracts — often score higher on NPS than their actual satisfaction warrants.

The psychology is straightforward. When a customer has already decided (consciously or unconsciously) that switching is impractical, cognitive dissonance motivates them to evaluate their current vendor more favorably. Admitting deep dissatisfaction with a product you cannot leave is psychologically uncomfortable. The result is what researchers call “satisficing” — reporting adequate satisfaction rather than confronting the underlying frustration.

In diligence, switching cost inflation creates a specific and dangerous illusion: the appearance of loyalty where only inertia exists. A customer base with high NPS driven by genuine product affinity will retain and expand under new ownership. A customer base with high NPS driven by switching costs will retain temporarily — until a competitor reduces the switching barrier, a contract renewal creates a decision point, or an internal champion departure eliminates the institutional memory that maintains the integration.

The distinction between these scenarios is invisible in NPS data. It is discoverable only through deeper investigation of the reasons behind the scores.

What to Measure Instead: A Due Diligence Measurement Framework


Replacing NPS does not mean abandoning customer measurement. It means replacing a single flawed metric with a multi-dimensional framework that captures the complexity NPS obscures.

Segment-Level Satisfaction With Driver Analysis

Measure satisfaction at the segment level — by customer size, industry, tenure, and use case — and for each segment, identify the top three drivers of satisfaction and the top three drivers of dissatisfaction. This produces a diagnostic map rather than a single score.

The driver analysis is critical because it converts measurement into action. “Satisfaction is 72 in the enterprise healthcare segment” is a measurement. “Satisfaction is 72 in the enterprise healthcare segment, driven by workflow integration (positive), clinical reporting (positive), and implementation speed (negative)” is a diagnostic that tells the operating team exactly where to invest.

AI-moderated interviews excel at driver analysis because the laddering technique — probing five to seven levels deep on every response — surfaces the specific, granular reasons behind satisfaction and dissatisfaction. A traditional NPS survey with an open-text comment field captures whatever the respondent chooses to write in 30 seconds. A structured interview captures the full causal chain.

Competitive Vulnerability Score

Competitive vulnerability measures the degree to which customers are aware of, interested in, and actively evaluating alternatives. It is the metric that NPS advocates assume the score already captures — but it does not.

A robust competitive vulnerability assessment answers four questions for each customer segment. Are customers aware of specific alternatives? (Awareness alone is not threatening, but unawareness signals a defensible position.) Have customers evaluated alternatives in the past 12 months? If yes, what triggered the evaluation? And what kept them from switching — product superiority, switching costs, contractual obligations, or simple inertia?

The resulting score is far more predictive of retention than NPS. A customer segment with moderate satisfaction (65) but low competitive awareness and no recent evaluation activity is more defensible than a segment with high satisfaction (80) where 40% of accounts have evaluated alternatives in the past year.

Expansion Willingness and Barrier Identification

Expansion potential is a critical value creation input, but it requires more nuance than a single willingness metric. The measurement should capture three dimensions: stated willingness to expand (would you increase usage, purchase additional modules, or extend to other departments?), the specific conditions required for expansion (what would need to happen first?), and the barriers that currently prevent expansion (budget constraints, competing priorities, product gaps, organizational resistance).

This three-dimensional view distinguishes between segments where expansion is probable (high willingness, low barriers), possible (high willingness, addressable barriers), and unlikely (low willingness or structural barriers that the vendor cannot influence). Each category requires a different investment approach, and misclassifying a “structural barrier” segment as “probable expansion” leads to wasted go-to-market investment and missed plan targets.

Champion Dependency Assessment

Many B2B software companies rely heavily on internal champions — individuals within customer organizations who advocate for the product, drive adoption, and defend the vendor relationship during competitive evaluations. Champion dependency is not inherently problematic, but it creates concentration risk that NPS does not capture.

The assessment identifies how many accounts depend on a single champion, whether backup champions exist, what would happen to the vendor relationship if the primary champion left the organization, and the historical pattern of champion turnover in each segment. Accounts with single-champion dependency and no succession plan represent a specific, measurable retention risk that can be addressed through deliberate relationship diversification.

Switching Cost vs. Genuine Loyalty Ratio

This is the metric that directly addresses NPS’s most dangerous failure mode. For each customer segment, the measurement distinguishes between customers who stay because of genuine preference (they choose the product because it is the best option) and customers who stay because of structural friction (they would consider alternatives if switching were easier).

The ratio is determined through structured interview questions that explore hypothetical scenarios: if switching were free and instant, would the customer stay? If a competitor offered equivalent functionality with better economics, what would happen? If the contract were month-to-month rather than annual, would behavior change?

A segment where 80% of retention is driven by genuine loyalty has fundamentally different economics than a segment where 80% of retention is driven by switching costs. Both may show identical NPS scores and identical current retention rates. Their forward-looking risk profiles diverge completely.

How AI Interviews Transform Customer Measurement


Traditional survey-based measurement — including NPS — is limited by the tradeoff between scale and depth. Surveys can reach hundreds of respondents but capture only surface-level responses. In-depth interviews provide rich insight but traditionally require human interviewers who limit throughput to 3-5 conversations per day.

AI-moderated interviews eliminate this tradeoff. Every interview applies the full depth of structured laddering — probing the “why” behind every response through five to seven levels of follow-up — while scaling to 50, 100, or 500 interviews without incremental interviewer cost or scheduling constraints.

This combination of depth and scale is what makes the measurement framework described above practical for due diligence timelines. Segment-level driver analysis requires enough interviews per segment to identify patterns. Competitive vulnerability scoring requires probing that goes well beyond a survey question. The switching cost vs. loyalty ratio requires hypothetical scenario exploration that surveys cannot facilitate. All of these methods demand conversational depth at statistical scale — precisely what AI-moderated interviews deliver.

The result is a customer evidence base that NPS partisans might not recognize. Instead of a single number that executives cite in board meetings, the deal team receives a multi-dimensional assessment that reveals where the business is strong, where it is vulnerable, why customers behave the way they do, and what specific levers the operating team can pull to improve outcomes.

The Right Role for NPS


None of this argues that NPS is useless in all contexts. As a longitudinal tracking metric within a single company using consistent methodology, NPS provides a useful directional signal. If the same survey, sent to the same customer population, using the same cadence, shows NPS moving from 40 to 55 over two years, that trend suggests genuine improvement.

But longitudinal tracking within a single company is not the due diligence use case. In diligence, NPS is typically encountered as a point-in-time number from an unknown methodology, presented by a management team with an incentive to highlight the most favorable interpretation. In this context, NPS is not a starting point for analysis — it is a distraction from the deeper investigation that the investment decision demands.

For teams building a more rigorous customer measurement practice, our commercial due diligence solution replaces NPS-centric approaches with the multi-dimensional framework described in this guide, and our NPS and CSAT solution shows how AI interviews transform satisfaction measurement from score collection into causal understanding.

Frequently Asked Questions

NPS fails because it obscures segment-level variance by averaging across customers with wildly different experiences, provides no competitive context, explains correlation not causation, and can reflect switching cost lock-in rather than genuine loyalty. A target company with a 52 NPS in a market where the average is 38 looks strong, but if churn rates are rising and NPS is held up by customers who can't leave easily, the score is actively misleading.
A multi-dimensional framework should include competitive vulnerability scores (how easily could customers switch?), switching cost ratios (economic vs. behavioral lock-in), segment-level retention rates by cohort and deal size, and voice-of-customer depth on renewal intent. These metrics reveal the actual durability of the revenue base rather than the averaged sentiment that NPS produces.
Most companies collect NPS at post-onboarding or annual renewal touchpoints — moments of either peak satisfaction or contractual obligation. This misses the erosion of satisfaction that happens mid-cycle, which is precisely when competitive conversations begin. Longitudinal customer interviews capture the full sentiment arc and reveal early warning signals that point-in-time NPS surveys structurally cannot.
User Intuition deploys AI-moderated customer interviews that go beyond NPS to probe competitive vulnerability, switching intent, and the specific product or service gaps driving dissatisfaction — all within 48-72 hours of program launch. At $20 per interview, a due diligence team can interview 100+ customers for a fraction of the cost of a traditional consulting engagement, producing the segment-level evidence required for defensible investment decisions.
Get Started

Put This Research Into Action

Run your first 3 AI-moderated customer interviews free — no credit card, no sales call.

Self-serve

3 interviews free. No credit card required.

See it First

Explore a real study output — no sales call needed.

No contract · No retainers · Results in 72 hours