Search “concept testing cost” and you’ll find something unusual for a market worth hundreds of millions of dollars annually: nobody publishes prices. Kantar won’t tell you. Ipsos won’t tell you. The boutique qualitative shop that does “consumer insight and innovation consulting” definitely won’t tell you until you’ve been through a 45-minute discovery call and a proposal process that takes two weeks.
This isn’t accidental. The reason pricing is opaque in concept testing is the same reason it’s opaque in any professional services market where markup is high and comparisons are difficult: transparency would force a reckoning with what things actually cost versus what clients are willing to pay.
This post breaks the silence. What concept testing actually costs at every tier, broken down by where the money goes. When the expensive option is genuinely worth it. When $200 is enough. And how to build a concept testing program that compounds over time instead of depleting your research budget on one study per year.
Why Pricing Transparency Matters Here
The pricing opacity in concept testing has a concrete cost: it makes most organizations systematically under-test. A brand team with a $50,000 annual research budget assumes they can afford one, maybe two major concept tests per year. They spend $40,000 on a comprehensive agency study, learn from it, and then — because the next study would consume the remaining budget — rely on those findings for the next 8-12 months regardless of how much the market has shifted.
Meanwhile, the concepts that weren’t tested, the positioning options that weren’t validated, the messaging iterations that weren’t explored — those decisions get made on gut instinct, internal debate, and whatever survey data the team can pull from existing research. Which means they frequently get made wrong.
The solution isn’t a bigger research budget. It’s understanding what concept testing actually costs at each tier — and recognizing that for most concept testing questions, the answer is a fraction of what most teams assume.
The Four Concept Testing Tiers
Tier 1: DIY Surveys ($500–$3,000)
The most accessible form of concept testing is also the most limited. A survey distributed through SurveyMonkey, Typeform, Qualtrics, or Google Forms — sent to your email list, posted in a community, or run through a basic panel service — gives you quantitative response data on a concept quickly and cheaply.
What you get: Purchase intent ratings (usually a 5-point scale), feature preference rankings, price sensitivity data, and demographic breakdowns. A 200-person survey on a concept can tell you that 34% of respondents said they were “very likely” or “extremely likely” to purchase. It can tell you that younger respondents rated the concept more favorably than older ones.
The depth problem: What a survey cannot tell you is why. It can’t tell you what drove the 34% who responded positively to be interested, or what specific objection caused the 66% to pass. It can’t tell you whether “I would definitely buy this” represents genuine purchase intent or social desirability bias — the well-documented tendency of survey respondents to report preferences that seem reasonable rather than ones that reflect actual behavior. It can’t probe the moment of hesitation, the competing product that came to mind, the price anchor that made the concept feel expensive or cheap.
Concept testing questions that surveys can answer: “Is the reaction directionally positive or negative?” “Which of three concepts generates the highest stated intent?” “Does this concept appeal more to segment A or segment B?”
Concept testing questions that surveys cannot answer: “Why aren’t people excited about this?” “What specific language triggers concern?” “What would need to change for this to be compelling to the people who passed?”
When it’s enough: Early screening of a large concept set where you’re trying to narrow 10 options to 3. Simple preference ranking across clearly differentiated options. Tracking shifts in sentiment across time when you’ve already done qualitative work to understand the underlying drivers.
Tier 2: AI-Moderated Interviews ($200–$5,000)
AI-moderated concept testing interviews are 1:1 conversations — 30+ minutes per participant — where a trained AI moderator guides the participant through a concept, probes their reactions with follow-up questions, and ladders down through surface responses to uncover underlying motivations and concerns. Unlike surveys, every answer generates a follow-up. Unlike human-moderated research, the AI applies consistent methodology across every conversation without moderator variability, fatigue, or the natural human tendency to move on when a respondent goes in an unexpected direction.
What you get: Depth equivalent to traditional qualitative research at dramatically lower cost. For $200 (20 interviews), you get 10+ hours of conversation data synthesized into thematic findings with verbatim quote evidence. You understand not just whether consumers respond positively to a concept, but what specifically resonates, what language they use to describe it, what competing alternatives come to mind, what would make it more compelling, and what would make them walk away.
Why it’s cheap: Traditional qualitative research is expensive not because the conversations are expensive, but because of everything around the conversations. Scheduling individual participants costs time and coordination overhead. Human moderators charge $150-$300/hour and can conduct only 4-6 interviews per day. Analysis and reporting require trained researchers spending 40-80 hours synthesizing findings into a deliverable. AI removes or automates most of this: scheduling is self-serve, moderation runs at scale without hourly fees, and synthesis is automated across all conversations simultaneously.
Sample size guidance:
- 20 interviews ($200): Directional reading on a single concept. Sufficient to identify dominant themes — what’s working, what’s not, what the main concerns are.
- 50 interviews ($500): Concept validation with meaningful pattern confidence. Enough variation to distinguish majority reactions from outlier perspectives. Appropriate for packaging decisions, messaging tests, and go/no-go decisions on development investment.
- 100 interviews ($1,000): Reliable segmentation — you can compare reactions across demographic groups, usage occasions, or competitive purchase behavior with enough within-segment depth to draw conclusions.
- 300 interviews ($3,000): Quantitative-level pattern detection within qualitative data. When 78% of participants raise a specific concern unprompted, that’s not a theme — it’s a finding you can cite with precision. This is where AI-moderated research starts to challenge what large-scale quant surveys can tell you.
User Intuition’s concept testing solution operates at all four budget levels. The methodology — 5-7 level laddering, non-leading language calibrated against research standards, evidence-traced findings to verbatim quotes — is identical regardless of whether you’re running 20 interviews or 300.
Tier 3: Specialized Quantitative Platforms ($5,000–$25,000)
Platforms like Zappi, Quantilope, and similar specialized concept testing tools occupy the middle tier: faster than traditional agencies, more structured than DIY surveys, and oriented around specific quantitative methodologies that produce normative scores and benchmarked outputs.
What you get: Concept testing with validated scales (purchase intent measured against category norms, breakthrough potential benchmarked against competitive concepts), MaxDiff analysis for feature prioritization, conjoint studies to model consumer trade-offs, and normative databases that let you say “this concept scores in the 73rd percentile for purchase intent in the personal care category.”
What the premium buys: The normative database is the real differentiator. If your organization needs to justify a major innovation investment to a board or senior leadership team and you need to demonstrate that your concept performed well against category benchmarks — not just that consumers said they liked it — a normative database has genuine value. These platforms have run thousands of studies and can tell you how your concept’s scores compare to what typically succeeds versus fails in your category.
Where the limitation is: Scale studies of this type prioritize measurement over understanding. MaxDiff tells you which features consumers prefer most among a defined set. Conjoint tells you the trade-off value of price versus features. Neither tells you why consumers rank things the way they do, what underlying need or concern drives the preference, or what you’d need to change about a low-scoring feature to make it compelling. For optimization decisions driven by measurable scores, these platforms are well-suited. For diagnostic questions — “why is this concept failing to land?” — they’re less useful.
When to use this tier: Late-stage concept validation before major investment, when you need to demonstrate concept performance against a normative baseline. Portfolio decisions across multiple concepts where you need a consistent measurement framework. Annual brand tracking that includes concept scores alongside brand health metrics.
Tier 4: Full-Service Agencies ($15,000–$75,000+)
Traditional qualitative research agencies — the Kantars, Ipsos Qualitatives, Hall & Partners of the world, plus dozens of boutiques — offer full-service concept testing that includes everything: study design, participant recruitment, moderation, analysis, and deliverable production. Many agencies now use AI-moderated platforms to deliver equivalent depth at a fraction of the cost. The quote for a standard concept testing study runs $15,000-$40,000 for a moderate scope. Complex multi-cell studies, in-home use tests, or studies requiring simultaneous multi-market fieldwork push to $40,000-$75,000 and beyond.
What you’re actually paying for: We’ll cover this in detail in the next section. The short answer is that you’re paying for a lot of human infrastructure around a core research act that, in most cases, doesn’t require it.
What you get: Deeply experienced research design. Moderator expertise that can navigate complex, ambiguous, or sensitive topic areas. Deliverables polished for executive audiences. Account management that handles every logistical detail. And, in some cases, the agency brand name on the research report — which has its own political value inside large organizations.
Cost Breakdown Table
| Method | Typical Cost | Turnaround | Sample Size | Depth of Insight | What It Answers | When to Use |
|---|---|---|---|---|---|---|
| DIY surveys | $500–$3,000 | 1–2 weeks | 100–1,000 | Low | Stated preference, directional reaction | Early screening, preference ranking |
| AI-moderated interviews | $200–$5,000 | 48–72 hours | 20–300 | High | Why, what would change it, verbatim language | Most concept testing use cases |
| Specialized quant platforms | $5,000–$25,000 | 1–2 weeks | 200–1,000 | Medium | Normative benchmarking, MaxDiff, conjoint | Late-stage validation vs. category norms |
| Focus groups (agency) | $12,000–$60,000 | 4–6 weeks | 8–12 per group | Medium | Group dynamics, social reaction | Stimulus testing requiring visible group response |
| Full-service agency (qual) | $15,000–$75,000+ | 6–12 weeks | 30–100 | Very high | Complete diagnostic with stakeholder packaging | Regulatory requirements, high-stakes launches |
A note on the “depth” column: it refers to access to actual consumer decision logic — motivations, barriers, emotional triggers, language consumers use when they talk about a concept. Surveys are shallow because there’s no probe. Focus groups are medium because group dynamics suppress individual candor (people adjust their answers based on what others in the room have said). AI-moderated 1:1 interviews produce high depth for the same reason human-moderated 1:1 interviews do: participants speak to a single listener without social pressure, and every answer generates a follow-up.
Different annual budgets produce very different concept testing programs. Here is what each tier realistically delivers across a year.
| Annual Budget | Recommended Approach | Studies/Year | Depth | Turnaround |
|---|---|---|---|---|
| Under $2,000 | AI-moderated interviews (per-study) | 4-8 screening tests | High (30+ min, 5-7 level laddering) | 48-72 hours |
| $2,000-$10,000 | AI-moderated iterative program | 8-15 (screening + development + validation) | High | 48-72 hours |
| $10,000-$50,000 | Blended: AI-moderated + specialized quant (Zappi, Quantilope) | 15-25 AI + 1-2 normative benchmarks | High + normative scoring | 48 hours - 2 weeks |
| $50,000-$150,000 | Full-service agency + AI supplements | 1-2 agency + 20-30 AI studies | Very high (includes stakeholder deliverables) | 2-8 weeks (agency), 48-72 hours (AI) |
| $150,000+ | Enterprise: agency + quant platform + AI continuous testing | Continuous pipeline screening + project validation | Comprehensive | Mixed |
The ROI of concept testing is driven by what it prevents. Here is the math on decisions that go wrong without evidence versus the cost of getting evidence first.
| Scenario | Cost of Getting It Wrong | Cost of Research | ROI Multiple |
|---|---|---|---|
| Failed product launch (development + manufacturing + marketing) | $500,000-$5,000,000 | $1,000 (50 concept validation interviews) | 500-5,000:1 |
| Wrong packaging direction (production run + retail reset + lost velocity) | $200,000-$2,000,000 | $500 (25 packaging concept interviews) | 400-4,000:1 |
| Messaging that doesn’t resonate (wasted media spend on flawed creative) | $200,000-$1,000,000 | $500 (25 message testing interviews) | 400-2,000:1 |
| Launching the second-best concept when the winner was in the set | $1,000,000+ in unrealized revenue | $2,000 (100 comparative concept interviews) | 500:1 |
What You’re Actually Paying for in a $40,000 Agency Concept Test
When a research agency quotes $40,000 for a concept testing study, the instinct is to treat this as the cost of concept testing. It isn’t. It’s the cost of concept testing when you buy it from a firm with a specific cost structure. Here is where the money goes.
Agency Account Management and Project Coordination: 30–40% of Budget
The largest single cost in most agency engagements doesn’t appear in any deliverable. It’s overhead: the account manager who runs your kickoff call, the project manager who coordinates scheduling and participant logistics, the agency principal who appears at the read-out presentation, the internal QA review of the discussion guide, the client services infrastructure that keeps everything moving. Industry standard professional services overhead runs 30-40% of total project cost.
On a $40,000 engagement, $12,000-$16,000 is agency overhead — before a single consumer has spoken to a moderator. This overhead doesn’t improve research quality. It coordinates logistics that exist because human-moderated research requires significant coordination.
Moderator Fees: $150–$300 Per Hour
Experienced qualitative researchers charge $150-$300 per hour for moderation. A 45-minute concept testing interview takes 60-90 minutes of moderator time when you include setup, the interview itself, and brief notes immediately after. At $200/hour, that’s $120-$180 per completed interview.
A study with 30 completed interviews — considered minimal for a serious concept test — costs $3,600-$5,400 in moderator fees alone. Add preparation time (reviewing discussion guide, concept stimulus, participant screener data: 4-8 hours per project, plus briefing calls), and moderator costs for a modest study easily reach $4,500-$7,000.
These rates reflect years of methodological training and genuine expertise. Experienced concept testing moderators know how to probe without leading, how to manage a participant who gives only surface answers, how to pivot when a concept isn’t landing the way the stimulus suggests. The skill is real. The question is whether you need to pay for it in every study, or only in the specific studies where the complexity genuinely requires it.
Focus Group Facility Rental: $1,500–$3,000 Per Day
Focus group concept testing requires dedicated facilities: viewing rooms with one-way mirrors, recording equipment, separate client observation areas, and typically catering for a full-day session. These facilities cost $1,500-$3,000 per day to rent in most major markets, and concept testing studies typically require 2-4 group sessions across 2 days to have any confidence in the findings.
Facility cost for a 4-group concept test: $6,000-$12,000. This buys you a room with a mirror. The research itself — the quality of the conversation, the depth of probe, the moderator’s ability to surface what’s actually driving consumer reaction — is unrelated to whether the room has a one-way mirror and a viewing lounge.
Remote concept testing has eliminated facility costs for most study designs. But if your agency quote includes in-person focus groups, you’re paying $6,000-$12,000 for infrastructure that improves logistics and executive experience (clients can watch from behind the glass) more than it improves research quality.
Participant Recruitment and Incentives: $75–$150 Per Person, Plus Buffer
Recruiting participants for a concept test requires a screener questionnaire, panel access fees, participant incentives, and a no-show buffer. Category buyers with specific usage characteristics — the right demographic, the right purchase frequency, the right competitive brand repertoire — cost more to recruit than general population respondents.
Incentives for a 90-minute concept testing session run $75-$150 per completed participant. No-show rates for in-person research run 20-30%, which means you need to recruit and pay (partial) incentives for more people than you need to complete. By the time you account for screening cost, completed incentives, and partial no-show incentives, recruitment cost runs $100-$200 per completed participant.
For a 50-person study: $5,000-$10,000 in recruitment alone.
Analysis, Reporting, and Presentation: 40–80 Analyst Hours
After fieldwork is complete, a senior researcher synthesizes 30-100 conversations into a deliverable: typically a 40-60 page slide deck with executive summary, key findings by theme, implication sections, verbatim quote evidence, and a recommendations appendix. Writing this deck takes 40-80 hours of senior researcher time. At agency billing rates ($150-$250/hour), that’s $6,000-$20,000 for the deliverable — before client revisions, which are typically two rounds included and $1,000-$3,000 per round after that.
The deck is often the most visible artifact of the engagement. It represents weeks of work and significant budget. And its useful life — the period when it’s actively referenced in decisions — is typically 60-90 days. After that, it lives in a shared drive and is cited occasionally in meetings until the next major research initiative starts fresh.
The Actual Insight: Maybe 20% of What You Paid
Add it up on a $40,000 concept testing study:
- Agency overhead and account management: $12,000-$16,000
- Moderator fees: $5,000-$8,000
- Facility rental: $4,000-$9,000
- Participant recruitment and incentives: $5,000-$10,000
- Analysis and deliverable production: $6,000-$12,000
The actual time in conversation with consumers — the core research act that generates insight — is embedded inside this. It is not itemized because it cannot easily be separated from the infrastructure required to produce it in the traditional model. That infrastructure is expensive. And in most concept testing contexts, most of it is replaceable.
When the Expensive Option Is Worth It
This guide would be dishonest if it suggested that AI-moderated interviews are the right choice for every concept testing situation. There are specific, identifiable circumstances where higher-cost approaches are genuinely justified. Knowing when those apply — and when they don’t — is the most valuable thing a research budget holder can know.
Regulatory and Legal Requirements
Claims substantiation for FDA-regulated products, FTC advertising substantiation for specific categories, HIPAA-compliant research protocols — these contexts require specific methodology documentation that satisfies regulatory standards. The research infrastructure around the methodology (documented chain of custody, certified moderation, legal review of discussion guides) is compliance-driven overhead, not quality-driven. If your concept will carry specific claims requiring regulatory substantiation, budget accordingly and verify that your research vendor can provide the required documentation.
Physical Product Testing
If your concept involves taste, texture, smell, tactile quality, or any other sensory dimension that can’t be communicated through a written description, image, or video stimulus — you need to get product into participants’ hands. In-home use tests, sensory evaluation sessions, and co-creation workshops requiring physical prototypes have legitimate logistics costs that exist regardless of the moderation methodology. A packaging concept can be tested remotely. A new flavor formulation cannot.
Normative Benchmarking
If your organization needs to know not just whether a concept performed well in your study but how it performed relative to category norms — “your concept scored in the 82nd percentile for purchase intent among personal care concepts in the Zappi database” — you need a platform with a validated normative database. No amount of AI-moderated interviews will give you that comparison, because the comparison requires a consistent measurement methodology applied across thousands of studies with consistent scales and stimulus protocols.
Normative benchmarking is genuinely valuable for portfolio decisions (which of our five concepts should we invest in developing, relative to what has historically succeeded in this category) and for board-level justification of innovation investment (“this concept significantly outperforms category benchmarks”). It is less valuable for diagnostic questions, early-stage screening, or iterative concept refinement.
Executive Stakeholder Management
Sometimes the brand name on the research report matters more to organizational decision-making than the content of the report. A strategic concept that requires C-suite or board buy-in may benefit from being validated by a recognized research firm rather than an AI platform, regardless of the comparative quality of the underlying research. This is a legitimate business consideration — organizational credibility is real — even if it has nothing to do with research quality. Be honest with yourself about when you’re spending research budget on research and when you’re spending it on internal politics.
High-Stakes Global Launches
A concept going to simultaneous launch across 10+ markets, each requiring local cultural calibration, translated and back-translated stimulus materials, regional recruitment networks with verified category expertise, and in-country moderation oversight genuinely requires agency infrastructure. The coordination cost is real and not easily replaced by a technology platform. The threshold for this is high — most “global” concept tests don’t actually require simultaneous multi-market execution on this scale — but when it applies, it applies.
When $200–$5,000 Is Genuinely Enough
The majority of commercial concept testing questions — the ones that drive the day-to-day decisions of brand teams, innovation leaders, and product managers — can be answered at a fraction of what most organizations spend. Here is what different budget levels get you in practice.
$200 (20 interviews): A directional reading on a single concept. Should we develop this further? Is the reaction broadly positive or broadly skeptical? What’s the main thing consumers like? What’s the main concern? Twenty 30-minute conversations will reliably surface the dominant patterns. You won’t have statistical significance, but you’ll have the themes — and the verbatim language consumers use when they talk about the concept, which is often more valuable than a purchase intent score.
This is the right budget for: early-stage concept screening before committing to development investment, go/no-go decisions where the question is whether to pursue an idea at all, quick reads on rough concepts before they’re refined enough to warrant a larger study.
$500 (50 interviews): Concept validation with meaningful pattern confidence. Enough variation to distinguish majority reactions from outlier perspectives, to identify which specific elements are driving positive and negative reactions, and to surface the key language themes that would make the concept more compelling. This is appropriate for a packaging decision, a headline test, a go/no-go on product development investment.
Eric O. at Turning Point Brands used AI-moderated concept testing to evaluate three packaging concepts across 72 hours — a timeline that would have required 6-8 weeks through a traditional agency process. The test identified which concept drove the clearest premium signal and surfaced the specific visual element that was creating confusion in the runner-up design. Total cost: a fraction of what a single traditional focus group session would have cost.
$1,000 (100 interviews): Reliable segmentation across meaningful groups. “Heavy users react differently than light users — here’s why.” “The 35-54 demographic raises a concern that younger consumers don’t.” “Competitive brand buyers see the concept differently than current customers.” At 100 interviews, you have enough within-segment depth to draw conclusions about how different consumer groups experience a concept differently, not just what the average reaction is.
$3,000-$5,000 (300+ interviews): Quantitative-level pattern detection within qualitative data. When 74% of participants raise a specific concern unprompted, that’s a finding with the precision of survey data and the explanatory depth of qualitative research. This budget level is appropriate for final concept validation before major launch investment, for studies where precision matters and directional findings aren’t sufficient, and for tracking concept performance across iterative development.
The complete guide to concept testing covers research design frameworks for different questions in more detail. The concept testing questions guide addresses how to structure your discussion guide at each stage of development.
How to Build a Compounding Concept Research Budget
Most organizations currently budget for concept testing the way they budget for a capital expenditure: large, infrequent, and difficult to justify. A typical brand team might commission one major concept test per year — spending $30,000-$50,000 on a comprehensive study — and then use those findings for the next 12 months regardless of how much the competitive landscape or consumer context has shifted.
This model made sense when the minimum viable concept test cost $15,000 and took 8 weeks. Running multiple tests per year was logistically and financially impossible. When studies cost $200-$1,000 and take 48-72 hours, the model becomes a choice rather than a constraint — and it’s a choice that most teams are making by default rather than by design.
The Episodic Model vs. The Compound Model
The episodic model: one study per year, high cost, high production value deliverable, finding that ages over 12 months. Most teams learn something important once a year and rely on it for too long.
The compound model: 10-15 smaller studies per year, each tied to a specific decision, each feeding a searchable knowledge base that builds on previous research. Test #7 benefits from what tests #1-6 revealed. By year three of continuous concept testing, the questions you bring to each new study are sharper because you understand the category deeply — not from one annual synthesis, but from dozens of conversations conducted across different concept iterations, different consumer segments, and different competitive contexts.
The compound model costs the same as or less than the episodic model in dollar terms. Ten studies at $500 each is $5,000 — less than a single traditional focus group session. The difference is that the compound model produces 10 learning cycles instead of one, each informing a specific decision instead of becoming a reference document that ages.
How Each Study Feeds the Next
User Intuition’s Intelligence Hub stores every conversation in a searchable, indexed knowledge base. When a brand team tests a new concept six months after testing a related concept, the earlier study doesn’t disappear — it’s in the hub, retrievable by theme, by consumer segment, by specific language pattern. A researcher asking “what have we learned about how consumers think about sustainability in this category?” gets an answer drawn from every study the team has ever run, not just the most recent one.
This is how institutional knowledge compounds instead of depreciating. In the traditional model, research delivers insight that’s referenced for 90 days and then effectively disappears. In the compound model, every study adds to a base that grows more valuable over time.
The concept testing vs. focus groups comparison explores the methodological differences in more detail. But from a budget perspective, the most important thing the Intelligence Hub does is eliminate the 90% waste rate: the fraction of insight that disappears from organizational memory because it was delivered as a static document rather than stored in a queryable system.
Sample Budget Scenarios
These scenarios illustrate what different annual concept testing budgets can realistically accomplish — and where the tradeoffs are.
CPG Startup: $2,000 Annual Budget
Allocation:
- 4 early-stage concept screens at $200 each: $800 (identifies which ideas are worth developing)
- 2 development-stage tests at $500 each: $1,000 (refines the 2 concepts that passed screening)
- 1 pre-launch validation at $200: $200 (final go/no-go before production investment)
Result: 7 research touchpoints across your innovation funnel in a year, at a total cost equivalent to one traditional focus group session. You’ve validated ideas before investing in development, refined concepts before investing in production, and built a growing knowledge base of consumer language and preference data in your category.
What this doesn’t cover: Normative benchmarking (you won’t know how your concept scores against category norms), physical product testing (requires logistics budget beyond research), or the kind of stakeholder-ready deliverable that impresses investors or retail partners who expect a Kantar logo on the research.
Mid-Market Brand: $10,000 Annual Budget
Allocation:
- 8 early-stage concept screens at $200-$400 each: $2,400 (broad pipeline screening)
- 4 development-stage tests at $500-$1,000 each: $3,000 (serious development for strong concepts)
- 2 pre-launch validation studies at $1,000-$2,000 each: $3,000 (robust validation for final concepts)
- 1 post-launch consumer reaction study at $1,000: $1,000 (closes the learning loop)
Result: 15 research touchpoints across the year. Every significant innovation decision has consumer evidence behind it. The post-launch study closes the loop — you learn whether the consumer reactions in testing matched actual market behavior, which calibrates how you interpret future tests.
What this doesn’t cover: High-stakes global launch validation (needs specialized multi-market coordination), normative benchmarking against category databases, or physical in-home use tests for sensory products.
Enterprise Innovation Team: $50,000 Annual Budget
Here’s where the honest answer diverges from how most enterprise teams actually spend their research budgets.
How most enterprise teams spend $50,000 on concept testing: One comprehensive agency study ($35,000-$45,000) with multiple concept cells, segmentation analysis, normative benchmarking, and a polished 60-page deliverable. Remaining budget covers ad hoc follow-up surveys. The team learns from one major study and flies blind for most of the year.
How they should spend $50,000:
- 20 early-stage concept screens at $200-$400 each: $6,000 (continuous pipeline screening)
- 10 development-stage tests at $500-$1,000 each: $8,000 (rigorous development cycle)
- 5 pre-launch validation studies at $1,000-$3,000 each: $12,000 (final validation for leading concepts)
- 1 normative benchmarking study via Zappi or Quantilope: $8,000 (category benchmark for top concept)
- 1 agency study for the concept going to major launch: $16,000 (stakeholder-ready deliverable with full methodology)
Result: 37 research touchpoints — versus the one study in the traditional model. The single agency study is reserved for the one concept that cleared all the earlier gates and needs stakeholder packaging and normative benchmarking. Everything else runs on AI-moderated infrastructure at a fraction of the cost.
Total spend: $50,000. Learning cycles: 37 vs. 1. The concepts that reach launch have been tested, refined, tested again, and validated. The ones that failed screened out early, before development investment.
The Right Question Is Not “How Much Does Concept Testing Cost?”
The right question is: what does it cost you to make concept decisions without evidence?
A product concept that fails in market after an $800,000 packaging and formulation investment cost the wrong amount. A messaging campaign built around positioning consumers don’t recognize cost the wrong amount. A feature set that launched with the wrong emphasis because nobody asked consumers which capabilities actually mattered — that cost the wrong amount too.
The research literature on innovation failure is consistent: 60-80% of new products fail within two years of launch. The majority of those failures are predictable in consumer research — not because research can guarantee success, but because research consistently surfaces the specific concerns, misalignments, and missing elements that separate concepts consumers are excited by from concepts they’re indifferent to.
A $200 concept screen that catches a fundamental positioning problem before $500,000 of development investment doesn’t cost $200. It saves $500,000. A $1,000 study that identifies which of three packaging concepts drives the clearest premium signal before a packaging production run doesn’t cost $1,000. It’s a 200x return on the cost of a bad decision.
The reason most teams don’t think in these terms is that the cost of not doing research is invisible and diffuse, while the cost of research appears as a line item on a budget. The bad launch gets attributed to market timing, competitive response, or execution. The study that would have prevented it was never commissioned.
Questions to Ask Any Concept Testing Vendor
Before you sign a research contract, these questions separate vendors who will give you honest answers from those who will run you through a sales process.
What is the per-interview cost, all-in — including recruitment, incentives, moderation, and reporting? Agency quotes obscure per-interview economics. Make them do the math. A $30,000 study with 20 completed interviews costs $1,500 per conversation. Know what you’re paying per insight before you approve the budget.
Who owns the data, and can I access full transcripts? Some vendors provide only synthesized outputs. If you want to run your own analysis, build your own knowledge base, or reference specific verbatim quotes in a presentation, you need full transcript access. Clarify this before engagement.
What does analysis include — themes only, or verbatim quotes traced to specific findings? Themes without evidence are interpretation without accountability. Insist on verbatim quotes tied to findings. If a vendor tells you “consumers expressed concern about the price point,” you should be able to read the actual words consumers used to express that concern.
Is there a setup fee, minimum commitment, or subscription? User Intuition charges $20/interview with no setup fees and no monthly minimums. Some platforms look cheap per interview but layer on platform fees that change the economics. Understand the total cost of a single study.
How are participants recruited, screened, and verified? A concept test is only as good as the participants. Ask specifically about fraud prevention — bot detection, duplicate suppression, professional respondent filtering. Ask about screener methodology. A panel of 4M+ with multi-layer fraud detection delivers fundamentally different participant quality than a general opt-in list.
How does the methodology prevent leading questions? Non-leading language is harder than it sounds. Moderators and question authors consistently introduce leading language that pushes participants toward expected responses. Ask how the vendor calibrates against this. User Intuition’s methodology was developed against McKinsey research standards — a specific, verifiable claim.
The Transparency This Industry Has Always Needed
The reason this pricing breakdown doesn’t exist anywhere else isn’t because the information is proprietary. It’s because publishing it creates competitive pressure in a market that profits from opacity. When clients know that a $40,000 agency concept test allocates $12,000-$16,000 to account management overhead, $5,000-$8,000 to moderator fees, and $6,000-$12,000 to deck production — and that the core research act can be replicated for $200-$1,000 in 48-72 hours — they make different decisions.
Some of those decisions will involve traditional agencies. Not because the agency model is wrong, but because there are legitimate concept testing situations where the infrastructure matters: regulatory requirements, physical product testing, normative benchmarking, high-stakes global launches, stakeholder management requiring a recognized brand name. Those situations are real. They’re also narrower than the research industry would have you believe.
For the majority of concept testing decisions that brand teams, innovation leaders, and product managers face — go/no-go on a rough concept, positioning validation, messaging iteration, development prioritization — the expensive option is a choice driven by habit and opacity, not necessity.
User Intuition’s concept testing solution starts at $200 for 20 conversations. Results in 48-72 hours. Every study stored in a searchable Intelligence Hub that compounds over time. 98% participant satisfaction. No setup fees, no minimum commitments.
If you’ve been making concept decisions without evidence because you assumed research was out of reach, the assumption was wrong. The question is what you do with that information.
Start with 20 interviews. See what consumers actually tell you. Then decide whether your next concept decision should cost $200 or $40,000 — and why.