Concept testing on User Intuition is an eight-step workflow: Create Study, Customize Plan, Review Goals, Setup Interview, Test Conversation, Launch, Invite Participants, and Review Insights. The product is built so a brand manager, a product marketing lead, or a CPG innovation director can launch a concept-evaluation study in an afternoon, then watch consumer transcripts, satisfaction-rated calls, and synthesized reports show up inside the Intelligence Hub within 48-72 hours. Pricing sits at roughly $20/interview on the Pro plan, and the 4M+ research panel covers 50+ languages, which means a continuous test-refine-retest program is viable at mid-market budgets. This tutorial walks through each step with the verbatim labels you will see in the product and the choices that matter most for concept testing specifically.
For the broader methodology (how to design a test, how to write the stimulus, when surveys or focus groups are the wrong tool), see the methodology guide. For pricing and comparison against Bases, Zappi, and Suzy, see the concept testing platform page. This piece stays in the workflow lane: what to click, where to click it, and why each setting matters when the cohort you care about is category buyers reacting to packaging, claims, ad creative, or product variants pre-launch.
What is concept testing, briefly?
Concept testing is the practice of presenting a concept (product idea, packaging design, message, brand name, ad creative, or pricing model) to target consumers before launch and understanding not just whether they react positively, but why. It sits between two failed substitutes: monadic surveys, which collect a top-2-box appeal score but never reach the reasoning underneath, and traditional focus groups, which produce 8 to 12 socially-mediated reactions contaminated by groupthink. Neither produces decision-grade intelligence.
Done well, concept testing answers five questions every brand or innovation team needs answered before committing budget. Do consumers want it (appeal)? Do they understand what it is (comprehension)? Does it solve a problem they have (relevance)? Is it meaningfully different from what they already use (differentiation)? Would they buy it, and if not, why (purchase intent)? Most concept testing methods capture appeal and ignore the other four.
The mechanism that makes it work is depth. A 30-minute conversation with a category buyer, with 5 to 7 levels of structured laddering, surfaces decision drivers that a 2-minute survey will never reach. Premium-vs-discount perception, comprehension failures, claim resonance by segment, shelf impact, these are the patterns underneath the appeal score.
Modern AI-moderated platforms compress the timeline from 6 to 12 weeks down to 48-72 hours and the cost from $25K to $75K per study down to $200 floor pricing. That changes program design. Instead of one big quarterly test, brand and innovation teams run 50 to 100 interviews on every iteration, feed transcripts into a searchable Intelligence Hub, and watch concept intelligence compound study by study.
Why doing this on User Intuition is different
Most concept testing tooling makes you choose between depth and speed. Surveys are fast and shallow. Focus groups are deeper but biased and slow. User Intuition is built to deliver both, and the product flow is shaped by that goal.
The AI moderator runs 30-minute structured laddering conversations with each consumer and probes 5 to 7 levels deep on appeal, comprehension, relevance, differentiation, and purchase intent. Every conversation runs the same methodology, so a study with 50 interviews and a study with 300 are directly comparable. No moderator drift across day three of fielding. Consumers report higher candor with AI than with vendor-affiliated researchers, and the 98% participant satisfaction rate reflects how the experience lands.
Recruitment is built in. The 4M+ research panel covers 50+ languages and lets you target by category usage, purchase frequency, geography, and demographics without building a list. For B2B concept testing, your customer or prospect list flows in via the CRM Contacts integration. Drop the embed widget into your post-purchase email and concept reactions flow continuously as part of a brand-tracking program.
Synthesis runs in the same product. Once interviews complete, the Reports tab generates an Executive Summary, Top Insights, Detailed Analysis, and Recommendations, all grounded in verbatim consumer quotes. The Intelligence Hub layers across studies, so concept tests from different quarters surface compounding patterns that any single-study report would miss. Pricing at $20/interview keeps the stack sustainable as a continuous program where agency programs die after one or two cycles. See the concept testing platform page for the full positioning.
The 8-step walkthrough
The product flow is the same for every study type. The choices below are tuned for concept testing.
Step 1: Create Study
Open the New Study screen and you will see six pre-built templates: Win/Loss Analysis, Churn Analysis, NPS and CSAT, Customer Onboarding, Brand Health, and Custom Design. For concept testing, pick the Custom Design card. The selected card highlights with a dark background. Click Save & Continue.
Custom Design is the right starting point because concept testing covers six distinct sub-types (product, packaging, messaging, naming, ad creative, pricing), each with different stimulus design and probe techniques. The Custom Design template gives Charles room to shape the conversation flow, screener defaults, and learning goals around the specific variants you are testing rather than slotting you into a fixed template. See Choose Your Study Type for the full template comparison.
Concept-testing-specific tip: hold the “What do people think of this concept?” framing in your head before you click. Custom Design is the canvas; the framing keeps Charles building toward consumer reaction depth instead of generic feedback. The same template handles single-concept evaluation and multi-variant comparison.
Step 2: Customize Plan
The Customize Plan screen opens a chat with Charles, our AI researcher. Type your context into the chat box and Charles drafts the research plan: Objective, Background, North-Star Learning Goal, Key Sub-Questions, Conversation Flow, and Interviewer Guidelines, all in a panel on the right. For concept testing, use the chat option (typing) rather than voice; you will paste in concept descriptions, claims, and stimulus details that are easier to compose in writing.
Give Charles four things. First, the concept type: “We are testing three packaging variants for a premium snack product.” Second, the variants themselves: name, visual or claim description, price point, callouts. Third, the target segment: “Category buyers, women 25 to 54, premium-snack purchase frequency at least monthly, North America.” Fourth, the success criteria: “We will pick the variant that scores highest on relevance and differentiation, and kill any variant where comprehension is below 60%.” Charles returns a plan probing initial reaction, comprehension, relevance, differentiation, purchase intent, and improvement.
Concept-testing-specific tip: tell Charles which questions you already think you know the answer to. He uses your hypotheses to design probes that test or break them. Read the Customize Plan docs for the full context questions.
Step 3: Review Goals
The Review Goals screen opens the research plan in a rich text editor. Study Name sits at the top (Charles suggests one, like “Fall Packaging Concept Test, Premium Snack”; rename it to fit your naming convention). Every section, Objective through Interviewer Guidelines, is editable. Highlight, retype, add, delete. Changes save automatically.
This is the step where most teams adjust the learning goals to cover both the comparative ranking and the why behind preferences. The default plan asks about appeal and intent, but for high-stakes packaging or campaign decisions you want explicit coverage of what drives the preference. Add sub-questions like “What specifically about Variant B makes it feel more premium than Variant A?” and “Is there anything in Variant C that confuses you or feels off?” Concept tests fail when they only collect the ranking; the why is the part that turns the result into a creative brief.
Concept-testing-specific tip: keep Key Sub-Questions to 4 to 6 themes with 2 to 4 questions each. Plans with 8+ themes produce interviews longer than 25 minutes and completion rates drop sharply. The Review Study docs have the full editing toolbar reference. Click Save & Continue when the plan reads the way you want.
Step 4: Setup Interview
The Setup Interview screen handles the participant-facing experience. Two voice cards: Elliot (Male, American) and Clara (Female, American). Click the play button on each card to preview. Pick the one whose tone fits your category audience and click the card to select. Below the voice cards, the Default Mode toggles control format: Chat, Audio, or Video. Audio is the recommended default for concept testing because voice transcripts capture tone, hesitation, surprise, and emotional reactions to visual or claim stimuli that chat compresses out.
Below the format toggles is the screener questions section. Add the highest-leverage concept testing screeners: category usage (“How often do you purchase {category}?”), recent purchase behavior (“Which of these brands have you bought in the last 6 months?”), and for B2B concept testing, decision-making role. Recruiting on category usage and behavior is the difference between a clean read on your target segment and a contaminated study averaged across the wrong audience.
Concept-testing-specific tip: pick Elliot for category audiences that skew male and Clara for audiences that skew female. The match between moderator voice and consumer demographic measurably improves rapport and depth. Read the Setup Interviewer docs for the full mode-selection guide.
Step 5: Test Conversation
The Test Conversation screen prompts you with “Ready to test your study?” Click Start Test Conversation and you experience the full participant flow: greeting, introduction, research questions, AI probing, and wrap-up. Test conversations do not count against usage limits. After the test, a feedback screen asks “How was the conversation?” with Good or Needs Work options.
This is the most important step for concept testing specifically. Concept tests are high-stakes and tone misalignment can taint reactions. Run two or three test conversations before launch. Run the first as an enthusiastic category buyer who likes the lead variant, and listen for whether the AI probes hard enough on what drives the preference (it should not accept “I like the colors” as final). Run the second as a skeptical buyer who finds the concept confusing, and listen for whether the AI probes past surface dismissal into the comprehension or differentiation gap. Run the third as a mid-funnel participant to confirm the AI handles ambiguous reactions without forcing a position.
Concept-testing-specific tip: if the AI accepts a surface answer, go back to Step 3 and tighten the Interviewer Guidelines with instructions like “Probe at least three levels deep on initial reactions, including category context and specific design or claim elements driving the response.” Then re-test. The Test Conversation docs cover what to look for.
Step 6: Launch
The Launch screen shows three summary cards (Study Type, Interviewer, Interview Type), the Study Name, and the full Research Plan Preview. Final review checklist sits in three accordions: Research Plan, Interviewer Settings, Study Configuration. When everything looks right, click the Save and Launch button (the one with the rocket icon). Your study goes live immediately. You are redirected to the Study Dashboard.
The detail brand managers miss: launching does not automatically send invitations. The study is live and ready to receive participants, but recruitment happens in the next step. This is intentional, because most teams want to attach concept stimulus assets, review with the cross-functional team, and then trigger recruitment on a specific day.
Concept-testing-specific tip: launch Wednesday morning and push recruitment the same day. Consumer response rates skew highest mid-week, so a Wednesday launch gives you 2 to 3 strong response days before the weekend. Attach concept variants as stimulus assets before pushing recruitment, so the first wave reacts to the same material as the last. The Launch docs cover what can and cannot change post-launch.
Step 7: Invite Participants
This step opens after launch. From the Study Dashboard, the Invites tab plus the Actions dropdown gives you four invitation methods that map directly to concept testing recruitment patterns.
Research Panel is the workhorse for CPG concept testing. It opens recruitment from the 4M+ panel across 50+ languages, with country and language selection happening here for multilingual tests. Filter by category usage and purchase frequency, not demographics alone. Invite Your Customers lets you import contacts manually, via CSV, or via a synced CRM (Salesforce, HubSpot). This is the right method for B2B concept testing or a customer-only test on a brand line extension. Share Link generates a unique URL you can paste into a personalized email or a Zapier workflow. Embed Widget drops always-on concept reaction collection into your customer portal.
Concept-testing-specific tip: combine methods. Panel for category buyers without contact info, customer invites for line-extension concepts. For a multi-region test (NA, LATAM, EU), set country and language at the panel step and run parallel cohorts inside the same study. The Recruiting overview covers all four methods.
Step 8: Review Insights
Once interviews complete (first ones inside hours, full cohorts within 48-72 hours), three tabs hold the output. The Calls tab lists every interview with email, status, date, duration, quality rating (High, Medium, Poor), and end reason. Click any row to expand the full transcript with audio playback. The Reports tab holds the AI-synthesized analysis: Executive Summary, Total Participants, Top Insights at a Glance, Detailed Analysis with verbatim quotes, and Recommendations. Click Generate Report once you have at least 2 completed interviews; click Regenerate Report as more land.
The third surface is the Intelligence Hub, accessed from the left navigation. Each Hub session is a workspace where you sync studies, upload supplementary files (PDFs, CSVs, prior tracker data), and ask natural-language questions across everything in the room. For a continuous concept testing program, this is the compounding layer.
Concept-testing-specific tip: create a single Intelligence Hub session called “Concept Intelligence” at the start of the year, sync every concept test into it, and keep adding sources. Then ask cross-study questions like “What design elements have driven premium perception across our last six packaging tests?” The session-level pattern recognition is the part that single-study reports cannot replicate. The Intelligence Hub docs cover the output surface.
Mini case study
Turning Point Brands ran the workflow above on a packaging concept test for a line extension to ALP, their breakout nicotine pouch brand. The team set up the study in a single afternoon: Custom Design at Step 1, three packaging variants briefed to Charles at Step 2, learning goals tightened at Step 3 to include premium-vs-discount perception and shelf-differentiation probes, Audio mode and category-usage screeners configured at Step 4, two test conversations run at Step 5, study launched mid-week at Step 6. Recruitment used the Research Panel filtered to nicotine pouch buyers with monthly-or-greater purchase frequency. All interviews completed within 72 hours.
The Reports tab and the Intelligence Hub surfaced a finding that contradicted the team’s internal favorite. The variant the brand and design team had championed scored highest on aesthetic appeal but lowest on premium perception. Consumers described it as “looks fine for a discount option” while describing the second variant as “looks like the expensive one on the shelf.” The team killed the favorite and advanced the second variant into production.
“We tested three positioning concepts in 72 hours and discovered one was perceived as a commodity data pipe despite premium pricing. User Intuition saved us from a costly repositioning miss before we launched the campaign.”
Eric O., COO, RudderStack
What gets the best results from concept testing on User Intuition?
Tips that consistently produce concept-decision movement.
Define the decision before designing the stimulus. Vague definitions (“we want to understand how consumers react”) produce vague findings that confirm whatever the team already wanted to do. Specific definitions (“we will pick A or B based on relevance scoring among 25-to-34-year-old female category buyers”) tell you what to measure and what would change the outcome.
Recruit on category behavior, not demographic boxes. A consumer who says they buy premium snacks monthly is a different respondent from one who checks “25-34 female” on a panel registration form. Screen on category usage and purchase frequency at Step 4. Every screener should require demonstrated category relevance.
Run two or three test conversations every time. One enthusiast, one skeptic, one mid-funnel. Concept tests are high-stakes and tone misalignment taints reactions. The single highest-leverage 30 minutes you can spend before launch is catching shallow probing while it is still cheap to fix.
Pick voice over chat for concept testing. Voice transcripts surface tone, hesitation, surprise, and emotional reactions to stimuli that chat compresses out. Reserve chat for international cohorts where typing is a better experience.
Test multiple variants in one study. Multi-concept testing with order rotation produces direct comparison data that single-concept tests run sequentially cannot. It is cheaper per concept and faster to field. Two to four variants per study is the sweet spot. See the comparison of concept testing platforms for how this differs from the monadic survey approach used by most legacy tools.
Iterate test-refine-retest within the same week. At $20/interview on the Pro plan with 48-72 hour turnaround, two rounds of iteration are cheaper than a single round of agency research. The team that tests once is hoping the first version is right. The team that tests twice has validated the fix. See the cost breakdown for continuous concept testing for the budgeting case.
Sync the Intelligence Hub once per month. Every completed concept test should land in the same Concept Intelligence session on a regular cadence. Cross-study pattern recognition is the highest-leverage output User Intuition produces, and it only works if every study lands in the same session. For teams running concept testing alongside upstream idea work, the idea validation methodology guide covers how the same compounding-intelligence logic applies earlier in the funnel.
Route insights inside 7 days. Concept findings degrade fast. Route specific findings to specific owners (brand, creative, packaging design, product development) with a defined SLA. The studies that change a launch decision are the ones whose findings reach a decision-maker the same week.
FAQ
The frontmatter FAQ block above answers the most common questions. The full canonical answers also surface inline in the workflow above.