← Reference Deep-Dives Reference Deep-Dive · 8 min read

Protomonadic Concept Testing: When to Combine Monadic and Comparative

By Kevin, Founder & CEO

What Protomonadic Testing Is


Protomonadic testing combines two evaluation approaches in a single interview: monadic evaluation followed by comparative reveal.

Phase 1 (Monadic): Each participant sees and evaluates one concept. They do not know that other concepts exist. Their reactions, ratings, and qualitative feedback reflect their unanchored response to that concept on its own merits.

Phase 2 (Comparative): After completing the monadic evaluation, the participant is shown one or more alternative concepts. They evaluate the alternatives and then provide a direct comparison — which do they prefer and why?

The structural insight is that these two phases answer different questions:

  • Phase 1 answers: “How does this concept perform in isolation, the way a consumer would encounter it in the real world?”
  • Phase 2 answers: “When a consumer can see the alternatives, which do they prefer and what drives that preference?”

Neither phase alone gives you the full picture. A concept might score well monadically but lose in comparison. Another might score modestly on its own but win decisively when compared to alternatives. Both patterns have strategic implications.

When Protomonadic Outperforms Other Designs


vs. Pure Monadic

Pure monadic testing evaluates each concept with a separate group of participants. It produces clean, unbiased data for each concept. But it cannot tell you which concept participants would choose when they see the alternatives side by side.

Protomonadic adds the comparative layer without sacrificing monadic data quality. The Phase 1 data is collected before the reveal, so it is just as clean as pure monadic data.

Choose protomonadic over pure monadic when:

  • You are down to 2-3 finalist concepts and need a final recommendation
  • Stakeholders want both “how good is each concept?” and “which one wins?”
  • The competitive context matters — your concept will face alternatives at shelf or in market

vs. Pure Sequential (Comparative)

Pure sequential testing shows each participant multiple concepts in sequence and asks them to compare. The problem: the first concept anchors the evaluation of everything that follows. Order effects contaminate the data, and participants evaluate later concepts relative to earlier ones rather than on their own terms.

Protomonadic eliminates the anchoring problem for Phase 1 because participants see only one concept initially. The comparison in Phase 2 is deliberate and controlled rather than an artifact of presentation order.

Choose protomonadic over sequential when:

  • You need reliable standalone metrics for each concept (sequential cannot provide this)
  • You want to understand how comparison changes reactions (the “reveal effect”)
  • Concept differences are subtle enough that order effects would distort results

When to Use Something Else Entirely

Protomonadic is not always the right choice:

  • Early-stage screening of 4+ concepts: Too many concepts for one participant to evaluate meaningfully. Use pure monadic with separate cells.
  • Exploratory research: If you are still generating concepts rather than choosing between them, the comparative phase adds little value.
  • Budget constraints requiring minimum sample: Protomonadic requires all participants to complete both phases, which means your effective sample per concept is your total sample (everyone sees everything in Phase 2). If you can only recruit n=30 total, protomonadic works. If you need n=30 per concept for statistical reliability in Phase 1, you need n=60+ total for a two-concept test.

The Two-Phase Interview Structure


Phase 1: Monadic Evaluation

The participant sees a single concept. The interview follows standard concept evaluation methodology:

  1. Initial reaction capture. Open-ended: what are your first thoughts? What stands out? What is this offering?
  2. Comprehension check. Does the participant understand what the concept is and what it does? Miscomprehension at this stage means the concept has a communication problem.
  3. Attribute evaluation. Rate the concept on key dimensions — appeal, uniqueness, believability, relevance, purchase intent. These ratings are the monadic data.
  4. Deep probing. AI-moderated interviews probe 5-7 levels into the reasoning behind reactions. Why is this appealing? What specifically about the [stated attribute] made you feel that way? What would make this even more appealing?
  5. Improvement suggestions. What would you change? What is missing? This qualitative data feeds directly into concept refinement.

The participant completes Phase 1 believing this is the entire interview. There is no hint that alternatives are coming.

The Transition

This is the most methodologically sensitive moment in the interview. The transition must:

  • Clearly signal that additional concepts will be shown
  • Frame the comparison as a new task (not a correction of their Phase 1 response)
  • Avoid language that implies their Phase 1 answers were wrong or incomplete

An effective transition: “Thank you for that evaluation. I’d now like to show you some alternative approaches to this same [product/packaging/message]. I’d like you to look at each one and then share your thoughts on how they compare.”

AI moderation handles this transition consistently. Human moderators may inadvertently signal which concept they expect the participant to prefer through vocal tone or body language. The AI moderator presents the transition identically for every participant.

Phase 2: Comparative Evaluation

The participant now sees the alternative concept(s). The evaluation follows a different structure:

  1. Alternative evaluation. Brief initial reactions to each new concept. Not as deep as Phase 1 — the goal is comparative context, not full monadic depth on every concept.
  2. Direct comparison. “Having seen all of these options, which do you prefer overall?” Follow with: “What makes [chosen concept] your preference?”
  3. Attribute-level comparison. “Which is most [appealing/unique/believable/relevant]?” This reveals whether the same concept wins across all attributes or whether different concepts lead on different dimensions.
  4. Switching analysis. If the participant preferred a different concept in Phase 2 than they evaluated in Phase 1, probe the switch: “What about seeing the alternatives changed your perspective?”
  5. Final preference strength. “How strong is your preference? Would you be satisfied with your second choice, or is there a clear winner?”

The Reveal Effect: How Comparison Changes Preferences


One of the most valuable outputs of protomonadic testing is the reveal effect — the measurable shift in preference and evaluation that occurs when alternatives become visible.

Common Reveal Effect Patterns

Concept holds after reveal. Participants evaluated Concept A monadically and rated it highly. After seeing Concept B, they still prefer A. This is strong validation — the concept withstands competitive comparison.

Concept collapses after reveal. Participants rated Concept A well monadically, but after seeing Concept B, they switch preference. This suggests Concept A performs adequately in isolation but has a competitive vulnerability. It is “good enough” until something better appears.

Concept strengthens after reveal. Participants gave Concept A moderate monadic ratings, but after seeing the weaker Concept B, they appreciate A more. Comparison provided context that elevated their assessment. This pattern sometimes indicates that the concept’s value proposition is relative — it works because it is better than alternatives, not because it is independently compelling.

Preferences diverge. Phase 1 ratings are similar across concepts, but Phase 2 reveals a clear winner. The concepts were hard to differentiate in isolation but easy to differentiate in comparison. This is common with packaging variants where the differences are visual and subtle.

Why the Reveal Effect Matters Strategically

The reveal effect tells you how your concept will perform in market. Products rarely exist in isolation — consumers encounter alternatives at shelf, in search results, on review sites. A concept that wins monadically but loses comparatively is a concept that will underperform in competitive markets.

Conversely, a concept with modest monadic scores that wins decisively in comparison may be more competitive than its standalone metrics suggest. The monadic data sets your expectations; the comparative data predicts competitive performance.

Sample Design for Protomonadic Testing


Minimum Sample Sizes

For a two-concept protomonadic test:

  • Total sample: n=60-100 participants
  • Phase 1 cell sizes: n=30-50 per concept (each participant evaluates one concept monadically)
  • Phase 2: All participants see both concepts

For a three-concept test:

  • Total sample: n=90-150 participants
  • Phase 1 cell sizes: n=30-50 per concept
  • Phase 2: All participants see all three concepts

Randomization Requirements

  • Phase 1 assignment must be randomized — participants are randomly assigned to their monadic concept
  • Phase 2 presentation order must be counterbalanced — half see concept B first, half see concept C first (to control for order effects within the comparative phase)
  • No repeat of Phase 1 concept as the first concept shown in Phase 2 — if a participant evaluated Concept A monadically, show them Concept B first in Phase 2 to maximize the “fresh eyes” comparison

Cost Implications

At $20 per interview, a two-concept protomonadic test with n=80 total costs $1,600. A three-concept test with n=120 costs $2,400. This makes protomonadic testing economically practical for decisions that previously defaulted to cheaper but less informative comparative-only methods.

How AI Moderation Handles the Two-Phase Transition


The two-phase structure of protomonadic testing creates a moderation challenge: the interviewer must seamlessly shift from deep monadic exploration to structured comparative evaluation. This transition is where human-moderated protomonadic studies most often break down.

Common human moderator issues:

  • Rushing Phase 1 because they know Phase 2 is coming, resulting in shallower monadic data
  • Telegraphing the reveal through framing or tone, which can make participants hedge their Phase 1 responses
  • Inconsistent transition language across participants, introducing variability
  • Difficulty managing the expanded interview length while maintaining engagement

AI moderation eliminates these issues. The moderator follows the same protocol for every participant — full depth in Phase 1, clean transition, structured Phase 2. The interview length extends naturally because the AI moderator manages pacing and engagement without fatigue.

Practical Applications


Packaging A/B Testing

Two packaging designs for the same product. Phase 1 captures how each design communicates on its own. Phase 2 reveals which design consumers choose when they see both — simulating the shelf decision.

Messaging Variants

Two or three positioning messages for the same product. Phase 1 identifies which message resonates independently. Phase 2 reveals whether consumers see meaningful differences or view the messages as interchangeable.

Concept Finalists

The last 2-3 concepts surviving from a broader screening process. Phase 1 provides final standalone metrics for each. Phase 2 gives the executive team a defensible preference ranking for the launch decision.

Protomonadic testing is one of several methodological choices covered in the monadic vs. sequential testing guide. For teams ready to implement, User Intuition’s concept testing solution supports the two-phase interview structure with consistent AI moderation across both phases.

Frequently Asked Questions

Protomonadic testing evaluates each concept individually first, then reveals alternatives for direct comparison. Pure monadic testing evaluates each concept in isolation with no comparison — capturing unanchored reactions but missing relative preference. Sequential testing exposes all concepts before rating any — introducing comparison bias into every reaction. The protomonadic design captures both unanchored reactions (from the monadic phase) and relative preference (from the comparison phase) in a single study, which neither pure approach can do.
The reveal effect is the shift in preference that occurs when participants see alternatives after rating a concept in isolation. When a concept that scored well in isolation looks less differentiated after comparison, the reveal effect indicates that its appeal was relative rather than absolute — it may not hold up at shelf where competitors are visible. When a concept that scored modestly in isolation outperforms alternatives in comparison, it suggests strong differentiated value that was undersold in the monadic phase. Distinguishing these patterns is essential for launch decisions.
Protomonadic outperforms when the decision requires both confidence that a concept has genuine standalone appeal and evidence about which of several alternatives is preferred. This is most common in packaging redesign, new product line extensions, and positioning variant testing — situations where launching the wrong choice is expensive and the team needs both 'is this good?' and 'which is better?' answered simultaneously. Pure monadic testing answers only the first question; sequential testing compromises both.
User Intuition's AI moderation manages the reveal moment — transitioning participants from independent concept evaluation to comparative assessment — in a way that captures genuine initial reactions before introducing comparison bias. The platform can administer protomonadic studies at $20 per interview across the full 4M+ panel, returning both phase-level data (monadic scores and verbatim reactions) and comparison-phase preference data in 48-72 hours. Teams can run a 60-participant protomonadic study across two or three concepts for $1,200 — a fraction of the cost of a facility-based concept test with equivalent depth.
Get Started

Put This Research Into Action

Run your first 3 AI-moderated customer interviews free — no credit card, no sales call.

Self-serve

3 interviews free. No credit card required.

See it First

Explore a real study output — no sales call needed.

No contract · No retainers · Results in 72 hours