Study design is the highest-leverage activity in user research. A well-designed study with mediocre execution produces useful insights. A poorly designed study with brilliant execution produces confident nonsense. Yet study design is consistently under-invested because teams are rushing to start recruiting before they have clarified what they actually need to learn.
Templates solve this by encoding proven study designs into reusable frameworks. They reduce planning time from days to hours while ensuring methodological consistency across studies, researchers, and teams. For organizations that democratize research — enabling product managers and designers to run studies through AI-moderated platforms like User Intuition — templates are the mechanism that maintains quality without requiring researcher involvement in every study.
This playbook provides complete templates for the five study types that cover 80% of user research needs. Each template is ready to adapt for your specific context.
How Should Discovery Research Be Designed?
Discovery research explores problem spaces, user needs, and market opportunities before solutions exist. It is the most strategically valuable and most commonly under-resourced research type because it answers questions the organization has not yet learned to ask.
Research objective template. Frame discovery objectives around the decisions the research will inform, not the topics it will explore. Weak objective: “Understand user needs around collaboration.” Strong objective: “Determine which collaboration pain points represent the largest opportunity for our Q3 product investment, measured by frequency, severity, and willingness to adopt new solutions.” The strong framing connects research directly to a decision, which keeps the study focused and makes findings immediately actionable.
Participant criteria template. Discovery research requires participants who represent the target user population broadly enough to reveal segment-level patterns. Define criteria along three dimensions: demographic or firmographic (role, company size, industry), behavioral (currently uses collaboration tools, manages a team of 5+, changed tools in the last 12 months), and attitudinal (considers collaboration a pain point, is actively seeking better solutions). Include exclusion criteria to filter out participants who cannot provide relevant insight: “Exclude participants who have never used digital collaboration tools” or “Exclude participants in companies with fewer than 10 employees.”
Sample size framework. Traditional discovery research uses 15-25 participants, which reveals individual patterns but rarely provides segment-level confidence. AI-moderated discovery with 50-100 participants reveals how different user segments experience the same problem space differently — a strategic insight that small-sample research cannot produce. For organizations serving multiple user segments, run 30-50 participants per segment to enable meaningful cross-segment comparison. At $20 per interview, a 100-participant discovery study costs $2,000 and completes in 48-72 hours.
Discussion guide structure. Structure the guide in four phases, spending approximately 25% of interview time on each. Phase 1: Context — understand the participant’s role, responsibilities, and current workflow. Phase 2: Pain points — explore frustrations, workarounds, and unmet needs in the target domain. Phase 3: Current solutions — map the tools, processes, and alternatives participants currently use and how they evaluate them. Phase 4: Opportunity exploration — present broad solution concepts or capability descriptions and gauge interest, concerns, and perceived value. Each phase includes 2-3 core questions with probing guidelines.
Analysis framework. Discovery analysis should produce three outputs: a problem priority matrix (pain points ranked by frequency and severity across participants), a segment map (how different user groups experience the problem space differently), and an opportunity assessment (which problems represent viable product opportunities based on participant willingness to adopt new solutions). Code transcripts along these three dimensions rather than creating open-ended theme taxonomies that proliferate categories without clarifying priorities.
What Makes an Effective Usability Study Template?
Usability research evaluates how well a specific design or product supports user tasks. The template must balance structured task observation with enough flexibility to capture unexpected interaction patterns.
Research objective template. Usability objectives should specify the tasks being evaluated, the success criteria, and the decisions that will be made based on results. Template: “Evaluate the [feature/flow] experience for [user type] performing [specific tasks]. Identify usability barriers that prevent task completion, measure task efficiency relative to benchmarks, and determine which issues require resolution before launch.” This framing ensures the study produces actionable prioritization rather than a generic list of observations.
Task design template. Design 4-6 tasks that cover the critical user paths under evaluation. Each task follows a consistent structure: Scenario context (a realistic situation that motivates the task), Task goal (what the participant needs to accomplish), Success criteria (how completion is defined and measured), and Time benchmark (expected completion time for comparison). Example: “Scenario: You just received an email notification that a team member shared a project with you. Task: Find the shared project and add a comment on the first item. Success: Comment appears on the correct item. Benchmark: Under 90 seconds.”
Participant criteria template. Usability research requires participants who represent actual users in terms of technical proficiency, domain knowledge, and product familiarity. For evaluating existing products, recruit a mix of experience levels: new users (first 30 days), regular users (active weekly for 3+ months), and power users (daily users who use advanced features). For evaluating new designs, recruit participants who match the target user profile but have not seen the design before. Avoid internal employees and stakeholders who cannot provide naive reactions.
Sample size framework. Traditional usability testing uses 5-8 participants based on the Nielsen Norman finding that 5 users find 85% of usability issues. This framework applies when the goal is issue identification. When the goal is measuring task success rates, efficiency, or comparing designs with statistical confidence, larger samples are needed. AI-moderated usability studies with 30-50 participants provide reliable task success rate data. For A/B design comparisons, 50-100 participants per variant produce statistically meaningful differences.
Observation and scoring template. Create a standardized observation grid for each task: Task completion (binary: completed/failed), Completion time, Error count and type, Assistance needed (questions asked, confusion expressed), Satisfaction rating (post-task), and Verbatim quotes capturing the participant’s experience. This structured scoring enables quantitative analysis of qualitative observations, making it possible to track usability improvement across design iterations.
Severity rating framework. Classify each identified usability issue using a consistent severity scale: Critical (prevents task completion, no workaround exists), Major (significantly slows task completion or requires workaround), Minor (causes momentary confusion but does not prevent or significantly slow completion), and Cosmetic (noticed but does not affect task performance). This classification drives prioritization and helps stakeholders understand which issues require immediate attention versus future iteration.
How Do You Template Concept Testing for Reliable Results?
Concept testing evaluates product ideas, design directions, or feature proposals before committing development resources. The template must prevent the two most common concept testing failures: false positives from leading presentation and false negatives from unclear concept communication.
Research objective template. Concept testing objectives should specify what dimensions of the concept are being evaluated and what constitutes a passing result. Template: “Evaluate [concept description] with [target users] to determine: comprehension (do users understand what it is and does), perceived value (would they use it and why), differentiation (how it compares to current solutions), and adoption barriers (what would prevent use). The concept proceeds to development if [specific threshold] of participants express genuine adoption intent with identifiable use cases.”
Concept presentation standards. How concepts are presented dramatically affects results. Use consistent presentation standards: Concepts should be described at the same level of fidelity (all high-fidelity mockups or all verbal descriptions — never mix). Present concepts in neutral language that describes what the concept does without implying it is good. Include both the concept and its limitations or constraints. Randomize presentation order when testing multiple concepts. Allow participants to examine the concept before asking evaluation questions.
Evaluation question sequence. Follow a consistent sequence that moves from comprehension to evaluation to comparison. First, verify comprehension: “In your own words, what is this and what does it do?” Miscomprehension invalidates all subsequent responses. Second, assess value: “How would this fit into your current workflow? What problems would it solve?” Third, evaluate differentiation: “How is this different from what you currently use?” Fourth, identify barriers: “What concerns or hesitations do you have about this?” Fifth, assess adoption intent: “How likely are you to try this? What would need to be true?” This sequence prevents enthusiasm contamination — starting with comprehension checks before introducing evaluative framing.
Multi-concept comparison template. When testing multiple concepts, use a structured comparison framework: Present concepts sequentially with evaluation after each, then present all concepts simultaneously for comparative assessment. For each comparison dimension (value, feasibility, preference), use forced-rank rather than absolute ratings — forced ranking reveals relative preference that rating scales obscure. Close with: “If you could only use one of these, which would it be and why?” The forced choice with explanation reveals the decision drivers that inform concept prioritization.
Analysis framework for concept testing. Analyze concept test results across four dimensions: Comprehension rate (percentage of participants who correctly described the concept), Net appeal (positive minus negative initial reactions), Adoption intent with qualification (separating genuine intent backed by specific use cases from polite interest), and Comparative preference with reasoning (which concept wins and why, disaggregated by user segment). These dimensions produce a decision-ready assessment rather than a collection of qualitative impressions.
How Should Satisfaction Research Be Templated for Ongoing Use?
Satisfaction research runs continuously rather than as one-off studies, making template consistency critical. Findings from this quarter must be directly comparable to findings from last quarter, which requires standardized methodology across waves that only well-designed templates and consistent execution can provide.
Research objective template for tracking studies. Objectives for ongoing satisfaction research focus on trend detection rather than snapshot assessment. Template: “Measure user satisfaction across [specific experience dimensions] for [user segments] to detect quarter-over-quarter changes in satisfaction drivers, identify emerging pain points before they affect retention metrics, and prioritize experience improvements based on user-stated impact. Results will be compared to Wave [N-1] findings from [date].”
Longitudinal question consistency. Core tracking questions must remain identical across waves to enable trend analysis. Design 8-10 core questions that will not change, covering: overall satisfaction, key experience dimensions (onboarding, daily use, support, value perception), recommendation likelihood with qualitative explanation, and the single most important improvement request. Reserve 20-30% of interview time for wave-specific questions that explore current issues or test new hypotheses.
Segment tracking template. Define consistent segment definitions across waves: user tenure (new, established, veteran), usage intensity (light, regular, power), and organizational role (individual contributor, manager, executive). Track satisfaction trends within each segment separately because aggregate trends mask segment-specific changes. A product that shows stable overall satisfaction may be losing power users while gaining new users — a pattern invisible in aggregate data but critical for retention strategy.
Sample size for tracking studies. Each wave needs sufficient participants per segment to detect meaningful changes. Minimum 30 participants per segment per wave enables detection of 15-20% shifts in sentiment themes. For three segments tracked quarterly, this means 90-120 participants per wave — a study that costs $1,800-$2,400 on AI-moderated platforms and would cost $30,000-$50,000 through traditional moderation. The economics of AI moderation make continuous tracking feasible for organizations that could never justify the cost of traditional longitudinal research.
Trend reporting template. Report tracking results in a standardized format: Executive summary (3-5 bullet points of notable changes), Dimension-by-dimension trend visualization (showing movement across waves), Segment-specific findings (highlighting where different user groups are moving in different directions), Emerging themes (new topics that appeared in this wave), and Priority recommendations (what should change based on the data). Consistent reporting format enables stakeholders to quickly orient and focus on what has changed.
AI-moderated platforms are particularly valuable for satisfaction tracking because they maintain identical methodology across waves — the same question phrasing, the same probing depth, the same analysis structure — eliminating the methodological variation that makes traditional longitudinal research noisy. Explore how consistent methodology at scale works at User Intuition.
What Template Structure Works for Competitive Perception Research?
Competitive research maps how users perceive, evaluate, and choose between alternatives in your market. The template must capture genuine competitive perception without biasing participants toward your product.
Research objective template. Template: “Map competitive perception across [market segment] to determine: how users categorize available options, what evaluation criteria drive selection decisions, how [our product] is perceived relative to [specific competitors], and what unmet needs create positioning opportunities. Findings will inform [specific strategic decisions: positioning, messaging, feature prioritization, pricing].”
Participant criteria for competitive research. Recruit three participant groups for comprehensive competitive insight: Current users of your product (to understand why they chose you and what they compare you against), Current users of competitor products (to understand what they value and what they perceive as your weaknesses), and Recent evaluators who considered multiple options (to understand the decision process itself). Equal representation across groups prevents confirmation bias.
Competitor neutrality protocol. The study must not reveal which company commissioned the research. Use third-party panel recruitment, present all competitors equally, and avoid questions that position any option favorably. Reveal sponsorship only after the substantive interview is complete if participants ask. AI-moderated platforms enforce this neutrality naturally because there is no moderator whose body language or tone might reveal preference.
Evaluation framework questions. Use a consistent framework to map competitive perception: Category understanding (“How would you describe the different types of [product category] available today?”), Evaluation criteria (“When you were choosing a [product], what were the most important factors?”), Perception mapping (“For each of these [3-5 competitors], tell me what they are best at and what they are worst at”), Decision drivers (“What was the single most important factor in your final decision?”), and Unmet needs (“What can none of the current options do that you wish they could?”).
Analysis framework for competitive research. Produce three strategic outputs: A competitive perception map showing how each competitor is positioned in users’ minds across the most important evaluation dimensions, A decision driver ranking showing which factors actually determine selection (often different from stated evaluation criteria), and A white space analysis identifying unmet needs that no current competitor addresses. These outputs directly inform product strategy, positioning, and messaging decisions.
Running competitive perception research at scale — 150-300 participants across multiple segments and competitor user groups — produces statistically defensible competitive intelligence. At $20 per interview on platforms like User Intuition, this costs $3,000-$6,000 and completes in 48-72 hours, enabling quarterly competitive tracking that was previously economically impossible for most organizations.
Frequently Asked Questions
How often should user research templates be reviewed and updated?
Review templates quarterly based on accumulated learnings from 5-10 studies per template. Assess whether questions consistently produce actionable insights, whether sample sizes are adequate for the analysis framework, and whether the reporting format resonates with stakeholders. Templates should evolve as your understanding deepens. After six months of continuous use on User Intuition, most teams find their templates have been refined into highly efficient instruments tailored to their specific product domain and user segments.
What is the minimum sample size for reliable user research findings?
Sample sizes depend on study type and analytical goals. For usability issue identification, 5-8 participants find 85% of issues. For attitudinal research requiring segment comparison, 30-50 participants per segment provide reliable theme prevalence data. For competitive perception research, 50-100 participants per competitor group enable statistically meaningful comparisons. AI-moderated platforms at $20 per interview make larger samples practical: a 100-participant study costs $2,000 and completes in 48-72 hours.
Can non-researchers use these templates to run credible studies?
Yes, especially on AI-moderated platforms where methodology is built into the interview process. The templates provide the research design framework (objectives, questions, criteria, analysis approach) while the platform handles moderation technique with consistent 5-7 level laddering. This is how User Intuition enables democratized research: the rigor lives in the template and the platform, not in the individual running the study. Researchers should review a sample of democratized study outputs periodically to ensure quality.
How do you build a question library from accumulated study experience?
After each study, identify which questions consistently generated the most actionable insights and which produced interesting but not decision-relevant data. Tag questions by study type, topic area, and effectiveness. Over 10-20 studies, patterns emerge showing which question structures, probing approaches, and behavioral anchors work best for your specific domain and user segments. Store the library in your research repository so new team members inherit proven methodology rather than designing questions from scratch. With User Intuition’s 4M+ global panel across 50+ languages and 98% participant satisfaction rate, teams can confidently recruit for any segment and geography.