Research Methods Taxonomy

12 min read
A MECE catalogue of 47 market and UX research methods ranked by commonness of use and rated for in-house versus outsource execution. Calibrated for startups and scaleups building digital products, with seed stage as the scoring baseline.

Tier 1 — Typical

Near-universal among companies conducting any structured research. Tooling is commoditised; skills are widely taught.

Method Category Domain Outsource need (B2B · B2C) Requirements
Online surveys Surveys Both Low · Low Survey platform (Qualtrics, Typeform, SurveyMonkey) · Questionnaire design skills · Owned audience or panel access
In-depth interviews (IDI) Interviews Both High · Medium Interview guide · Recording and consent · Thematic analysis · Participant recruitment is the binding constraint — see scoring notes
NPS / CSAT tracking Surveys Both Low · Low Survey platform · CRM or product integration · Benchmark data for interpretation
A/B / multivariate testing Behavioral UX Low · Low Experimentation platform (Optimizely, VWO, or native) · Statistical literacy · Sufficient traffic for significance
Session recording / heatmaps Behavioral UX Low · Low Analytics tool (Hotjar, FullStory, Clarity) · GDPR compliance setup · Basic behavioural interpretation skills
Unmoderated usability testing Behavioral UX Low · Low Testing platform (UserTesting, Maze, Useberry) · Prototype or live product · Task scenario writing
Competitive intelligence Secondary Market Low · Low Structured analysis framework · Web tools (SimilarWeb, BuiltWith, LinkedIn)
Search behaviour analysis Secondary Both Low · Low Keyword tool (Ahrefs, Semrush, Google Keyword Planner) · Basic data interpretation

Tier 2 — Common

Used regularly by companies with a dedicated research function or meaningful product maturity. Require specialist facilitation, deliberate participant recruitment, or a methodological framework to learn.

Method Category Domain Outsource need (B2B · B2C) Requirements
Moderated usability testing Behavioral UX Medium · Low Trained facilitator · Screener and recruitment · Recording setup · Lab or controlled remote environment adds rigour
Focus groups Interviews Market High · Medium Skilled moderator · Viewing facility or video platform · Participant recruitment and incentives
Card sorting Behavioral UX Low · Low Card sort tool (Optimal Workshop, Maze) · IA expertise to interpret dendrogram output
Tree testing Behavioral UX Low · Low Tree testing tool (Optimal Workshop, Treejack) · IA expertise to build and analyse navigation structure
First-click testing Behavioral UX Low · Low Testing tool (Optimal Workshop Chalkmark, Maze) · Prototype or screenshot · Task writing
Expert / KOL interviews Interviews Both High · Medium Expert network (GLG, Atheneum, Respondent) or proprietary contacts · Screener · NDA and consent protocol
Concept testing Surveys Both Medium · Low Stimulus design (mocks, wireframes, concept boards) · Panel or recruited participants · Monadic vs comparative study design knowledge
Contextual inquiry Observation UX Medium · Low Trained facilitator · Access to participant's natural environment · Field notes or recording protocol
Kano model Modeling Both Low · Low Kano questionnaire design knowledge · Survey platform · Kano analysis tool or spreadsheet
Co-creation / design workshops Interviews Both Medium · Low Facilitation expertise · Stimulus materials · Virtual whiteboard (Miro, Mural) · Synthesis and theming skills
Diary / experience sampling Observation Both High · Medium Diary platform (dscout, Indeemo) · Longitudinal participant management · Thematic analysis over multiple time points · B2B participant attrition over diary duration is a significant risk
Cognitive interviews Interviews UX Low · Low Verbal probing protocol · Survey or instrument under test · Basic qualitative synthesis
Van Westendorp pricing sensitivity Surveys Market Medium · Medium Survey with four price-sensitivity questions · Panel (n ≥ 150) · Cumulative curve analysis — specialist interpretation recommended
Advertising / copy testing Surveys Market Medium · High Test stimuli (ads, copy variants) · Target-audience panel · Norms database for benchmarking — higher barrier in B2C where consumer norms expectations are more established
Social listening / sentiment analysis Secondary Both Low · Medium Boolean query design skills · Mid-market tools viable for B2B volume; enterprise platform (Brandwatch, Sprinklr) needed for B2C scale
Third-party panel data Secondary Market High · Medium Subscription access (GWI, Statista, YouGov) · Data literacy · B2B professional panels are scarcer and costlier than B2C consumer panels
Trend / horizon scanning Secondary Market Low · Low Signal-scanning tools (Exploding Topics, SparkToro) · Analytical framework (PESTLE, futures wheel)
Customer advisory boards (CABs) Panel Both Low · Medium Participant recruitment and vetting · Facilitation · Ongoing incentive and relationship management · B2B CABs often managed in-house with existing customers; B2C boards typically require external recruitment
JTBD / switch interviews (qualitative) Interviews Both Low · Low 8–12 semi-structured retrospective interviews · Switch interview protocol · Thematic synthesis · No statistical tools required — see scoring notes
Outcome-Driven Innovation (ODI) surveys Surveys Both Medium · Medium ODI survey design knowledge (Ulwick methodology) · Large sample (n ≥ 180) · Outcome statement writing · Opportunity algorithm scoring — see scoring notes

Tier 3 — Rare

Largely absent from startup and scaleup research programmes without specific investment. Require specialist hardware, proprietary platforms, clinical credentials, or panel infrastructure built over years. Sorted from least to most specialist.

Method Category Domain Outsource need (B2B · B2C) Requirements
Ethnographic / field research Observation Both High · High Trained ethnographers · Multi-site access permissions · Extended fieldwork (days to weeks) · Thick-description coding expertise
Mystery shopping Observation Market High · High Trained agent network · Briefing and scoring protocol · Data aggregation platform
Brand tracking studies Surveys Market High · High Panel infrastructure for continuous or wave sampling · Historical benchmark database · Longitudinal survey design
Max-diff / best-worst scaling Modeling Market Medium · Medium Experimental design software (Sawtooth) · Survey panel (n ≥ 200) · Statistical analysis skills
Perceptual / brand positioning mapping Modeling Market Medium · High Correspondence analysis or MDS expertise · Survey data on brand attributes · Statistical software (R, SPSS) · B2C requires larger consumer panel for reliable mapping
Market sizing (primary research) Modeling Market Medium · Medium Panel or custom sample · Volumetric estimation survey design · Projection and extrapolation methodology
Patent analysis Secondary Market Medium · Low Patent database access (Espacenet, Google Patents, Derwent Innovation) · IP literacy · Systematic classification methodology
Conjoint / discrete choice modelling Modeling Market High · High Sawtooth Software / Lighthouse Studio licence · Statistician with choice modelling expertise · Large sample (n ≥ 300) · Attribute and level stimulus design
Customer segmentation (cluster analysis) Modeling Market High · High R or Python · Multivariate statistics expertise · Large survey dataset · Panel access for validation
Online research communities (ORCs) Panel Both High · High Community platform (Recollective, Forsta, Qual360) · Ongoing moderation · Participant incentive structure
Consumer panels (quantitative) Panel Market High · High Panel infrastructure and sampling frames · Quota and weighting management · Longitudinal tracking system
Longitudinal / cohort studies Panel Both High · High Multi-wave research design · Participant tracking and attrition management · Extended timeline (months to years)
Eye tracking Physiological Both High · High Tobii / SR Research hardware (£15k–£40k+) · Calibration lab · Tobii Pro Lab or iMotions software · Gaze analysis expertise
Galvanic skin response (GSR) Physiological Both High · High Biofeedback hardware (Shimmer, iMotions biosensor) · Signal synchronisation with stimulus · Physiological data processing expertise
Facial coding / emotion AI Physiological Both High · High Proprietary platform licence (Affectiva, Noldus, iMotions) · FACS-trained analyst or AI model · Informed consent protocol
Implicit association testing (IAT) Physiological Market High · High Reaction-time testing platform (Millisecond Inquisit, iMotions) · Cognitive psychology expertise · Stimulus design for paired associations
EEG / brainwave measurement Physiological Market High · High Research-grade EEG system (BrainProducts actiCHamp/BrainAmp) or consumer-grade alternative (Emotiv — lower signal fidelity) · Neuroscientist or cognitive psychologist · Signal processing software · Controlled lab environment
Voice / tone biometric analysis Physiological Both High · High Acoustic analysis platform (Nemesysco, Vokaturi) · Linguist or trained voice analyst · High-quality audio capture
fMRI studies Physiological Market High · High MRI scanning facility · Radiographer · Neuroscientist · Ethics board approval · Budget of ~£500–£2,000 per participant session (scanner time typically £300–£600/hr; 1–2 hr sessions; higher figures in commercial packages including recruitment, analysis, and reporting)

Legend

Categories definitions

Interviews — Facilitated conversations (individual or group) designed to surface motivations, mental models, experiences, and attitudes. The researcher is the primary instrument. Includes qualitative depth interviews and structured expert elicitation.

Observation — Methods where the researcher watches participants behave in context, either in situ (ethnographic, field) or via a structured task (mystery shopping). Behaviour is observed rather than self-reported.

Surveys — Self-report data collected at scale via structured questionnaires. Captures attitudes, preferences, and stated behaviour. Includes general-purpose instruments and specialised pricing, satisfaction, and prioritisation methodologies.

Behavioral — Task-based or system-recorded data about what users actually do (clicks, navigation paths, task completion, error rates) rather than what they say they do. Largely digital; tooling is the primary enabler.

Physiological — Biometric and neurological measurement methods that capture involuntary physical responses (eye movement, skin conductance, brain activity, facial expression) as proxies for attention, emotion, and cognitive load. Require specialist hardware or certified platforms.

Modeling — Advanced statistical and quantitative methods that take survey or behavioural data as input and produce structured outputs: segmentations, preference weights, positioning maps, market size estimates. Require statistical expertise and often specialist software.

Secondary — Research using existing data sources rather than primary collection. Includes competitive analysis, search data, syndicated panel data, trend platforms, and patent databases. Speed and cost advantages; limited to what already exists.

Panel — Methods built around longitudinal or community-based participant groups, recruited, maintained, and studied over time. Generates tracking data and deep familiarity with a stable cohort. High infrastructure overhead.


Scoring notes and edge cases

In-depth interviews (IDI): why High (B2B) and not Low across the board

The craft elements of IDI — guide writing, facilitation, note-taking, thematic analysis — are learnable in-house, and the tools required are minimal. On that basis, Low might seem defensible. However, the binding constraint for most startups and early scaleups conducting IDI is not craft. It is participant recruitment.

In B2B contexts, reaching qualified participants involves small and tightly protected populations. Cold outreach response rates are low. Gatekeeping is common. Subjects command high incentive expectations and have unpredictable availability. Agency networks such as GLG, Atheneum, and Respondent provide access but at significant per-interview cost (often £200–£500 per completed session). Without a standing panel of engaged customers, recruiting a sample of 12 qualified B2B participants is a logistical project in itself.

In B2C contexts, the problem is substantially more tractable: Prolific, UserInterviews.com, and similar platforms provide screened consumer participants at reasonable cost, and social media recruitment supplements this well. Respondent also serves B2C audiences alongside its well-known B2B offering. Medium reflects the residual coordination effort required even there.

The principle this illustrates: outsource need should reflect the full execution burden — including recruitment — not only the craft of the method itself.


JTBD: two different methods sharing one name

The term "JTBD" in practice refers to two methodologically distinct approaches that share a theoretical ancestor but differ significantly in execution. This document lists them as separate entries.

JTBD / switch interviews (qualitative)
JTBD as a theoretical frame was conceptualised by Tony Ulwick around 1990 and introduced to Clayton Christensen in 1999; Christensen later popularised the term in The Innovator's Solution (2003). The switch interview method was developed by Bob Moesta (with Greg Engle and Chris Spiek) through applied practice. The method involves a small number of semi-structured retrospective interviews — typically 8–12 — focused on a specific switch: the moment a customer decided to hire (or fire) a product in favour of an alternative. The goal is to surface the causal mechanism behind a purchase or adoption decision — the push and pull forces, the anxieties, the social and emotional dimensions.

This approach requires no large sample, no statistical framework, and no specialist software. It is interpretive and qualitative. It is also often practised without any interviews at all — through desk-based job mapping or synthesis from prior research. Outsource need is Low · Low.

Outcome-Driven Innovation (ODI) surveys
Developed by Tony Ulwick, founder of Strategyn (est. 1991), first applied at Cordis Corporation in 1992 and named ODI around 1999. ODI treats the job executor's desired outcomes as measurable variables. Participants rate each outcome on two dimensions — importance and current satisfaction — and an opportunity algorithm scores underserved areas. This requires a large sample (n ≥ 180, per Strategyn's stated minimum; typically 180–3,000), carefully constructed outcome statements written in a prescribed format, and statistical analysis to produce segment-level opportunity scores.

This is the variant that warrants Medium outsource need and a statistical tooling requirement. It is considerably less common in practice than qualitative JTBD, and more closely resembles a structured quantitative survey programme.

When practitioners say "we do JTBD," they almost always mean the qualitative, interview-based approach — not ODI.


Methodology

Frequency tiering methodology

Methods are assigned to one of three tiers based on how commonly they appear in research programmes within the calibration range, drawing from practitioner surveys, UX and market research curricula, and industry benchmark reports (see references).

  • Tier 1 — Typical: Near-universal among companies that conduct any structured research. Low barrier to first use; tooling is commoditised and skills are widely taught. A company with one part-time researcher will likely use all Tier 1 methods within its first year.
  • Tier 2 — Common: Regularly used by companies with a dedicated research function or sufficient product maturity to demand more rigorous insight. Require specialist facilitation, deliberate participant recruitment, or a methodological framework that takes meaningful time to learn.
  • Tier 3 — Rare: Largely absent from startup and scaleup research programmes without specific investment. Methods at this tier require physical lab infrastructure, clinical credentials, proprietary software with high licence costs, or panel infrastructure built over years.

Frequency ratings are approximations, not empirical measurements. They describe prevalence within the calibration range. Some Tier 3 methods are common practice in industries such as FMCG, pharma, or financial services — the rating reflects digital product company norms, not the broader research industry.


Outsource need methodology

The outsource need score reflects the difficulty of executing a method entirely in-house within the calibration range. It is evaluated across four binding constraint types:

  1. Participant access — Can qualified participants be recruited without an agency or specialist panel? Difficulty increases where populations are hard to reach or where volume requirements exceed what organic outreach can supply.
  2. Specialist knowhow — Does execution require a trained specialist — ethnographer, statistician, neuro-researcher — whose profile takes years to develop internally?
  3. Hardware and software — Does the method require equipment with significant capital cost or platform licences inaccessible at standard SaaS pricing?
  4. Normative and benchmark data — Does accurate interpretation require proprietary databases of comparable results that only established research firms possess?

Low: None of the above constraints meaningfully apply. In-house execution is realistic with standard tools and moderate skill.
Medium: One or two constraints apply, or one applies significantly. Outsourcing buys meaningful gains in quality, scale, or speed, but in-house execution remains viable with effort.
High: Three or more constraints apply, or one is categorically inaccessible. In-house execution without specialist support would produce unreliable results.


B2B vs B2C outsource need

B2B and B2C contexts are scored separately because participant dynamics differ substantially.

B2B raises outsource need when: decision-maker populations are small and protected by gatekeepers; NDA requirements are common; subjects command high incentive expectations and have limited availability; and professional panel infrastructure (GLG, Atheneum, Respondent) carries significant per-interview costs.

B2C lowers outsource need when: large participant pools are accessible affordably (Prolific — primarily academic and behavioural research recruitment; UserInterviews; Hotjar Engage); social media recruitment is viable; and sample size targets are achievable without specialist agency support.

The outsource need column always shows values as B2B · B2C, including where both scores are the same.


Calibration detail

All scoring in this document is calibrated to the operational context of startups and scaleups building digital products. The document is product and industry vertical agnostic — methods and scores apply equally across digital product categories.

The calibration range spans a significant spread of internal capability. A pre-seed startup may have no budget, no CRM, and no customer base to draw participants from. A Series B scaleup may have a small research team, an established customer base, and a five-figure per-study budget. Scores reflect the constraints that apply across this range.

Note: This document uses funding stage as a practical proxy for operational maturity. The OECD/Eurostat definition (Oslo Manual, 2005) and Nesta's application of it define scaleups by growth rate — ≥20% annualised growth in employees or turnover over three years, with ≥10 employees at the start — not by funding round or headcount. The funding-stage framing used here is common in practitioner and ecosystem contexts but does not map precisely onto the OECD definition.

Note: Some US-centric sources (Crunchbase, Y Combinator) use "startup" loosely through Series C or beyond as a cultural designation. This document uses the European definition, where stage reflects operational maturity rather than growth ambition.

Comparison with mature enterprise:

Constraint Startup (pre-seed–seed) Scaleup (Series A–B) Mature enterprise (Series C+)
Research headcount Zero; research done by founders or PMs 0–2 dedicated researchers Dedicated UX research and/or market insights teams
Participant access None; outreach from scratch Small but growing customer base CRM databases of thousands; pre-negotiated supplier panels
Per-study budget Near zero to ~£5k ~£5k–£25k £25k–£200k+
Internal research history None or early accumulation (months of data) Limited — 1–2 years, patchy coverage Longitudinal record spanning multiple years and research cycles
Research repository None Ad hoc or early-stage Centralised; institutional memory
Tooling Freemium Mid-market Enterprise licences (Qualtrics, Sawtooth, iMotions)
Specialist knowhow Unavailable; generalist execution Rare hire or contractor Available internally or via standing supplier relationships
Cycle time Sprint-aligned; highly compressed Sprint-aligned with more planning runway Programmatic quarterly cycles

Note on internal research history: this refers to proprietary longitudinal data accumulated through repeated research cycles — past NPS waves, prior usability benchmarks, historical concept test results — that give interpretive context for new findings. It is distinct from industry or normative benchmarks, which are available externally regardless of company maturity.

Implication for scores: The baseline for all outsource need scores is a seed-stage startup. Scores were not averaged across the full startup–scaleup range — they reflect what a small, under-resourced team with no existing participant pool faces. Constraints ease progressively as companies move through Series A and beyond.


Data sources and references

Primary method references

Stage definition sources

  • ScaleUp Institute. (Annual). Annual ScaleUp Review. scaleupinstitute.org.uk — the primary ongoing annual report on UK scaleup activity, published since 2016. (Note: The Scale-Up Report on UK Economic Growth was a one-time independent report by Sherry Coutu CBE, 2014, commissioned by the UK government.)
  • OECD/Eurostat. (2005). Oslo Manual: Guidelines for Collecting and Interpreting Innovation Data (3rd ed.). OECD Publishing.
  • Startup Europe / European Commission. Startup and scaleup definitions used in EU innovation policy frameworks: ec.europa.eu/growth/sectors/digital-economy/start-ups

Frequency tiering sources

  • ESOMAR. (Annual). Global Market Research Report. esomar.org
  • Nielsen Norman Group. (2022). When to Use Which User-Experience Research Methods. nngroup.com
  • Dovetail. (2023). State of User Research Report. dovetail.com
  • User Interviews. (2023). UX Research in the Wild. userinterviews.com

JTBD methodological sources

  • Christensen, C.M., Hall, T., Dillon, K., & Duncan, D.S. (2016). Competing Against Luck: The Story of Innovation and Customer Choice. HarperBusiness.
  • Ulwick, A.W. (2005). What Customers Want: Using Outcome-Driven Innovation to Create Breakthrough Products and Services. McGraw-Hill.
  • Moesta, B., & Engle, G. (2020). Demand-Side Sales 101: Stop Selling and Help Your Customers Make Progress. Lioncrest Publishing.
  • Christensen Institute — JTBD overview: christenseninstitute.org/video/what-is-jobs-to-be-done-theory
  • Strategyn — ODI methodology: strategyn.com

General research practice

  • Goodman, E., Kuniavsky, M., & Moed, A. (2012). Observing the User Experience: A Practitioner's Guide to User Research. Morgan Kaufmann.
  • Portigal, S. (2013). Interviewing Users: How to Uncover Compelling Insights. Rosenfeld Media.
  • Krug, S. (2014). Don't Make Me Think, Revisited. New Riders.
  • Hall, E. (2013). Just Enough Research. A Book Apart.
  • Rohrer, C. (2022). When to Use Which User-Experience Research Methods. Nielsen Norman Group. nngroup.com

Biometric and physiological research

  • Duchowski, A.T. (2017). Eye Tracking Methodology: Theory and Practice (3rd ed.). Springer.
  • iMotions. (2020). Biometric Research Methods: A Guide. imotions.com

Industry bodies

  • ESOMAR — global research industry standards: esomar.org
  • Market Research Society (MRS) — UK standards and qualifications: mrs.org.uk
  • UXPA (User Experience Professionals Association): uxpa.org

47 methods across 8 categories. Calibrated for startups (pre-seed, seed) and scaleups (Series A, Series B) building digital products. Last updated March 2026.