Skip to content
research-methodsjtbdproduct-research

Four schools of Jobs to be Done: how a milkshake story became four research methods

In 1999, a consultant named Tony Ulwick walked into Clayton Christensen’s office at Harvard Business School and showed him a framework for measuring what customers want. Christensen was impressed. He included Ulwick’s examples in his 2003 book The Innovator’s Solution and gave the concept a catchy name: Jobs to be Done.

Around the same time, another consultant — Bob Moesta — was working with Christensen on a different angle. Moesta was less interested in what tasks people perform and more interested in why they make the decisions they make. He studied home buyers, mattress shoppers, and fast-food customers, reconstructing the emotional story behind each purchase.

By the mid-2010s, both Ulwick and Moesta were running successful consulting firms, both claiming to practice “JTBD,” and both meaning entirely different things by it. Then in 2016, Christensen himself published Competing Against Luck, articulating a distinct approach that differed from both Ulwick’s quantitative method and Moesta’s purchase-timeline interviews. Christensen’s version focused on discovering the job through observation and theory-driven conversation — probing not just what happened, but what progress people were trying to make and why non-consumers were choosing to do nothing at all. A fourth approach — Jim Kalbach’s practical canvas — appeared in 2020, bridging the gaps for teams that just wanted a usable tool.

Today, if you google “How to do JTBD research,” you will find advice from all four schools mixed together, often on the same page, without any indication that the methods are not interchangeable. A researcher who follows Moesta’s interview script while trying to fill Ulwick’s opportunity scorecard will produce confused data that answers neither question.

This article explains what each school actually does, where their philosophies diverge, and how to choose the right one.

The disagreement at the root

The four schools agree on one premise: people do not buy products for their features. They “hire” products to get something done in their lives. Beyond that, the agreement ends.

Christensen’s question: What job is this person hiring the product to do, and what are the functional, emotional, and social dimensions of that job? The answer reveals the job itself — the progress a person is trying to make in specific circumstances, including non-consumers who have not hired any solution.

Moesta’s question: Why did you switch? What was your life like before, and what pushed you to change? The answer reveals motivation — the emotional and situational context behind a purchase decision.

Ulwick’s question: What are you trying to accomplish, and how well does the current solution perform? The answer reveals unmet outcomes — measurable gaps in how well a task gets done.

Kalbach’s question: What is the job, and can the whole team agree on it? The answer produces alignment — a shared canvas that product, design, and engineering can reference.

Alan Klement, a JTBD theorist who worked with Christensen’s team, put the distinction between Moesta and Ulwick sharply. He calls Moesta’s approach Jobs-As-Progress and Ulwick’s approach Jobs-As-Activities, and argues they are not just different methods but incompatible models of why people buy. Christensen’s own approach sits closer to Jobs-As-Progress — he cares about the progress people seek — but adds a theory-driven layer that neither Moesta nor Ulwick emphasize: structured observation, non-consumption analysis, and the explicit decomposition of every job into functional, emotional, and social dimensions.

Moesta’s model says people do not want to mow the lawn. They want a home that looks cared for, and they will hire any solution that delivers that — including a lawn service, an automated mower, or artificial turf. The activity is incidental.

Ulwick’s model says people do want to mow the lawn, and the product should help them do it with fewer passes, less noise, and a more even cut. The activity is the unit of analysis.

Christensen’s model asks a different question: what are the circumstances that cause the job to arise? A suburban homeowner with a large lawn and a retiree with a small patio experience the same category (“lawn care”) but have different jobs. Christensen probes the circumstances and dimensions, not just the purchase decision or the task performance.

All four models have produced real business results. The question is not which is theoretically correct but which matches what you need to learn.

The mini periodic table of JTBD

We are building a periodic table of research methods — an interactive tool for choosing the right method for the right question. Four of the elements in that table are JTBD methods. Here they are, side by side.

Proactive (future)
Reactive (past)
Qualitative
Quantitative
Filter by question
Filter by role

Each element links to a full method guide with step-by-step instructions, AI prompts, and a checklist:

We take your business from where it is to where it needs to be
From project design to actionable, implementable insights.
Cloud Research Global →

How each method works in practice

Christensen Interview: discovering the job through theory and observation

Christensen’s approach begins not with an interview but with observation. The researcher watches people in context — in stores, at work, in their homes — looking for the circumstances that cause a “job” to arise. The milkshake study that launched the entire JTBD movement is the canonical example: Christensen’s team stood in a McDonald’s parking lot at 7 AM counting who bought milkshakes and asking them what “job” the milkshake was hired to do. The answer — a long, boring commute that needed one-handed food — would never have emerged from a focus group or a feature survey.

After observation, the researcher conducts interviews with three groups: active users who hired the product, lapsed users who hired and then fired it, and non-consumers who experience the same circumstances but have not hired any solution. This third group — non-consumers — is what distinguishes the Christensen approach most sharply from the others. Moesta reconstructs why someone switched; Christensen asks why someone did not switch at all.

Every job is decomposed into three dimensions: functional (what needs to happen), emotional (how the person wants to feel), and social (how they want to be perceived). This three-dimensional view often reveals that the emotional or social dimension dominates the hiring decision, even when the functional dimension is what the product team has been optimizing for. The retirement condo case from Competing Against Luck is the classic example: the functional job was “move to a smaller home,” but the emotional job — letting go of a lifetime of possessions — was the real barrier to purchase.

A typical study involves 10-15 interviews, includes an observation phase, takes 3-5 weeks, and produces job definitions with circumstance maps and a non-consumption analysis. The output is most valuable for innovation strategy, new market creation, and situations where the team does not yet know what job exists.

Switch Interview: the documentary about your customer’s decision

Moesta describes his interview technique as shooting a documentary. You sit down with someone who recently bought your product (or left it) and ask them to tell you the story of how they got there. Not “what features did you like” but “tell me about the day you decided to sign up — what was happening?”

The interviewer maps the story onto six stages of the buying timeline: first thought, passive looking, active looking, deciding, onboarding, ongoing use. At each stage, four forces are in play:

  • Push (F1): Frustration with the current situation. “Monday reporting was eating my entire morning.”
  • Pull (F2): The attraction of something new. “I saw a demo and imagined getting Mondays back.”
  • Anxiety (F3): Fear of change. “What if I migrate everything and it doesn’t work?”
  • Habit (F4): Comfort with the status quo. “My whole team knows the old spreadsheet.”

The switch happens when push + pull outweigh anxiety + habit. The most actionable finding is usually in F3 — a specific fear that the product team can address through design, messaging, or onboarding.

A typical study involves 8-12 interviews, takes 2-4 weeks, and costs relatively little beyond participant incentives. The output goes straight to marketing (positioning, messaging, ad targeting) and product (onboarding redesign, churn prevention).

ODI: the scorecard for unmet needs

Ulwick’s approach starts from the premise that every “job” (an activity customers want to perform) consists of dozens of desired outcomes — measurable criteria for how well the job gets done. “Minimize the time it takes to create a playlist in the correct order” is an outcome for the job “listen to music.”

The research happens in two phases. First, qualitative interviews with 15-20 people who perform the job. The researcher extracts 50-150 desired outcome statements, each following a strict format: direction of improvement + metric + object of control. These are not features — they are performance criteria customers use to judge any solution.

Second, a quantitative survey asks 50-200+ respondents to rate each outcome on importance and current satisfaction. The Opportunity Score (importance + the gap between importance and satisfaction) reveals which outcomes are underserved. Statistical segmentation finds groups of customers who share similar unmet needs — segments defined not by demographics but by what they struggle with.

A typical study takes 4-8 weeks, requires statistical expertise, and costs significantly more than a Switch Interview. The output goes to product strategy: which outcomes to target, which segments to serve, which innovation strategy to pursue (differentiated, dominant, disruptive, or discrete).

Canvas Workshop: the alignment tool

Kalbach’s approach is less about discovery and more about structured decision-making. A cross-functional team sits down in a workshop, defines the domain (where do we want to innovate?), the job performer (who are we innovating for?), and the main job (what are they trying to accomplish?).

After the workshop, 5-10 investigation interviews with real job performers validate and enrich the canvas. The team fills in job steps, desired outcomes, emotional and social aspects, and job differentiators. A second workshop prioritizes these elements. An optional survey validates the prioritization at scale.

The canvas itself becomes the shared reference document — a single artifact that product, design, engineering, and marketing can all point to when making decisions. GitLab published their entire JTBD process as an open-source playbook, making the Canvas approach accessible to any team.

A typical study takes 3-6 weeks and works best when the goal is team alignment rather than deep discovery or market sizing.

Which role benefits from which school

Different teams extract different value from different schools. This is not about theory — it reflects how the outputs of each method map onto daily work.

Marketing gravitates toward the Switch Interview. The four forces framework produces the exact language customers use when describing their situation — that language becomes ad copy, landing page headlines, and email sequences. Trigger events from the interviews become targeting criteria: reach people at the moment they feel the push. Moesta named his book Demand-Side Sales for a reason — the method was built for understanding how demand forms.

Sales also relies on the Switch Interview, but for different outputs. The anxiety force (F3) maps directly onto buyer objections. When a salesperson knows that the most common anxiety is “What if my team doesn’t adopt it?”, they can address it before the prospect raises it. The competitive map — built from what buyers actually considered, not what the company assumes — becomes battle cards.

Product management leans toward ODI. Opportunity scores give product managers a quantitative answer to “what should we build next?” that does not depend on the loudest stakeholder or the most recent customer complaint. Outcome-based segments tell engineering who they are building for and what “better” means in measurable terms.

Strategy and C-level use both Christensen Interviews and ODI, depending on the question. The Christensen approach answers “should we enter this market — is there a job here?” while ODI answers “which unmet needs should we target within a market we already understand.” The Christensen Interview is the tool for portfolio-level questions; ODI is the tool for product-level prioritization.

Innovation and R&D gravitate toward the Christensen Interview. The method was designed for new market creation and disruptive innovation — situations where the existing product category may not be the right frame. Non-consumption analysis reveals markets that do not yet exist, which is exactly the question R&D teams need answered before committing resources.

UX researchers often start with the Canvas Workshop to align the team on the job, then run Christensen or Switch Interviews for deep discovery. The Christensen Interview works when the team does not yet know what the job is; the Switch Interview works when they know the job but need to understand purchase dynamics. ODI enters when quantitative validation is needed. The Canvas serves as the shared artifact that product, design, and engineering reference throughout the project.

Customer success benefits from Switch Interviews conducted with churners — people who “fired” the product. The four forces reveal what pushed them out and what could have retained them, giving CS teams specific retention levers rather than generic “check in more often” advice.

Founders and startups almost always start with Switch Interviews. The method is cheap (8-12 interviews, no survey infrastructure), fast (2-4 weeks), and produces immediate GTM insights. ODI requires resources most startups do not have. The Canvas is useful once the team grows beyond 3-4 people.

RolePrimary schoolWhat they get from it
MarketingSwitch InterviewCustomer language for copy, trigger events for targeting
SalesSwitch InterviewObjection handling (F3 anxiety), competitive battle cards
Product managementODIFeature prioritization by opportunity scores
Strategy / C-levelChristensen + ODIJob discovery for new markets + outcome-based segments for existing ones
Innovation / R&DChristensen InterviewNon-consumption analysis, new market opportunities, job definitions for products that do not yet exist
UX researcherCanvas + Christensen/SwitchTeam alignment + deep discovery (Christensen for job finding, Switch for purchase dynamics)
Customer successSwitch (with churners)Retention levers from the four forces
Founder / startupChristensen + SwitchChristensen discovers the job; Switch validates demand with 8-12 interviews

Decision guide: which school for which question

Your questionMethodWhy
What job exists here — is there a market?Christensen InterviewDiscovers the job through observation and non-consumption analysis before any solution exists
Why do people buy (or not buy) our product?Switch InterviewReveals purchase motivation, not task performance
Which features should we build? Where is the biggest unmet need?ODIMaps the full outcome space and quantifies the gaps
How do we align the team on what we are building for?Canvas WorkshopCreates a shared artifact for cross-functional teams
Is there demand in a new market?Christensen first, then SwitchChristensen discovers the job and non-consumers; Switch validates purchase dynamics
Why is churn so high?Switch Interview (with churners)Four forces reveal what drives people out and what could retain them
How do we position against competitors?Switch InterviewThe competitive map comes from what alternatives buyers actually considered
We need to segment our market by needs, not demographicsODIOutcome-based segmentation finds hidden segments no survey can
What emotional and social barriers block adoption?Christensen InterviewThe three-dimension decomposition reveals barriers the functional analysis misses

How AI changes JTBD research — and where it does not

All four JTBD methods involve labor-intensive steps that AI can accelerate, but each method has a different bottleneck and a different human-dependent core. The table below maps what changes and what stays the same.

StepChristensen InterviewSwitch InterviewODICanvas Workshop
PreparationAI can draft a job hypothesis and identify likely circumstances from market data. The researcher still defines the observation plan and recruits across all three groups (users, lapsed, non-consumers).AI can draft a timeline framework and suggest probing questions for the six stages. The researcher still defines the switching decision to study and recruits participants with recent memory.AI can generate a hypothesis job map with 15-20 steps and candidate outcome statements. The researcher still defines the job-to-be-done — the single decision that shapes everything downstream.AI can populate an initial canvas with job steps, likely emotions, and circumstances in under an hour. The team still needs to agree on the job performer and main job through discussion, not delegation.
InterviewsAI cannot replace the observation phase or conduct the discovery interview. The method depends on following the participant’s story to find the job — the researcher must notice what the participant does not say and probe the circumstances that triggered the hiring decision.AI cannot conduct a Switch Interview. The method depends on real-time rapport, emotional probing, and the interviewer’s ability to follow the participant’s story wherever it leads. Synthetic respondents miss the surprising details that make JTBD research valuable.AI cannot replace outcome discovery interviews. The researcher must push past feature requests to extract measurable, solution-independent outcome statements — a skill that requires live judgment.Kalbach is explicit: AI is “incomplete and lacks real-world context.” At least 6-8 real interviews are needed. AI can supplement, not substitute.
AnalysisAI decomposes transcripts into functional, emotional, and social dimensions and maps circumstance patterns across interviews. The researcher validates the three-dimension split and identifies which dimension dominates — a judgment call AI consistently gets wrong.AI reduces transcript analysis from 2-3 hours per interview to 20-30 minutes. An LLM can extract the four forces, map the purchasing timeline, and identify trigger events. The researcher reviews, corrects, and spots patterns across interviews.AI cuts the consolidation of 300+ raw outcomes from days to hours. An LLM clusters duplicates, standardizes wording, and flags outcomes that violate ODI formatting rules. The researcher validates every statement.AI extracts JTBD elements (steps, criteria, emotions, circumstances) from transcripts and identifies which appear most frequently. Cross-interview pattern detection drops from a full day to minutes.
SynthesisAI can draft job definitions and circumstance maps. It cannot assess whether a job is worth pursuing, identify the non-consumption opportunity, or decide between sustaining and disruptive innovation strategy — those require business judgment Christensen explicitly frames as human work.AI can draft force diagrams and job stories. It cannot decide which stories matter most for the business or how to position the product — that requires competitive awareness and strategic judgment.AI can calculate opportunity scores, run cluster analysis, generate segment profiles, and draft the opportunity landscape. It cannot decide which segments to target or what innovation strategy to pursue.AI can draft job stories and HMW statements. It cannot facilitate the team workshop where these artifacts get debated, refined, and turned into commitments.

The pattern across all four methods: AI compresses the mechanical middle — transcription, extraction, scoring, pattern matching — but leaves the strategic bookends intact. The researcher still frames the question at the start and makes the decisions at the end. Skipping either bookend produces fast but unreliable results.

Timeline impact: teams with AI support report roughly 40-50% reduction in total project time, concentrated in the analysis phase. A Christensen Interview project drops from 3-5 weeks to 2-3 weeks. A Switch Interview project drops from 3-4 weeks to 2 weeks. An ODI study drops from 16-24 weeks to 8-12. A Canvas Workshop drops from 3-4 weeks to 1.5-2 weeks. The interviews themselves take the same time — AI does not make conversations shorter.

The risk to watch for: AI-generated JTBD artifacts (force diagrams, outcome statements, job stories) look plausible even when they are wrong. An LLM will produce a clean force diagram from a transcript where the interviewer never actually probed the anxiety force — it will infer what the anxiety “probably” was. The researcher must treat AI output as a first draft to verify, not a finished product to deliver.

What does not work

Mixing methods in the same study. Running ODI-style outcome interviews and Moesta-style switching interviews with the same participants, in the same session, confuses them and contaminates the data. Pick one method per study.

Applying Switch Interview findings to feature prioritization. Job stories tell you what people want to become, not which feature to build next. If you need feature-level prioritization, you need ODI’s opportunity scores.

Using ODI to fix churn. Opportunity scores tell you where the task is underserved, but they will not tell you what emotional barrier stopped someone from switching or what habit kept them with a competitor. For that, you need the four forces.

Using Christensen Interviews to prioritize a feature backlog. The Christensen Interview discovers the job and its dimensions — it does not produce a ranked list of features. For feature prioritization, you need Ulwick’s opportunity scores. The Christensen output feeds into innovation strategy decisions (should we enter this market?), not sprint planning.

Skipping the method and just “doing JTBD.” The phrase “we do JTBD” without specifying which school is like saying “we do statistics” without specifying whether you mean descriptive, inferential, or Bayesian. The tools are different, and using the wrong one gives you data that does not answer your question. With four schools in the field, the risk of mixing approaches has only grown.

The books worth reading

If you want to go deeper into any school:

For the Christensen school, read Competing Against Luck (2016) — it is the definitive articulation of Christensen’s own approach to Jobs Theory, with the milkshake study, the retirement condo case, and the distinction between Big Hire and Little Hire. For the earlier theoretical foundations, read the relevant chapters in The Innovator’s Solution (2003).

For the Switch Interview school, start with Moesta’s Demand-Side Sales 101 — short, practical, full of interview examples. For a free deep dive into the theoretical distinction between Jobs-As-Progress and Jobs-As-Activities, download Klement’s When Coffee and Kale Compete.

For ODI, read Ulwick’s Jobs to Be Done: Theory to Practice. It is the most complete description of the ODI process, with case studies from Strategyn’s consulting work.

For the Canvas approach, read Kalbach’s The Jobs to Be Done Playbook. It is hands-on, with templates and workshop facilitation guides. Also check the GitLab JTBD Playbook, which is open-source and battle-tested.