AI-Moderated Interview at Scale

A figure at a laptop with multiple chat bubbles fanning outward, each containing a small silhouette

In Brief

AI-moderated interviews are structured qualitative interviews conducted by an AI agent with dozens or hundreds of participants at the same time. Participants respond asynchronously on their own schedule via text or voice, while the AI moderator follows a pre-designed interview guide, asks follow-up questions, and probes for detail. The output is qualitative data at a scale that would otherwise require a team of trained interviewers — recurring themes, language clusters, and segment differences across a large participant pool. This method excels at breadth; use human interviews when you need depth.

Common Use Case

You have already done a handful of customer interviews and found three possible customer segments. You want to know which segment is largest and whether each segment describes the problem differently. You set up an AI-moderated interview that runs asynchronously across all three segments, giving you broad pattern data in a week. Even 20–30 participants can surface useful patterns if your segments are clearly distinct — you don’t need 150 to learn something.

Helps Answer

  • What patterns show up across a large group of potential customers?
  • How do different types of customers describe the same problem?
  • What words do customers naturally use when talking about this topic?
  • Are there distinct sub-groups within our target market?
  • What themes come up repeatedly when many people answer the same questions?
Setting up an AI-moderated interview study typically takes 1 to 2 days: designing the interview guide, configuring the AI moderator, recruiting participants, and running a pilot test. The interviews themselves run for 3 to 7 days asynchronously, with each participant spending 15 to 30 minutes over one or more sessions.Analysis takes another 1 to 2 days, depending on the number of responses and whether you use AI-assisted analysis tools. With AI-assisted clustering and quote extraction, what used to be a multi-week thematic-coding pass on a hundred-plus transcripts collapses to a one-day workflow — the human time goes into reading a sample of raw transcripts to gut-check the model, not into manual coding.
Costs range from $500 to $2,000 for a typical study of 50 to 200 participants. This includes platform fees (typically $1 to $10 per completed interview) and participant incentives ($5 to $20 per person). Compared to hiring a team of human moderators to conduct the same number of interviews, AI moderation is substantially faster and lower-cost — you trade depth for breadth and parallelism, not money for money.You’ll need someone on the team who can design a solid interview guide — the AI moderator is only as good as its instructions. A researcher or founder experienced with customer discovery interviews should design and pilot the guide before scaling it up.

Description

AI-moderated interviews are structured qualitative interviews where an LLM-driven moderator runs your discussion guide against participants over text or voice, asynchronously, at a scale a human team can’t match. The mechanism is simple: you write the guide, configure the AI’s tone and probe behavior, and the model conducts each conversation in parallel — surfacing patterns, language clusters, and segment differences across the full transcript corpus instead of one interview at a time.

The method became viable in 2023–2024 as conversational LLMs became reliable enough to follow a discussion guide without going off-rails. The first peer-reviewed academic comparison — Wuttke et al. 2024 (arXiv:2410.01824) — found AI conversational interviewers produce data “comparable to traditional methods” with the scalability advantage you’d expect, though the published study was small (university students, political topics) and the authors flag generalizability constraints. Practitioner platforms (Strella, Conveo, Listen Labs, Outset, dscout, Maze) have since built productized workflows on top of the same core capability.

AI moderation gives you breadth, not depth. The AI moderator can’t read pauses, body language, or affect; it asks every participant the same question 200 times even when the question is poorly worded; and people who agree to be interviewed by an AI skew younger and more tech-comfortable. That’s why this method fits after you’ve done enough human discovery to know your segments and questions are roughly right — and why the canonical move is to use AI moderation to find the 5 to 10 most interesting participants, then talk to those people yourself.

How to

Prep

  1. Design the interview guide. Write 5 to 8 core questions that map to your learning goals. For each question, write 2 to 3 potential follow-up probes. The guide should flow like a natural conversation: start broad, get specific, close with open-ended reflection. Use the same principles you’d use for a human-moderated interview — open-ended questions, no leading, past behavior over future intent.

  2. Configure the AI moderator. Set up the AI’s instructions in your chosen platform. Key configuration decisions:

    • Tone and persona (professional researcher vs. friendly conversationalist)
    • How aggressively to probe (some participants clam up if pushed too hard)
    • When to move on vs. dig deeper
    • Maximum interview length
    • How to handle off-topic responses
    • Whether to allow voice, text, or both
  3. Pilot test with 5 to 10 participants. Run the full interview flow with a small group. Review every transcript. Look for:

    • Questions where the AI’s follow-ups fall flat
    • Points where participants seem confused
    • Moments where a human moderator would have caught something the AI missed
    • Whether the interview is too long or too short
  4. Revise the guide based on the pilot. This step is critical. Most first-draft AI interview guides need significant revision. Pay special attention to the follow-up logic — this is where AI moderation is weakest.

  5. Recruit participants. Use screener surveys to qualify participants who match your target persona. Recruit 50 to 200 participants depending on the diversity of your target market. For a single well-defined segment, 50 is often sufficient. For cross-segment comparison, aim for 30+ per segment.

Execution

  1. Launch the study. Send interview links to qualified participants. Monitor the first 10 to 20 responses as they come in to catch any issues with the guide or the AI’s behavior. Catching a broken question after 15 transcripts is cheap; catching it after 200 is not.

  2. Analyze at scale. Use AI-assisted analysis to:

    • Cluster responses into themes
    • Identify language patterns and frequently used phrases
    • Compare responses across demographic or behavioral segments
    • Flag outlier responses that deserve human attention
    • Generate a thematic summary with supporting quotes
  3. Deep-dive with human interviews. Select 5 to 10 participants whose responses were particularly interesting, surprising, or representative of key themes. Conduct follow-up human interviews to go deeper on the patterns you identified. The biggest single payoff from running 200 AI interviews is usually the shortlist of people you decide to talk to yourself.

Analysis

The output is a thematic analysis across a large sample, not a set of individual stories. Interpret accordingly.

Strong signals at this scale:

  • Theme saturation. When the same theme appears unprompted across 30%+ of responses, it’s likely a genuine pattern rather than an artifact of the question design.
  • Language convergence. If many participants independently use the same words or phrases to describe a problem, those are strong candidates for your marketing copy and positioning.
  • Segment divergence. Where two segments give systematically different responses to the same question, you’ve found a real segmentation boundary.

Be cautious about:

  • Emotional intensity. AI moderators are poor at gauging how strongly someone feels about something. A participant typing “yeah, that’s annoying” and one typing “that drives me crazy” might have the same underlying emotional intensity — or vastly different. You can’t tell from text alone, and the AI certainly can’t.
  • Social desirability at scale. Participants may perform for the AI the same way they would for a human interviewer. The async format may reduce this somewhat (no eye contact, no real-time social pressure), but it doesn’t eliminate it.
  • Depth of insight. A human moderator notices the pause before an answer, the shift in body language, the thing someone almost said but didn’t. AI misses all of this. The insights from AI-moderated interviews are real but shallow compared to what a great human interviewer can surface.
Biases & Tips
  • AI misses emotional cues The AI moderator cannot read tone of voice (even in voice interviews, current models miss nuance), facial expressions, or body language. It will treat a sarcastic response the same as a sincere one. It won’t notice when someone is holding back.

  • Participant effort bias Async interviews attract participants who are comfortable expressing themselves in writing (or speaking to a bot). People who think out loud, who need a conversation partner to develop their ideas, or who are less comfortable with technology will be underrepresented or give lower-quality responses.

  • Question design amplification In a human interview, a skilled moderator can recover from a poorly worded question by rephrasing in real time. An AI moderator will ask the same suboptimal question 200 times, amplifying the bias across your entire dataset.

  • Analysis automation bias When AI tools summarize 200 interviews into themes, there’s a tendency to trust the summary without reading individual transcripts. Always read a random sample of 15 to 20 raw transcripts to gut-check the automated analysis.

  • Recruitment self-selection People who agree to be interviewed by an AI agent are not representative of all customers. They skew younger, more tech-comfortable, and more open to novel experiences. Factor this into your interpretation.

  • “AI interviews give you breadth. Human interviews give you depth. You need both.” - @TriKro

  • “If you wouldn’t trust a junior research assistant to improvise follow-up questions, don’t trust the AI to do it without extensive piloting.” - @TriKro

  • “The most valuable thing that comes out of 200 AI interviews is the shortlist of 10 people you want to talk to yourself.” - @TriKro

  • Pilot before scaling: run 5 interviews manually (founder reads every transcript), revise the guide, run 20 more, revise again — only then scale to hundreds. An untested guide at 100+ participants amplifies the flaw across the entire dataset.

  • Structure the guide with open-ended questions in the first half and specific/targeted questions in the second. Themes that emerge unprompted are more reliable than themes prompted by direct questions.

  • When reporting findings, include the prompting context: “47% mentioned file sharing — but 80% of those were responding to a direct question about it” is meaningfully different from “47% spontaneously mentioned file sharing.”

Next Steps

  • Review a random sample of transcripts manually to verify AI moderation quality.
  • Cluster responses into themes and compare with any prior qualitative research.
  • Follow up with human-moderated Customer Discovery Interviews on the most nuanced or surprising themes.
  • Use findings to inform your next experiment or product iteration.
  • Use a Closed-Ended Survey to quantify the most common themes across your AI-moderated interviews.
  • Use a Synthetic Persona Screening to pre-test your discussion guide on simulated personas before running interviews at scale.
Learn more

Case Studies

Unilever (via Conveo)

Conveo, an AI-moderated video research platform, has cited Unilever as a customer using its tooling to study food preparation behavior at a scale and turnaround that would have been difficult with a human-moderated panel. Conveo’s CEO Dieter De Mesmaeker has said: “We can’t imagine anyone doing it the old way two years from now; 90% of the time in qualitative research is spent on activities that can be automated.”

Read more

Listen Labs

Won the Greenbook Insight Innovation Competition (IIEX) in 2024 for its AI-moderated interview platform. Customers cited on its site include Microsoft and Google.

Read more

Strella (“AI of Oz” origin story)

Co-founders Lydia Hylton and Priya Krishnan validated demand by manually faking the AI; Priya joined calls camera-off with a robotic voice and a five-second response delay to simulate model latency before raising capital and building the real platform. Bessemer Venture Partners’ profile notes the case as a worked example of Wizard-of-Oz-style demand validation in the AI-research category.

Read more

Strella customer use-case taxonomy

Strella analyzed its own customer base and found AI-moderated interviews concentrate in seven repeatable use cases: concept testing, customer purchase journey, ad/messaging testing, customer feedback, usability testing, competitive analysis, and general discovery. Useful as a vendor practitioner taxonomy for “where this method actually gets used.”

Read more

Further reading

Got something to add? Share with the community.