Product Growth
The Growth Podcast
How to Do AI-Powered Discovery (Step-by-Step with Live Demo) | Caitlin Sullivan
0:00
-1:12:40

How to Do AI-Powered Discovery (Step-by-Step with Live Demo) | Caitlin Sullivan

The exact workflow a UXR expert uses to analyze surveys and interviews plus how to build agents that do it for you.

Check out the conversation on Apple, Spotify and YouTube.

Brought to you by:

  1. Maven: Get 15% off Caitlin’s courses with code AAKASHxMAVEN

  2. Pendo: The #1 software experience management platform

  3. Jira Product Discovery: Plan with purpose, ship with confidence

  4. Kameleoon: AI experimentation platform

  5. Amplitude: The market-leader in product analytics


Today’s Episode

Discovery might be the most important core PM skill for building great products.

But most PMs are unprepared to do discovery in AI. PMs run surveys incorrectly, conduct interviews poorly, and end up with poor insights.

Today will give you the roadmap to avoid all those mistakes.

Caitlin Sullivan is a user research expert who runs courses teaching PMs how to do AI-powered discovery. And in today’s episode, she shows you exactly how she does it.

We’re talking live demos. Step-by-step workflows. Real survey data. Real interview transcripts.

This is a masterclass in discovery. The kind that moves the needle.

Apple Podcast

Spotify


Your Newsletter Subscriber Bonus

As a thank you for having me in your inbox, I’ve written up a complete guide to building your AI native PM operating system. It’s the ultimate composable stack playbook:

Today’s guide covers:

  1. Why AI Changes Discovery Completely

  2. The Step-by-Step Survey Analysis Workflow

  3. The Step-by-Step Interview Analysis Workflow

  4. How to Build Discovery Agents with Claude Code

  5. Advanced Techniques for Production-Grade Analysis


1. Why AI Changes Discovery Completely

Most PMs treat discovery as a checkbox.

Run some user interviews. Send out a survey. Check the box. Move on to building.

The problem is the analysis never actually happens.

You end up with:

  1. Survey responses sitting in a spreadsheet

  2. Interview recordings you never transcribe

  3. Notes scattered across documents

  4. Zero actionable insights

I’ve seen this pattern on every product team I’ve worked with.

The issue isn’t that PMs don’t care about discovery. It’s that traditional discovery takes too much time.

A proper interview analysis used to require:

  1. Transcribing the recording (30+ minutes per interview)

  2. Reading through the entire transcript

  3. Extracting key themes and quotes

  4. Organizing findings by topic

  5. Synthesizing across multiple interviews

For a standard discovery sprint with 10 interviews, that’s 10+ hours of analysis work.

Most PMs don’t have 10+ hours. So the analysis doesn’t happen.

AI Solves the Time Problem

With AI, the same analysis takes minutes instead of hours. (If you haven't set up your AI stack yet, start with my bundle.)

Here’s what the workflow looks like for analyzing a single interview:

  1. Upload transcript to Claude

  2. Run structured analysis prompt

  3. Get organized findings with timestamps

  4. Export to markdown file

Total time: 3-5 minutes per interview.

For 10 interviews, that’s 30-50 minutes instead of 10+ hours.

This changes the economics of discovery completely.

When analysis takes 10 hours, you only do it for the most critical projects. When analysis takes 30 minutes, you can do it for everything.

The Three Discovery Use Cases

AI-powered discovery works for three main scenarios:

Use Case 1 - Feature Validation

You’ve built a prototype. You need to validate it with users before investing engineering time.

  • Traditional approach: Run 5-10 user tests, take notes, report findings verbally to the team.

  • AI approach: Record every session, get transcripts analyzed automatically, share structured findings document with timestamps.

The difference is you actually have data to reference later. When someone asks “why did we build it this way?” you have timestamped quotes to point to.

Use Case 2 - Retention Analysis

Your product has a retention problem. You need to understand why users churn.

  • Traditional approach: Look at analytics, make guesses about why people leave.

  • AI approach: Interview churned users, analyze transcripts for patterns, identify specific friction points with evidence.

In the episode, we walked through this exact process using interviews with meditation app users to understand retention drivers.

The output included:

  1. Value anchors (what users care about most)

  2. Friction points (where the experience breaks down)

  3. Competitive gaps (what alternatives provide that you don’t)

  4. Risk segments (which user groups are likely to churn)

All with timestamped evidence from actual user conversations.

Use Case 3 - Market Research

You’re exploring a new product area. You need to understand user needs and competitive landscape.

  • Traditional approach: Read competitor websites, run a few interviews, make educated guesses.

  • AI approach: Systematic survey analysis, interview synthesis, competitive analysis with structured outputs.

This is where AI really shines. You can process hundreds of survey responses and dozens of interviews to find patterns that would be impossible to spot manually.

The key insight is that AI doesn’t replace the need for good discovery questions. It just removes the friction from analysis.

You still need to know your users. You still need to ask good questions. But now you can actually act on what you learn.

Continue Reading Online


2. The Step-by-Step Survey Analysis Workflow

Let’s get tactical.

Setup

You’ve just finished a survey. You have responses in a spreadsheet.

For the demo, Caitlin used a meditation app retention survey with these questions:

  1. How long have you been using the app?

  2. What made you start using it?

  3. What keeps you coming back?

  4. What would make you stop using it?

  5. Have you considered alternatives?

Standard retention research questions. Nothing fancy.

The key is how you analyze the responses.

Step 1 - Export Your Data

Get your survey data into a clean format.

Google Forms automatically exports to Sheets. Typeform exports to CSV. Whatever tool you use, get a clean spreadsheet with:

  1. One row per response

  2. One column per question

  3. Timestamps if you have them

For Caitlin’s workflow, she just copies this data directly. No need to sanitize or clean it first.

Step 2 - Open Claude Projects

Here’s the first key decision point.

You can analyze surveys in two ways:

Option A - Single conversation in regular Claude

Good for: One-off analysis, quick insights, learning the workflow

Downside: No memory between sessions, have to reupload data each time

Option B - Claude Projects with uploaded files

Good for: Ongoing research, multiple rounds of analysis, building institutional knowledge

Downside: Requires paid Claude Pro subscription

The better path is Claude Projects. The advantage is you can upload all your survey data and interview transcripts once, then reference them in multiple conversations.

Create a new project called something like “Meditation App Retention Study.”

Upload your survey responses spreadsheet.

Step 3 - Write the Analysis Prompt

This is where most people struggle.

Generic prompts like “analyze this survey” produce generic outputs. You need structure.

Here’s the exact prompt structure from the episode:

I have survey responses from [product] users about [topic].

Please analyze the data and provide:

1. Overall patterns and themes
2. Key insights by question
3. User segments (if patterns emerge)
4. Actionable recommendations

For each insight, include:
- The finding
- Supporting evidence (quotes)
- Implications for product decisions

The key elements:

  1. Context - Tell Claude what this is (product, topic, goal)

  2. Structure - Define exactly what outputs you want

  3. Evidence requirement - Ask for quotes to back up claims

  4. Actionability - Force conclusions that inform decisions

This structure ensures you get usable output instead of generic observations.

Step 4 - Review Initial Analysis

Claude processes the data and returns structured findings.

For the meditation app survey, it identified:

  1. 3 stable subscriber segments (daily meditators, therapy seekers, sleep users)

  2. 2 at-risk segments (price-conscious users, feature explorers)

  3. 4 key retention drivers (habit formation, therapy integration, sleep quality, guided content)

  4. 3 churn risk factors (price increases, feature removal, competitor features)

Each finding included supporting quotes from actual survey responses.

This initial analysis takes about 2-3 minutes to generate.

Step 5 - Dig Deeper with Follow-Up Prompts

The real power comes from iterating.

The follow-up prompt:

For the "therapy seekers" segment, can you:

1. Extract all relevant quotes
2. Identify their specific pain points
3. Compare what they get from our app vs traditional therapy
4. Recommend features that would increase their retention

This produces a focused deep-dive on one segment.

The beauty of Claude Projects is you can come back tomorrow and ask “What did we learn about therapy seekers?” and Claude remembers the analysis.

You’re building institutional knowledge that persists.

Step 6 - Export to Markdown

Once you have findings you want to share, export them.

The approach that works best: Ask Claude to format everything as a markdown document with:

  1. Executive summary at the top

  2. Detailed findings by section

  3. All supporting quotes included

  4. Recommendations clearly called out

Copy the markdown. Paste into your note-taking system. Share with your team.

Now you have a permanent record of the research, not just scattered conversation history.

Common Survey Analysis Mistakes

Mistake 1 - Not providing enough context

If you just upload a spreadsheet with no explanation, Claude has to guess what it’s looking at.

Always include: What product, what goal, what decisions this research informs.

Mistake 2 - Accepting the first output

The initial analysis is never complete. Iterate. Ask follow-ups. Dig into specific segments or questions.

Mistake 3 - Not exporting findings

If you leave everything in the chat, you’ll forget it. Export to markdown, save to your docs, make it permanent.

Mistake 4 - Analyzing surveys in isolation

Surveys tell you what people say. Interviews tell you what they mean. Usage data tells you what they do.

The best discovery processes combine all three sources.


3. The Step-by-Step Interview Analysis Workflow

Survey analysis is useful. But interview analysis is where AI really shines.

Interviews are rich, unstructured conversations. Extracting insights manually is time-consuming and inconsistent.

With AI, you can systematically analyze every interview the same way.

The Setup

You’ve conducted user interviews. You have transcripts.

For the demo, Caitlin analyzed meditation app retention interviews. The interviews explored:

  1. Why users started using the app

  2. What value they get from it

  3. What would cause them to stop using it

  4. How it compares to alternatives

Standard retention interview questions.

Step 1 - Get Clean Transcripts

Record your interviews (with permission). Use a transcription service.

Options:

  1. Otter.ai (automatic transcription during Zoom calls)

  2. Descript (upload recordings, get transcripts)

  3. Riverside.fm (records and transcribes automatically)

  4. Claude itself (upload audio files directly)

The transcript quality matters. Clean transcripts with speaker labels produce better analysis.

Format should be:

Interviewer: Why did you start using the meditation app?

User: I was going through a really stressful period at work...

Interviewer: What keeps you coming back?

User: Honestly, the therapy chat feature...

Step 2 - Define Your Analysis Framework

Before analyzing, decide what you’re looking for.

The framework that produces the best results is value anchors:

Value anchors are the core reasons users care about your product. The things that, if you removed them, users would leave.

For each interview, extract:

  1. Value anchors - What specifically provides value

  2. Supporting evidence - Exact quotes with timestamps

  3. Competitive context - How this compares to alternatives

  4. Retention risk - What might cause this user to churn

This framework ensures consistent analysis across all interviews.

You can use different frameworks depending on your goal. For feature validation, you might focus on usability issues and feature requests instead.

Step 3 - Create the Analysis Prompt

Here’s the exact interview analysis prompt:

I'm analyzing user interviews for retention research.

Please analyze this interview transcript and provide:

## Phase 1: Retention Assessment
- Retention segment (stable, moderate risk, high risk)
- Key retention drivers
- Churn risk factors

## Phase 2: Value Anchors
For each value anchor:
- Specific feature or capability they value
- Why it matters to them (quote with timestamp)
- How it compares to alternatives
- Risk if this value anchor were removed

## Phase 3: Recommendations
- What would increase this user's retention
- What product changes would reduce churn risk
- What competitive gaps we should address

Include exact quotes with timestamps for all findings.

This prompt produces a structured analysis document for every interview.

Step 4 - Analyze One Interview

Upload the first transcript. Run the analysis prompt.

Claude processes the transcript and returns:

Retention Assessment: “Stable subscriber - high engagement, multiple value anchors, low churn risk”

Value Anchors:

  1. Therapy chat feature (”This therapy chat feature directly addresses a gap that traditional therapy couldn’t fill” - 00:08:32)

  2. Daily habit formation (”I’ve built it into my morning routine” - 00:14:12)

  3. Sleep tracking integration (”The sleep quality improvements are measurable” - 00:21:45)

Recommendations:

  1. Protect therapy chat feature (core retention driver)

  2. Build on habit formation (push notifications, streaks)

  3. Expand sleep analytics (more detailed insights)

Total analysis time: 3-5 minutes.

Compare that to the 30-60 minutes it would take to manually review the transcript, extract quotes, and organize findings.

Step 5 - Analyze All Interviews

Repeat for every interview transcript.

In the demo, we analyzed 5 interviews. That’s 15-25 minutes total.

Each interview gets its own structured analysis document.

The consistency is the key benefit here. Every interview is analyzed using the exact same framework, making patterns easier to spot.

Step 6 - Synthesize Across Interviews

Once you’ve analyzed all individual interviews, synthesize across them.

Prompt:

I've analyzed 5 interviews. Please synthesize findings across all of them:

1. What value anchors appear in multiple interviews?
2. What retention patterns are consistent across segments?
3. What churn risks are most common?
4. What recommendations apply across user types?

Organize findings by frequency (how many users mentioned each theme).

This produces your executive summary.

The output shows you which findings are one-offs vs which are patterns that matter.

For the meditation app:

  1. Therapy chat - Mentioned by 4/5 users (critical feature)

  2. Habit formation - Mentioned by 5/5 users (universal need)

  3. Price sensitivity - Mentioned by 2/5 users (segment-specific)

Now you know where to focus.

Step 7 - Export Everything

Create a master document with:

  1. Executive summary (synthesized findings)

  2. Individual interview analyses (detailed breakdowns)

  3. Supporting evidence (quotes with timestamps)

  4. Recommendations (prioritized by frequency)

This becomes your research deliverable.

Share it with your team. Reference it in product strategy discussions. Use it to inform roadmap decisions.

The key is you have real evidence behind every decision. When someone questions why you’re prioritizing the therapy chat feature, you can point to 4 separate interviews with timestamped quotes.


4. How to Build Discovery Agents with Claude Code

Manual analysis works. But agents are better.

The difference between manual and agent-based:

  1. Manual approach - You run prompts, review outputs, iterate, export results

  2. Agent approach - You define the workflow once, the agent runs it automatically on all your data

For ongoing research programs, agents are worth the setup time.

When to Build Discovery Agents

Build agents when:

  1. You’re running regular user research (weekly, monthly)

  2. You need consistent analysis across many interviews

  3. Multiple people on your team do research

  4. You want to scale research without hiring more researchers

Don’t build agents when:

  1. This is a one-time research project

  2. You’re still figuring out your analysis framework

  3. You don’t have technical resources to maintain the agent

Start with manual workflows. Once you’ve done 10+ interviews the same way, automate it.

The Architecture

Discovery agents need three components:

Component 1 - Data Pipeline

The agent needs access to your research data.

For surveys: Google Sheets integration For interviews: Transcript files in a shared folder

Google Sheets works well for survey data. Local files work for interview transcripts.

Component 2 - Analysis Logic

This is where your prompts live.

The agent needs instructions for:

  1. How to process survey responses

  2. How to analyze interview transcripts

  3. What outputs to generate

  4. Where to save results

You’re essentially converting your manual workflow into code.

Component 3 - Output Generation

The agent needs to save results somewhere useful.

The agent outputs markdown files with:

  1. Executive summaries

  2. Detailed findings

  3. Supporting quotes

  4. Recommendations

These files automatically save to a shared folder where her team can access them.

Step 1 - Set Up the Project File

Create a file called instructions.md that defines your agent’s behavior.

Here’s what the instructions file looks like:

# Discovery Analysis Agent

## Purpose
Analyze user research data (surveys and interviews) for the meditation app retention study.

## Data Sources
- Survey responses: [Google Sheets link]
- Interview transcripts: /transcripts folder

## Analysis Framework

### For Surveys
1. Read all responses
2. Identify patterns and themes
3. Segment users by behavior
4. Extract actionable insights
5. Output to survey-analysis.md

### For Interviews
1. Process each transcript individually
2. Extract value anchors with timestamps
3. Assess retention risk
4. Provide recommendations
5. Output to interview-analysis-[name].md

## Output Format
All outputs should be markdown files with:
- Executive summary
- Detailed findings
- Supporting evidence
- Recommendations

This file becomes the agent’s operating manual.

Step 2 - Build the Survey Analysis Script

Claude Code can write the script for you.

Prompt:

Based on the instructions.md file, write a Python script that:

1. Connects to the Google Sheets survey data
2. Analyzes responses using the Claude API
3. Generates a structured markdown report
4. Saves the report to survey-analysis.md

Use the analysis prompt from my previous manual workflow.

Claude Code generates the script.

The script uses:

  1. Google Sheets API to read survey data

  2. Claude API to analyze responses

  3. File system operations to save results

You don’t need to write any of this yourself. Claude Code handles the implementation.

Step 3 - Build the Interview Analysis Script

Same process for interviews.

Prompt:

Write a Python script that:

1. Reads all transcript files from /transcripts folder
2. Analyzes each interview using the Claude API
3. Generates individual analysis reports
4. Saves reports to interview-analysis-[name].md

Use the value anchors framework from my previous manual workflow.

Claude Code generates another script.

This one loops through all transcript files and processes them automatically.

Step 4 - Test the Agent

Run both scripts on your existing data.

Check the outputs:

  1. Do the survey findings match your manual analysis?

  2. Are interview analyses using the correct framework?

  3. Are quotes properly extracted with timestamps?

  4. Are markdown files properly formatted?

Iterate on the prompts and scripts until outputs are consistent with manual analysis.

This testing phase is critical. You’re essentially training the agent to think like you.

Step 5 - Schedule Automatic Runs

Once the agent works reliably, schedule it.

Options:

  1. Run manually when new data arrives

  2. Schedule weekly runs (for ongoing research programs)

  3. Trigger runs when new files appear in the transcripts folder

The simplest approach: Run the agent whenever you complete a new batch of interviews.

Drop transcript files in the folder, run the script, get analysis reports in minutes.

The Results

In the demo, the agent analyzed:

  • 1 survey dataset with 20+ responses

  • 5 interview transcripts

Total runtime: Less than 2 minutes.
Total output: 6 markdown files with structured findings.

The analysis quality matched the manual workflow. But it happened automatically.

This is the power of AI agents for discovery. You define the workflow once, then scale it infinitely.


5. Advanced Techniques for Production-Grade Analysis

The basic workflows work. But here are the advanced techniques that separate good analysis from great analysis.

Technique 1 - Comparative Analysis

Don’t just analyze your own data. Compare it to competitive alternatives.

The advanced version of the survey prompt includes:

For users who mentioned considering alternatives:

1. What alternatives did they consider?
2. What features attracted them to those alternatives?
3. What kept them with our product instead?
4. What would cause them to switch?

Organize findings as a competitive feature matrix.

This produces a structured comparison of your product vs alternatives, based on what actual users say.

You can use this to inform product strategy and identify competitive gaps.

Technique 2 - Longitudinal Tracking

If you run regular research, track how findings change over time.

Create a master synthesis document that combines:

  1. This month’s survey findings

  2. Last month’s survey findings

  3. Trend analysis (what’s changing)

Prompt:

Compare the survey results from [current month] to [previous month]:

1. Which themes are strengthening?
2. Which themes are weakening?
3. Are new patterns emerging?
4. What retention risks have increased?

Highlight any significant changes.

This helps you spot trends before they become problems.

Technique 3 - Sentiment Analysis

Beyond extracting what users said, analyze how they said it.

Advanced prompt:

For each interview, assess:

1. Overall sentiment (positive, neutral, negative)
2. Emotional intensity (calm, frustrated, excited)
3. Confidence level (certain, uncertain, conflicted)

Use this to prioritize which feedback matters most.

A user who’s calmly explaining a minor annoyance is different from a user who’s frustrated about a major blocker.

Sentiment analysis helps you prioritize.

Technique 4 - Hypothesis Testing

If you have specific hypotheses, test them against the data.

Hypothesis: “Users who integrate the app into their morning routine have higher retention.”

Prompt:

Analyze all interviews for evidence supporting or refuting this hypothesis:

"Users who integrate the app into their morning routine have higher retention."

For each interview:
- Does the user have a morning routine?
- Is the app part of it?
- What's their retention segment?

Calculate the correlation.

Claude can process your research data to validate or invalidate specific hypotheses.

This is similar to how you’d use product analytics to test quantitative hypotheses. But you’re doing it with qualitative data.

Technique 5 - Risk Scoring

Build a simple scoring system to prioritize users by retention risk.

Scoring framework:

Assign each user a retention risk score (0-10) based on:

- Number of value anchors (more = lower risk)
- Price sensitivity mentioned (yes = higher risk)
- Alternative consideration (yes = higher risk)
- Feature dependency (fewer dependencies = higher risk)

Output a ranked list from highest to lowest risk.

This helps you identify which users to focus retention efforts on.

Technique 6 - Feature Prioritization Matrix

Combine research findings with usage data to prioritize features.

Prompt:

Create a feature prioritization matrix:

Columns:
- Feature name
- Number of users who mentioned it
- User segment requesting it
- Retention impact (high/medium/low)
- Competitive gap (yes/no)
- Recommended priority (P0/P1/P2)

Sort by recommended priority.

Now you have a data-backed roadmap prioritization framework.


Final Words

Bad discovery happens when you treat it as a phase. Good discovery happens when you treat it as a habit.

The best PMs continuously gather evidence, analyze patterns, and adjust course.

AI makes this possible.

Start with the manual workflows Caitlin showed. Analyze one survey. Process one interview. Get comfortable with the prompts.

Then build it into your regular rhythm. Weekly surveys. Monthly interviews. Continuous synthesis.

You’ll have superpowers soon enough.

Loading...

Where to Find Caitlin Sullivan


Related Content

Newsletters:

Podcasts:


PS. Please subscribe on YouTube and follow on Apple & Spotify. It helps!

Discussion about this episode

User's avatar

Ready for more?