Visual Intelligence

Study overview screen showing publish approval required before launching a research study

What's new

  • Introduces the ability for the AI moderator to see and interpret visual information during research

  • Adds three new intelligence types: Digital Intelligence (screen activity), Physical (participant-uploaded images or streamed video), and Emotional (real-time facial expressions or intonation)

  • Surfaces visual observations as timestamped, evidence-backed insights alongside transcripts

Why it matters Research has always been limited to what participants say — but what they do, show, and feel tells a different story. Visual Intelligence closes that gap by giving the AI moderator eyes: it can watch a usability session, analyze a photo of a participant's pantry, or detect confusion on someone's face — and use all of it to ask smarter follow-up questions and deliver richer insights. The result is faster time to insight, less manual video review, and findings that combine what people say with what they actually do and feel.

Used for

  • Running usability tests where the AI observes screen activity and probes friction points automatically

  • Collecting and analyzing participant-submitted photos or video streams mid-interview

  • Detecting emotional signals during video interviews and tying them to specific moments

How to use it

  • Add a screen-share task to your study to enable Digital Intelligence

  • Add a participant upload/stream prompt to collect and analyze images in context for Physical Intelligence

  • Enable Emotional Intelligence in your org settings to surface facial expression data on transcripts and insights pages

Read the full release >