UX Evals

Announcement graphic introducing UX evals powered by AI-moderated research

Evaluate AI-generated experiences with structured user research

What’s new

  • Run studies focused on evaluating AI or LLM-powered experiences

  • Capture user reactions to generated outputs

  • Analyze how well AI responses meet user expectations

Why it matters
As more products rely on AI-generated outputs, understanding how users perceive and interact with those experiences becomes critical. UX Evals help teams test and improve AI-driven features with real user feedback.

How researchers are using it

  • Testing chatbot responses and identifying where users lose trust

  • Comparing multiple prompt variations to see which performs better

  • Evaluating end-to-end AI workflows before shipping

How to use it

  • Define the AI experience or output you want to test

  • Run interviews or studies with target users

  • Analyze feedback to refine the experience

Read the full announcement