Testing Assumptions &
Building Trust

Testing Assumptions &
Building Trust

How comprehensive user research revealed what engineers really needed from AI and changed our product roadmap

Due to confidentiality, respect of privacy policies, and enterprise compliance, some visuals are recreated or simplified for portfolio purposes. A detailed version is available upon request.

CONTEXT

Role & Impact

UX Research Lead | 3 months, Aug 2024 - Oct 2024 | Houston, TX

Challenge: After the initial launch of our AI assistant, the product was being designed in the dark, building features based on assumptions rather than evidence.

Outcome: Led the platform's first structured user research, uncovering critical trust gaps and workflow misalignments that directly shaped our next several design sprints and influenced product strategy.

THE PROBLEM

We Were Building Features Without Understanding How Engineers Actually Used Them

Our AI assistant had 1,600+ active users, but we lacked systematic understanding of how they used it, what frustrated them, or where opportunities existed. In safety-critical environments, unused features represent missed opportunities to prevent incidents.

The Stakes:

  • Building features based on assumptions, not evidence

  • No visibility into trust barriers or workflow gaps

  • Roadmap decisions lacked user validation

  • About to invest in new capabilities without validating existing ones

RESEARCH APPROACH

I Designed a Mixed-Methods Approach to Uncover Both Usability Issues and Strategic Opportunities

Objectives:

  • Identify usability friction across 17 core features

  • Understand workflow integration patterns

  • Explore unmet needs and expansion opportunities

Participants:

  • 11 engineers across disciplines, experience levels, and time zones

  • Early career to 20+ year SME veterans

  • Global representation (Houston, AGT, Sunbury)

Methodology:

  • Heuristic evaluation using Nielsen's principles

  • 11 moderated remote usability tests (~45 min each)

  • Contextual inquiry observing real-world usage

🧠

🧭

🧪

Heuristic Evaluation

Hypothesis

Insights & Design Recommendations

Conclusions

User Testing: Usability Tests + Contextual Inquiry

Experimentation

Thematic Analysis

Analysis

Heuristic Evaluation Chart

Heuristic Evaluation Chart

ANALYSIS

Turning Raw Interview Data Into Actionable Product Strategy

After 11 interviews, I created systematic frameworks to spot patterns and prioritize insights:

Rose, Bud, Thorn Summary - What's working, opportunities, and pain points

User Profile Matrix - Behavioral patterns by role, region, and experience

Prioritized Issues List - Recommendations mapped by impact and effort

Thematic Clustering - Grouped into: discoverability, trust, workflow integration, expansion

Rose, bud, thorn chart

Rose, bud, thorn chart

User profile matrix

User profile matrix

Organized notes

Organized notes

  1. If you wanted to rename a chat, how would you do that?

Participant Name

Body of notes

Participant Name

Body of notes

Participant Name

Body of notes

Participant Name

Body of notes

Participant Name

Body of notes

Participant Name

Body of notes

Participant Name

Body of notes

Sample question setup

Sample question setup

Summary of section

Summary of section

Participant Name

Participant Name

Notes taken by notetaker

Notes taken by notetaker

Participant Name

Participant Name

Notes taken by me

Notes taken by me

Sample notes setup

Sample notes setup

KEY INSIGHTS

Engineers Could Use Features, They Just Couldn't Find Them

6 Insights:

No Clear Chat Organization Strategy

Accuracy > Speed

Feature Discoverability > Comprehension

Help & Settings Misaligned with Expectations

Blue Text Was Domain Convention, Not Confusion

Engineers Want Document Interaction

Users struggled to find features but easily used them once informed. The problem was visibility, not complexity.

The Data:

  • Only 27% could find "Rename Chat"

  • Just 18% knew where "Dark Mode" was

  • 36% discovered "Copy Chat" organically

Design Impact:

  • First-time user walkthrough

  • Changelog banners for new features

  • Moved dark mode to Settings with default

Feature Discoverability %

No Clear Chat Organization Strategy

Accuracy > Speed

Feature Discoverability > Comprehension

Help & Settings Misaligned with Expectations

Blue Text Was Domain Convention, Not Confusion

Engineers Want Document Interaction

Users struggled to find features but easily used them once informed. The problem was visibility, not complexity.

The Data:

  • Only 27% could find "Rename Chat"

  • Just 18% knew where "Dark Mode" was

  • 36% discovered "Copy Chat" organically

Design Impact:

  • First-time user walkthrough

  • Changelog banners for new features

  • Moved dark mode to Settings with default

Feature Discoverability %

BEHAVIORAL PATTERNS

Observing Real Workflows Revealed Deeper Behavioral Patterns

Primary Use: Verification, not ideation - Engineers were using the assistant to confirm known facts rather than exploring abstract ideas.

Trust Concerns: Users were worried AI might lag behind process updates and document versions. They wanted to ensure data was up to date.

Workflow Complexity: Multi-step workflows needed simplification to reduce cognitive load.

Scale Desires: Users expected bulk actions, configurable interfaces, and personalization.

Onboarding Gap: Users consistently requested walkthroughs and contextual tips to make the product easier to adopt, especially for first-time users.

IMPACT

Research Identified Clear Wins and Critical Gaps That Shaped Our Roadmap

What Worked:

  • Significantly faster than legacy workflows

  • High satisfaction when features were discoverable

  • Strong excitement about document upload

Critical Gaps:

  • Feature discovery is the biggest adoption barrier

  • Trust and accuracy need work

  • Workflows under-supported for deeper research

Voice of the User:

"This is like a million times faster than finding the source myself."

"Creating the RACI charts—it did the work for us. I didn’t have to slog through documents."

"Saved me a day's worth of work—I did it in an hour."

WHY THIS MATTERED

The Most Valuable Insights Came from Challenging Our Assumptions

The Blue Text Revelation: I flagged blue text as a problem in my heuristic evaluation based on UX principles and accessibility standards. But user interviews revealed it was actually valuable because it mirrored internal documentation norms.

  • Learning: Users need clarity and control, not just uniformity.

Speed vs. Accuracy: The team assumed users prioritized speed. 82% preferred accuracy even if slower. This redirected our technical priorities from performance optimization to reliability.

  • The Real Value: Shifted the team from designing in isolation to building based on real-time user needs.

TRANSFORMING INTO DESIGN

This project is apart of a larger system of work

This research directly informed the design strategy detailed in my AI Assistant for Safety-Critical Engineering case study. The insights discovered here regarding trust, transparency, and workflow integration, directly informed the product's foundational design principles and feature roadmap.

See how I translated these research findings into shipped features like Workspaces, Agentic Tools, and transparency-first interfaces in the companion design case study.