User Research Case Study
The Challenge
MedLaunch Concepts is a healthcare technology startup developing the DNV Healthcare Portal, a platform designed to replace outdated accreditation systems that have long frustrated hospital surveyors and compliance officers.
The existing system was plagued by abandoned forms, manual workarounds, and workflows that felt like obstacles. Users frequently gave up mid-task, resorting to pen and paper.
Our mission: evaluate the new portal's usability and identify friction points preventing users from completing critical work that impacts patient safety.
8
Test Sessions
4
Core Tasks
60 min
Avg. Session
3
Key Findings
"Like there's nothing that's kind of like leading me as to how to use the tool or what to do first."— Participant 1, during survey creation task
My Role
Amen Hailu
UX Researcher, Client Liaison, Team Coordinator
Arsh Kaushik
UX Researcher
Mars Nevada
UX Researcher
Shlok Belgamwar
UX Researcher
Our Approach
Phase 01
We initiated contact with Robert Dostert, MedLaunch's project manager, to understand the product landscape. The platform serves two user groups: hospital-side users (quality directors, compliance officers) and DNV-side users (accreditation surveyors).
Key insight: users were "not technical, used to pen and paper." This informed our entire approach.
Strategic Decision
We focused on the Surveyor Tool, the newest and most complex and part of the platform.
Phase 02
I wrote the research proposal outlining our methodology, objectives, and success criteria. Then I designed a screening survey targeting healthcare professionals with accreditation experience.
We recruited 8 participants, including two ideal candidates, P3 (Clinical Psychology PhD) and P4 (Healthcare Attorney/former RN), whom I personally interviewed.
Methodological Consideration
While proxy participants couldn't fully represent end users, their experience with regulated workflows provided valuable insights.
Phase 03
I authored a comprehensive testing script with a realistic scenario: participants would set up and conduct an internal audit for the emergency department at a fictitious hospital.
We employed concurrent think-aloud (CTA) and concurrent probing (CP) techniques across 60-75 minute remote sessions via Zoom.
Phase 04
After sessions, we conducted affinity mapping to cluster observations. Patterns emerged: terminology confusion, hidden call-to-actions, and inflexible workflows appeared across nearly every session.
Scope Decision
We focused on pages requiring major reconstruction: Survey Page, Agenda, and Submissions/Entries Review.
Our synthesis organized our many observations into thematic clusters across seven key areas.
P1
Project Manager
20+ years in NYC advertising with QC testing expertise
P2
Project Manager
Health & wellness advertising experience
P3
Clinical Psychology PhD
4-7 years healthcare research at NYU Langone
Ideal CandidateP4
Healthcare Attorney
Former RN, 8+ years healthcare regulatory
Ideal CandidateP5
Project Manager
Software development at healthcare startup
P6
Former Project Manager
Game design, workflow analysis expertise
P7
Product Designer
Healthcare compliance startup freelancer
P8
Senior Product Designer
Enterprise tools at Uber, healthcare contracts
What We Found
59.4
Grade D — Below Industry Average
Product Usability Scale Score
Task 1
75%
Survey Creation
Task 2
25%
Agenda Management
Task 3
62.5%
Control Center Navigation
Task 4
87.5%
Entry Form Submission
A Note on Design Recommendations
After synthesizing the insights, my team members did a great job mocking up the design recommendations. I made minor edits to elevate the designs, which I highlight in blue throughout the mockups below.
Users were unclear how to get started creating a survey and got easily lost during the multi-step process and navigating the different status types.
Problem 1A: Scheduling vs Creating Confusion
Users were unsure how to create a survey, despite the presence of a "Schedule Survey" button. "Scheduling" created confusion as it's generally understood as managing deployment time rather than creation. The button was inside two layers of boxes, making it less prominent. It's also the same color as the Historical Data table header.
The "Schedule Survey" button suffers from terminology, placement, and color hierarchy issues
Insight
Users desired a more explicit survey creation guided experience.
Problem 1B: Form Entry Frustration
Users relied on learned behaviors from other tools, resulting in frustration when designs didn't conform to expectations.
The tool isn't connected to a database like Microsoft Teams or Pulse, which allows pre-loaded teammate selection.
Form fields lack familiar patterns: requirement clarity, date pickers, input labeling, and user management
Insight
Users desired a form entry experience that aligns with common design standards across other tools.
Problem 1C: Status Confusion
Users were confused about status terminology and what options meant. They expected status changes that didn't happen (e.g., updating a survey resulted in a draft, not an updated live survey).
Multiple workflow and terminology issues create confusion about survey states and actions
Insight
Users want more clarity about status types and when they change, as well as flexibility in editing surveys after confirmation.
Redesigned main page with "Get Started" guidance, relocated CTA, filter system, and clearer status chips
Stepped form with unique name assistance, dynamic checks, industry-standard date pickers, and user invitation flow
Obscured timeline controls and inconsistent role labeling hide critical context, increasing the risk of coordination errors during the survey.
The agenda interface suffers from hidden time context, rigid column assignments, missing priority indicators, and no inline editing
Problem 2A: Missing Temporal Context
Surveyors are managing a complex 3-5 day narrative where finding specific survey milestones currently requires memorization rather than visual recognition. The current single-day constraint forces users to manually click through days to find high-priority items, increasing cognitive load and the risk of missing a critical "Hard Stop." 3 out of 8 users specifically compared this limitation to standard tools (like Google Calendar) that allow "zooming out" to assess the week's workload at a glance.
Insight
Surveyors must balance "High Risk" systems against routine checks. Without a Weekly/Monthly view or "Priority Star," they cannot visually strategize their time across the full survey duration.
Problem 2B: Rigid Column Assignments
The system incorrectly assumes a static, one-to-one relationship between a surveyor and a track (e.g., "Lisa" is always the Generalist). In reality, audit teams act as a dynamic mesh where responsibilities shift hour by hour. If a surveyor gets delayed in one area, another team member must step in to cover specific events. The current UI "locks" the column, making this mandatory compliance flexibility impossible without changing the entire day's assignment. This was pointed out by 5 out of 8 users.
Insight
If the Physical Environment Specialist gets stuck in another inspection, the Generalist often covers their next interview. The UI's rigid "Whole Column" assignment prevents this mandatory compliance flexibility.
Problem 2C: No Inline Editing
The current workflow breaks the user's mental model of direct manipulation as implied by 3 out of 8 users. When a user spots an error during a high-pressure briefing (e.g., a "Morning Team Meeting"), the system forces a context switch—navigating away to a separate form rather than allowing immediate correction. This friction discourages users from keeping the schedule up to date, as they fear losing their place or wasting time during critical "live" discussions.
Insight
During audit milestones, the agenda often changes instantly based on new findings. Forcing navigation away from the view creates a workflow gap where data entry errors occur, or worse, where the surveyor abandons the digital tool entirely in favor of a notebook.
These changes shift the interface from a simple daily list to a strategic planning tool, allowing the team to visualize the full 3-5 day survey and anticipate hard stops.
Day/Week view toggles with prominent chevrons and star icons for marking important tasks
This addresses user frustration regarding rigid assignments and supports the real-world compliance workflow where surveyors must cover for one another during complex inspections.
Draggable task cards enable quick reassignment to different roles or time slots following Tracer methodology
This capability is critical for "live" audit scenarios (such as Morning Briefings or Exit Conferences) where the team must rapidly adjust the schedule based on immediate findings without losing their visual context or focus.
Task edit modal enables quick on-the-spot editing adhering to the tracer methodology
Duplicate tabs with similar information and purpose cause decision friction, reducing surveyor productivity during critical documentation tasks.
The My Submissions and Entry Review tabs create confusion with duplicate functionality, poor CTA visibility, and redundant information display
Problem 3A: Confusing Tabs & Poor CTA Communication
Users (Surveyors) experienced confusion during their initial interaction with the My Submissions page. Their expectation was to quickly submit entries from the surveys they had conducted, but the interface increased their interaction cost and reduced their productivity. Most users were unable to distinguish between My Submissions and Entry Review, questioned the complexity of the flow, and needed initial guidance to understand the purpose and actions within each tab.
They also expected to navigate back to understand the screen flow better but couldn't, due to the absence of a back button. Additionally, many users overlooked the primary CTA because it did not effectively communicate itself as the main action.
Insight
4 out of 8 users faced confusion differentiating these functionally similar tabs. The redundant navigation structure and hidden CTA created unnecessary friction during time-sensitive documentation tasks.
Problem 3B: Unclear Entry Review Structure
Under the 'Entry Review' tab, users were unable to clearly understand what was happening because two separate information sections—a table and a card—were displayed simultaneously. This created confusion and increased their cognitive load, leading some users to abandon the task altogether, while others needed external assistance to navigate the page.
Insight
5 out of 8 users faced this issue. The simultaneous display of redundant information sections violated the principle of progressive disclosure, overwhelming users who simply wanted to complete their documentation.
These changes allow users to focus on the actual task they come to this page for, without overwhelming them with information overload, resulting in faster submission and review times.
Consolidated 'Submissions' tab with prominent CTA, bulk action checkboxes, row-specific actions, and cleaner table design
"I like a sleek platform, and it's really aesthetically pleasing to the eye. What I really like about this is having the different columns."— Participant 3, noting positive aspects
The Outcome
The MedLaunch team received our findings enthusiastically. Robert Dostert expressed that our solutions aligned with development priorities and would be prioritized for near-term implementation.
Field testing is essential. Since surveyors use this tool while walking through hospitals with spotty connectivity, we recommended tablet-specific usability sessions in real clinical environments.
This project reinforced that domain expertise doesn't compensate for interface issues. Even healthcare-background participants struggled with workflows.
Implementing these recommendations could move the Product Usability Scale score from 59.4 (Grade D) to 70-80 range (Grade B-C) within one to two development cycles.