A rework of Pulse's existing usage dashboard

Rethinking the data visualization experience in Pulse

COMPANY

Microsoft

ROLE

Lead Designer

Duration

3 months

YEAR

2024

Summary

Pulse's usage dashboard is an admin-level data tool that helps users understand how Pulse is being used within their organization.

As the lead and sole designer on this project, I leveraged customer feedback to create targeted improvement to the existing dashboard. The updated design gave users additional requested metrics and centered the dashboard around taking action on insights.

Results

20% increase in survey response rates

01. Users want more
The existing experience was missing the mark for users

This MVP version was falling short of what users needed due its rigidity in data displays and lack of actionable guidance

Auditing customer feedback

I started by auditing feedback submissions from users to understand their frustrations and needs. This quote represented how most users seemed to be feeling:

" The metrics help only at a surface level, I can't do much with it… "

02. Teamwork makes the product work
User interviews + workshopping created strong team alignment

To start, I conducted user interviews to get a deeper sense of user's current pain points with the dashboard

Key user frustrations learned from interviews

Leveraging a 6-person user interview, I identified core frustrations users were experiencing with the current product.

1

Data dead-ends

Users are unable to close the loop on insights they see or do not have the metrics they want

2

External action agitation

The added work from having to go outside the dashboard to view or parse metrics

3

Overly broad data

Data provided is too broad and can’t be applied effectively to smaller groups

Establishing goals and generating early ideas through workshopping

So I finished up user interviews and now I've got all of this qualitative data… But what was I supposed to do with it?

First, I setup a cross-functional workshop that would progress through generating HMWs to frame the problem for users and then help us establish goals to focus on and generate ideas through a crazy 8's workshop.

An alignment on our goals and design direction

This workshop was impactful as it helped us to define and refine user and business goals and align on early design direction

03. First pass on high-fidelity designs
Alignment on goals and features created a smooth transition into design

With alignment on design direction and the goals the team wanted to focus on, I moved into wireframes and then onto a high-fidelity design to take through user testing.

First high-fidelity iteration breakdown

A focus on taking action

Additional metrics incorporated to highlight setup progress.

A "Take action" section that houses top-of-mind areas for users to leverage in managing settings and communicating with teams

Staying in the flow-of-work

Surfacing in-line modals that allow users to take key steps

New metrics and filters

More granular filter options to drill down by specific departments along with metrics that provide a better picture of user's engagement in Pulse and offer different chart views when possible

04. Pivot
User testing revealed that Users felt the in-app email wouldn't help

Back to the drawing board

Through low-fidelity designs, I explored a few different ideas that would provide more targeted actions for users while keeping things simple and confirming with engineering that any ideas would be technically and timeline feasible

Lightweight testing the new designs

I ran lightweight tests using clickable low-fidelity versions of three different ideas in front of users.

Feedback was the most positive on a design that allowed users to pick audience grouped by response rate level and write a message to them.

Shift in design visualized

Users felt having to add teams and then draft an email was basically something they already did. They wanted something fast so they can easily reach groups with low response rates

The original author flow only took into account survey time and length


This update would allow users to only target low response rate groups if they wanted, and also leveraged a design that would keep them more within the experience

05. A dashboard that delights
The shipped design that helped boost user engagement

After implementing updates from usability testing, I was able to put together a final prototype and get final buy in from all stakeholders… and soon after, a new dashboard rolled out to users!

06. did it work?
design outcomes

Leveraging metrics that product, engineering, and I had partnered together to determine how to measure success post-launch as well as user satisfaction surveys and feedback submissions, we felt the launch was (mostly) a success!

40%

Click-through on metric help

40% of users had repeat and sustained use use of the new metric help buttons

20%

Response rate increase

Of 40 random users surveyed, a reported 20% increase on average in user engagement at their orgs

-5%

Drop in external actions

Less users were relying on external actions such as data export, however this number could be still be improved

07. Learning is power
Project lessons

Thank you for reading through this case study! Below I've shared a couple of moments that felt worth reflecting on from this project:

Missing the mark and adapting

Despite having data going into the design, you might not always get what users were hoping to solve the problem.

I learned to validate designs with users at multiple stages if possible and how to best adapt while including cross-functional partners if there is a need to pivot the design.

Running workshops

Running workshops are excellent tools for gaining alignment. I learned a lot about how to smoothly run a multi-stage workshop and how to make the most out of the outcomes.

Next step - investigating additional external actions users take

Due to the lower than unexpected drop in external actions done by users, the next steps for this project would be talking to users to drill down further on what specific actions users take outside of the dashboard. Then, determine how those actions can be better incorporated.