Improving Admin Center setup and deployment times

Revising the Viva product suite's setup and management dashboard to improve setup times by creating a consistent experience

Overview

The Viva app management dashboard in the Microsoft admin center allows admin users to setup, deploy, and manage the Viva apps purchased by an admin’s company

The driving principle of the dashboard is to help customers ensure all app features are fully setup and running. Tasks for users include things like giving proper admin permissions; audience and role assignment; content sources; and other app specific steps

Solution

An updated setup dashboard that reports status and a consolidated and more standardized setting configuration process

Results

Reduction of setup and deployment times

Constraints

Working within a rigid design system
6 week design and ship timeline (+ 2 weeks of A/B testing)

The problem:


Based on submitted feedback and interview data, users were having difficulty with the app setup process. Without a way to check app status and complete the setup process in one place, users were missing key steps and task-switching.


This could cause low usage rates and impact customer retention due to a lengthy time-to-value perception.

01. Finding the cause of the problem

Auditing the design

Using data gathered from user interviews, I ran a product audit to identify specific areas where the problems were occurring

Notable findings:

1

Lack of information that left users in the dark about what is complete, in progress or even needs to be started

2

The look and feel of setup pages could vary wildly and sometimes occur outside of the Admin Center almost entirely

3

Users had a general uncertainty about what steps needed to be done first or if the steps needed to be done at all

02. lightweight cross-functional workshopping

Collaborating on ideas

Design approach

I used the audit results to help support a collaborative session with my PM and engineering partners. We first discussed what our key solutions should encompass.

From there, we generated some early ideas via a crazy 8s session and then dot-voted to narrow down our top three. This helped to build alignment in the group about potential approaches.

Sorting ideas

To decide on the areas where we could most easily create impact, I guided the team to sort ideas from the workshop into three categories of potential approaches:

1

Notify

Ideas that focused on improving how users were informed of any action needed

2

Organize

Approaches that could present info in a more consistent way to streamline the process

3

Consolidate

Ways to bring all setup steps into the same admin center dashboard

The highest and fastest areas we could create impact in were identified as the notify and organize areas. Putting our effort here first would allow us to maximize the effectiveness of our work, while making sure things were kept in scope and feasible.

With the group now aligned around a core idea, I was able to develop a hypothesis that underlined what our focus would be.

After discussing, we decided our focus should be on the areas where we could most easily affect change (+ minimize engineering effort) and generate noticeable impact.

This was identified as the notify and organize levels. Putting our effort into these areas first would allow us to quickly generate insights that could guide future iterations of the dashboard.

The hypothesis

By surfacing relevant info upfront and providing a consistent experience across apps, users will finish setup faster and be confident that they completed their task.


This will improve setup times and setup success rates. As a result, app usage and customer churn rates will also improve across the Viva product suite.

03. Iterating

Exploring design options

As I explored designs I wanted to make sure I kept in mind the following:

  • A minimal disruption to users typical workflow

  • Staying within Microsoft Admin Center's rigid design system guidelines

  • Creating a scalable, flexible design for different app setup needs and future updates

Design Flow A

  • Separate settings out in a table similar to the landing page

  • Surface the priority of each setting on the app setup page

  • Place each setting's individual steps in a fly-out component

Design Flow B

  • Surface setting setup upon entering page

  • Sort each setting into a guided wizard

  • Mark "priority" steps with a relevant icon

04. Guiding direction with Quantitative data

A/B/C testing to validate our decisions

Measuring success

To ensure that we gained value from the test, my project PM and I met to establish what would define a “successful” test

1

Success rate

Users in the non-control groups should be able to setup or configure an app successfully more often than the control

2

Setup time improvement

Metric should have a noticeable improvement of at least 5 minute setup time per app over the control version

3

Negative qualitative notes

Did we receive any noticeably negative qualitative feedback from users during the test

Testing

Once the new designs were built by engineering, they were pushed to production where user traffic was split across the 3 options (control vs. the a/b variations). The experiment ran for close to 3 weeks.

I worked with engineering to make sure we frequently and prominently surfaced our feedback widget to gather qualitative data in addition to quantitative. I also worked closely with a data science team to help finalize the metrics and understand the outcomes.

The test stats

Over the 3 week testing period our experiment had approximately 5,000 admin users

Results

20

Minute decrease (Version A)

Version A resulted in users decreasing their setup time by over 10 minutes. A 60% reduction in setup time per app!

2.5

Minute decrease (Version B)

Version B resulted in users decreasing their setup time by 2.5 minutes. While positive, users did better with version A.

98%

Success rate

Users had a success rate when setting up any app of over 98%!

Based on the results from the test, we determined that the version that focused on showing status of all apps upfront provided the best time improvement for users and also returned more positive qualitative feedback.

05. Collected qualitative data pushes for updates

Iterating from feedback

Following user feedback

While the results from the status and table version already showed our designs were successful, some users had mentioned that the sudden shift from the old design confused them.

Finding a solution

We hypothesized that the shift in design may have strayed too far off from a user's existing mental model for setup in the admin center.

To counter this, we also added in an onboarding process to guide users through the new system.

This shift in product strategy helped to anticipate adoption risks and mitigate them through design.

Key impacts

Task time reduction

Users were able to setup and deploy Viva apps faster, Leading to decreased help ticket submission from setup.


Users were able to setup and deploy Viva apps 60% faster than with the old design. Leading to decreased help ticket submission from setup.



Users were able to setup and deploy Viva apps 60% faster than with the old design. Leading to decreased help ticket submission from setup.


Setup success rate increase

The setup success rate of apps in the admin center increased shortly after launch



Users were able to setup and deploy Viva apps 60% faster than with the old design. Leading to decreased help ticket submission from setup.



Users were able to setup and deploy Viva apps 60% faster than with the old design. Leading to decreased help ticket submission from setup.


Time-to-value improvements

Post-launch, customers gave qualitative feedback that they felt value quicker from Pulse.



Project takeaways

Areas to improve on

While we had strong metrics to understand the strength of the designs presented in A/B testing, it would have been worth setting up methods to gain additional qualitative feedback. This would have allowed us to better measure usability.

Next steps - Implementing opportunities to promote premium plans

While not within the scope of the project, I identified an opportunity to leverage unused white space for promotion premium features (known as "add-ons").

This would not only make a user aware of available features that would improve their experience, but also take could work to notify users if a "add-on" was purchased and needed setup.

Unfortunately, this idea was tabled after I moved to a different product team, but at the time, it was noted by product leadership as an idea that could boost sales and adoption rates, while also increasing value customers received from their Viva apps!

NEXT PROJECT

Data
Dashboard

Improved response rates

Improved response rates