OPENTEXT WEBROOT

Enhancing security training insights

SAT campaign summary report
Overview

Webroot's Security Awareness Training(SAT) is designed to help small/medium business managed service providers(MSPs) and IT admins educate employees or clients on cybersecurity best practices by completing campaigns.

A campaign summary report is an essential tool for our core users, as it provides detailed insights into the effectiveness of the training campaigns they organize.

However, an improvement had to be made since user feedback indicated that the existing report design was difficult to interpret and a want for interactive features.

Project Scope

A revamped campaign summary report that offers MSPs and IT administrators easier error resolution and enhanced readability for analyzing training performance.

Role

UX Design

UI Design

Tools

Figma

Jira

Confluence

Team

1 Designer

1 Product Manager

1 User Researcher

2 Software Engineers

Duration

6 months

Overview

Webroot's Security Awareness Training(SAT) is designed to help small/medium business managed service providers(MSPs) and IT admins educate employees or clients on cybersecurity best practices by completing campaigns.

A campaign summary report is an essential tool for our core users, as it provides detailed insights into the effectiveness of the training campaigns they organize.

However, an improvement had to be made since user feedback indicated that the existing report design was difficult to interpret and a want for interactive features.

Project Scope

A revamped campaign summary report that offers MSPs and IT administrators easier error resolution and enhanced readability for analyzing training performance.

Role

UX Design

UI Design

Tools

Figma

Jira

Confluence

Team

1 Designer

1 Product Manager

1 User Researcher

2 Software Engineers

Duration

6 months

ABOUT THE PRODUCT
What is a "campaign summary report"?
Who is it for?
What is a "campaign summary report"?
Who is it for?
What is a "campaign summary report"?
Who is it for?

The campaign summary report is a feature provided as a part of Security Awareness Training (SAT) from Webroot that provides insights into the effectiveness of cybersecurity training campaigns and phishing simulations. Primary users of Webroot are managed service providers (MSPs) and IT admins who take care of IT for operations their clients. They are externally hired by mostly small to medium businesses who don't have their own IT department in their organizations.

Thus, it is MSPs' jobs to educate their clients on cybersecurity best practices.

THE PROBLEM

MSPs are having a hard time reading and resolving campaign errors, often leading to frustration and delays in campaign progress for their clients.

This can lead to a decline in training performance outcomes, making employees and clients vulnerable to cyber attacks.

SOLUTION PREVIEW
PROCESS OVERVIEW
EMPATHIZE
USER INSIGHTS
USER INSIGHTS
USER INSIGHTS
Listening to the voices of our users and being on their eye-level was crucial.

The project came to surface when my team was going over feedback they gathered from direct contact with customers. Most of the feedback had a similar theme, which meant that many users were facing similar challenges. It became clear that their pain points needed to be addressed, and these user insights were a good starting point for us to discuss priorities. This gave us guidance on where to start for the revamp of the campaign summary report.

“I wish I could interact with the data more, like filtering results or drilling down into specifics.”

“I have to scroll through a lot of irrelevant details to find what I need. It’s not efficient.”

“I don't always get what's written in the table. It takes some time to absorb the info.”

USER PERSONA
USER PERSONA
USER PERSONA
Meet Alex, a 38-year old with 15 years of experience in cybersecurity.
Meet Alex, a 38-year old with 15 years of experience in cybersecurity.
Meet Alex, a 38-year old with 15 years of experience in cybersecurity.

Based on user insights, I created Alex, a fictional persona, that represents the core users of Webroot. Laying out and understanding Alex's needs and pain points allowed me to focus on the critical areas that required attention. This persona served as a key reference throughout the project, ensuring that every decision was made with a clear understanding of the users’ real-world challenges.

IT Manager
Chicago, IL
Background

Alex works at an MSP that provides IT services to small and medium-sized businesses. His primary responsibility is to oversee cybersecurity solutions for multiple clients, ensuring that their systems are protected against threats and compliant with industry regulations.

Pain Points
  • Difficulty in interpreting insights

  • Challenges in communicating complex data to clients

  • Time consuming security training analysis

Needs
  • Clear visualization of security metrics

  • Insights easy to communicate to non-technical clients

  • Identify security training success/fail rates

ANALYZE

Scroll inside the laptop to see the old design!

Scroll inside the laptop to see the old design!

Scroll inside the laptop to see the old design!

OLD DESIGN PROBLEM OVERVIEW
OLD DESIGN PROBLEM OVERVIEW
OLD DESIGN PROBLEM OVERVIEW
Critiquing the existing design with a fresh set of eyes.

Before starting on this project, I was fairly new to the team and was working on another part of Webroot. However, redesigning SAT's campaign summary report became top of priority, and I was assigned to this as my first leading project. Since I was still in the process of familiarizing myself with the SAT section of Webroot, it was a perfect opportunity for someone like me to analyze and define the problems of the existing UI.

Before starting on this project, I was fairly new to the team and was working on another part of Webroot. However, redesigning SAT's campaign summary report became top of priority, and I was assigned to this as my first leading project. Since I was still in the process of familiarizing myself with the SAT section of Webroot, it was a perfect opportunity for someone like me to analyze and define the problems of the existing UI.

Before starting on this project, I was fairly new to the team and was working on another part of Webroot. However, redesigning SAT's campaign summary report became top of priority, and I was assigned to this as my first leading project. Since I was still in the process of familiarizing myself with the SAT section of Webroot, it was a perfect opportunity for someone like me to analyze and define the problems of the existing UI.

Problem 1

Lack of information hierarchy

Problem 2

Unclear data visualization

Problem 3

Poor table readability

Problem 4

Ineffective
layout
IN-DEPTH ANALYSIS
IN-DEPTH ANALYSIS
IN-DEPTH ANALYSIS
  1. Lack of information hierarchy
  • All of the information in the report has similar visual weight, making it a challenge to differentiate between primary and secondary data.

    For example, the campaign overview (at top), phishing results, and the user results table contain key information about the campaign that must be communicated to MSPs at a glance. However, there is no visual emphasis or any kind of visual aid that draws in attention.

    Primary data

    Secondary data

  • The report lacks a flow of information, as it does not have a logical progression from a high-level overview to more in-depth data.

    The report starts from a high-level overview of the campaign, then continues with supporting information that does not have much significance in communicating the current progress of the campaign. The supporting information acts more like a disruptor of flow rather than an enhancer in the report. The report then continues with in-depth data that contains results-oriented insights.

    Supporting information

    In-depth data

  • Result-oriented content is placed near the bottom of the report, making the information look less important.

    Since MSPs prioritize campaign performance by focusing on insights and metrics, placing result-related information in the middle or at the end of the report diminishes its importance and hinders users from fully leveraging the report's potential. This can lead to critical oversights, such as missing errors when a campaign hasn’t been successfully delivered or when participating clients fail to complete the training.

  1. Unclear data visualization
  • Although the bar graph is straightforward, it takes up a good amount of real estate and does not follow the best practices for data visualization.

    The bar graph does not maximize the use of that space to convey more information, which could include additional data points, annotations, or trends that may be useful in helping MSPs understand the current state of a campaign.

  • The legend is too prominent, taking the user’s attention away from the graph itself.

    Additionally, the legend is not effective as it requires a description under each category to explain to MSPs its meaning. Although the categories are jargons that may need explanation for some, there are more effective ways to communicate that in data visualization.

  • The graph is missing an overview label on the x-axis, and there is no numerical scale on the y-axis.

    The specific terms used as labels on each bar may lack context without the legend, making it unclear what the bars represent by itself. Additionally, the absence of a numerical scale on the y-axis, despite percentage labels on the x-axis, hinders MSPs from accurately gauging the overall campaign success, failure, and error ratios.

  • The labels are too small for legibility at a glance.

    Lack of readability is an accessibility issue. This can result in hindering data interpretation.

  1. Poor table readability
  • Table column sizing is disproportionate, as columns such as Notes should have wider space compared to Status.

    The Notes column may contain detailed comments or explanations that require more space to be fully visible. If this column is too narrow, the text might be truncated or wrapped awkwardly, making it difficult for MSPs to read the entire note at a glance.

  • There is no toolbar in the table that help with finding data such as a search bar, a filter, and any other actionable items.

    Without a toolbar, it is difficult for MSPs to find any specific data they are looking for. Important data such as statuses that need immediate attention (e.g., Not Delivered) should be accessible with ease to resolve any issues.

  • There is no sort for ascending/descending data in each column for quicker navigation.

    Similar issue to a lack of toolbar, a lack of column sort can be a challenge for MSPs when looking for specific data or reviewing the campaign progress/status for their participating clients.

  • Lack of visual aids such as colour codes for important information such as Status to quickly communicate information at a glance.

    All of the status types are accompanied by a representative icon for better communication. However, in the case a table is full of data, the icons can be distracting and it is hard for MSPs to view important information at once. Furthermore, the icons are monochromatic, which ultimately does not help in proritizing data.

  1. Ineffective layout
  • The layout doesn't clearly distinguish between different sections, leading to a cluttered appearance and decreased efficiency in navigating the report.

    Without clear distinctions between sections, the report can appear cluttered and overwhelming. When different types of information (such as summaries, detailed data, and visual elements) are not visually separated, MSPs are confronted with a large, continuous block of content. This can make it difficult to focus on specific information.

    Campaign Materials

    Campaign Results

  • Elements that are not as important take up a lot of space, making the rest of the information on the report look less important.

    As unimportant elements such as Email Used and Inforgraphic Used are placed at the top of the report and also take up more space, the important elements (i.e., result-oriented data) appear crammed into smaller areas, resulting in inconvinience and inefficiency in viewing data.

HEURISTICS EVALUATION
HEURISTICS EVALUATION
HEURISTICS EVALUATION
Doing a quick assessment before moving onto brainstorming!

To progress to the next step, I decided to assess the old design using a UX principle-based approach, incorporating both qualitative and quantitative evaluation.

I chose Jakob Nielsen’s 10 Usability Heuristics as the framework for this assessment, as this was the most established guideline to follow. Given that my role was not centred on UX research, I was tasked with a general heuristic evaluation.

For the quantitative evaluation, I came up with a severity rating scale to rate each heuristic.

5 - Severe Issue

5 - Severe Issue

5 - Severe Issue

Significant usability problem that needs to be addressed

4 - Major Issue

4 - Major Issue

4 - Major Issue

Important usability problem that needs to be addressed

3 - Moderate Issue

3 - Moderate Issue

3 - Moderate Issue

Some usability concerns, but nothing critical

2- Minor Issue

2- Minor Issue

2- Minor Issue

Slight usability concerns that may not need immediate attention

1 - No Issue

1 - No Issue

1 - No Issue

Follows the guidelines well with no usability concerns

Interact with the chart to view details!

Interact with the chart to view details!

Interact with the chart to view details!

DEFINE
KEY CHALLENGES
KEY CHALLENGES
KEY CHALLENGES
What are the biggest challenges to solve and implement in Phase 1?

Once completing the analysis of the current single site training campaign, I had an understanding of the problems that users were facing on a deeper level. I thought about how the current design impacted our primary users (MSPs) and secondary users (the clients) not only during the experience, but beyond. On the surface, it was mostly superficial problems that were visible as discussed earlier, but once analyzing in-depth, I uncovered more profound issues that affected the overall usability and effectiveness of the tool. I then listed the challenges I determined were top priority to focus its improvement on for the redesign.

Helping MSPs interpret important information in a short period of time.

Translating complex data into bite-sized information that is communicable for both MSPs and their clients.

Allowing complex data to be handled in a flexible manner without tool constraints.

Distinguishing important information from supporting content while keeping the design minimalistic.

PHASE 1 GOALS
Translating key challenges into project goals!

Since Webroot operates as a B2B platform, it was essential to establish project goals that aligned with our users' business objectives. After discussions with the product manager, we identified four key project goals, each derived from the major challenges we uncovered.

Enhance UI accessibility for information.

Simplify complex data for easy communication.

Increase flexibility in data handling and navigation.

Keep a minimalist design while prioritizing key information.

DESIGN
POTENTIAL SOLUTIONS
POTENTIAL SOLUTIONS
POTENTIAL SOLUTIONS
First, we focused on making major changes.

Click zoom in for better readability🔍

I was then tasked with creating different versions of the Sankey diagram.

We explored various options to determine which best illustrates campaign status with clear detail while maintaining readability.

Option 1

Sankey diagram with legend. There is no separate label to indicate each status on the diagram, and users can hover to view the number of users under each status.

Option 2

Sankey diagram with labels. Statuses are displayed on the diagram, and on hover, users can see both the number and percentage of users at each stage.

Option 3

Sankey diagram with labels and numerical data. Number and percentage displayed on diagram. On hover, a tooltip containing meaning of each status appears.

PHASE 1 DESIGN
PHASE 1 DESIGN
PHASE 1 DESIGN
Redesign of single site report was ready for release.

Here are the improvements I made:

Sankey Diagram

We chose the sankey diagram with labels that displays a tooltip of status definition upon hover.

Course Completion Table

The new table was given a search bar and filter, as well as the number of courses completed.

Thumbnail Size

The size of the thumbnail was significantly decreased to use space more efficiently and to put emphasis on primary elements.

Courses Layout

Originally a grid view with detailed description, courses section was minimized with a horizontal overflow.

ITERATE
FEEDBACK
FEEDBACK
FEEDBACK
Gaining insight from existing users to prepare for Phase 2.

When we released Phase 1, we already anticipated on working on Phase 2 right away, as we had received additional insights on what users wanted to see and use in the campaign summary report. Before delving right into Phase 2, the product manager reached out to users to receive feedback for Phase 1 to use as a guide for Phase 2. Through this, we identified several pain points from Phase 1.

Phase 1 Feedback:
  • “The Sankey diagram is helpful for seeing the overall flow, but it’s hard to track progress at the individual user level. I need a way to pinpoint which users are facing what issues.”

  • “The visual representation of errors could be more distinct. It's easy to miss issues like delivery errors, especially in a busy report.”

  • “I like how the two tables were combined into one, but I feel like it's missing important info like error details.”

Phase 2 Requests:
  • “Changes in this release was a level-up. Now, I want to see the same improvements in multi-site.”

  • “Bridge the gap between what I'm seeing on screen and what is being printed. There's a lot of discrepencies in the PDF version.”

  • “When errors come up, there’s no way to take action directly from the report. I want to be able to resolve issues on my own without having to contact support everytime.”

SCOPING PHASE 2
Thanks to user feedback, the scope for Phase 2 was clear.

During our weekly Wednesday meeting, we learned that Phase 2 scope was bigger than Phase 1. The product manager discussed feature prioritization with the team, and this is what we decided to implement for Phase 2:

Independent error resolution flow

Delivery error acknowledgement feature

Re-send email feature

Multi-site report improvement to align with Single-site

PDF report redesign

PHASE 2 DESIGN
PHASE 2 DESIGN
PHASE 2 DESIGN
To understand how users would navigate in the new design, I created a user flow.

New features and feature redesigns have been highligted!

Phase 2 introduced a significant number of enhancements compared to Phase 1.

Below are the before and after of existing features😎

Before
After

1

Header: Icons were added for visual representation of each category. Purpose of the colour was to bring in more attention to the header content, as it was easily missed.

2

Data Visualization: Colours in the Sankey diagram were to mainly communicate errors, so the various colours in Phase 1 were removed and only blue (normal activity), red (error) and green (completion) were kept. Additionally, a segmented controller was added so users can switch between two kinds of charts: Sankey diagram and bar chart.

3

Table: Changed from Course Completion Details table to User Progress table. The previous table didn't provide any methods for users to resolve errors, whereas the new table allows users to acknowledge errors and re-send emails when necessary. They are also able to expand and collapse each row to view their clients' activity in more detail. The new table now displays accurate time for each activity, so users can manage their clients' trainings in real-time.

Below are designs of additional features✨

Multi-site Report
Designing the multi-site report for consistency.

The multi-site report is a feature that helps users set up training campaigns for multiple organizations simultaneously. At its core, the feature is similar to a single site report, but it gives a high-level view of a campaign before getting into the nitty gritty.

After I redesigned the single-site report, the team recognized the need to improve the multi-site report to ensure a consistent user experience and eliminate any discrepancies.

From multi-site, users can see which organizations are experiencing email errors or training delays, and which organizations had already completed their training.

PDF Report
Thinking of the user even after the experience ends.

The PDF report is an additional report that is often used when users want to share training progress insights directly with their clients (organizations).

My task for the PDF report redesign was to bridge the gap between the digital and printed (PDF) versions of the report, making sure that the users and their clients experience consistent communication regarding errors and training progress.

IMPLEMENT
FINAL DESIGN
FINAL DESIGN
FINAL DESIGN
Gaining insight from existing users to prepare for phase 2.

When we released phase 1, we already anticipated on working on phase 2 right away, as we had received additional insights on what users wanted to see and use in the campaign summary report. Before delving right into phase 2, the product manager reached out to users to receive feedback for phase 1 to use as a guide for phase 2. Through this, we identified several pain points from phase 1.