We get a lot of questions from our clients about benchmarking research. What is it? How do we do it? Will it help us meet our goals? It’s easy to get lost in the storm and feel stuck not knowing how to proceed. Over the years, we’ve refined our benchmarking best practices to address all of these questions and create a streamlined, effective approach for our clients.
At its core, UX benchmarking is a comprehensive test of usability, pairing hard metrics with qualitative insights to get a high-resolution perspective on the effectiveness of your products. Unlike small scale usability tests, benchmarking aims to understand how the components of a system contribute to overall usability, and how a product as a whole meets the needs of its users.
Benchmarking any product, digital or hardware, starts by asking: What is this product intended to do, and how well is it doing it?
Benchmarking provides both comprehensive “big picture” frameworks for overall product health and detailed usability information in observational and participant feedback data. The combination gives you available knowledge and a holistic, prioritized picture of product health.
The pairing of observational and participant feedback provides data for several analytical approaches, giving you a range of benefits and outcomes. Here are some benefits of benchmarking programs we see with our clients:
Benchmarking is most valuable when repeated, to maximize the value of data comparisons over time and to capture the impact of changes in the interface. While both are repeatable approaches, depending on what you hope to accomplish with your benchmarking program you may want to pursue periodic comprehensive benchmarking or take a progressive benchmarking approach.
You’ve likely heard of comprehensive benchmarking before, as it’s the most common approach in UX research. This form of benchmarking tests all aspects of a product at periodic intervals such as quarterly, biannually, or annually. This approach is a complete assessment of usability, testing everything your product is “meant to do” in order to gauge overall product health and identify high and low performing areas.
Comprehensive benchmarking is best suited when your need is for a complete or comprehensive dataset at one point in time. It may also be the best research design if your product’s components are dependent on one another and testing individual components would isolate them in a way that was highly unrealistic to the live experience.
Additionally, periodic comprehensive benchmarking is a useful approach when you want to gauge impact over time, such as year-over-year changes or when you would like to make comparisons with competitors. It’s good to use when launching or relaunching, or seeking foundational data on which to build product roadmaps.
Progressive Benchmarking is a more paced approach than comprehensive benchmarking, intended to provide a consistent flow of benchmarking data to aggregate into a whole while making consistent usability improvements over time. In this approach, individual components of a system are tested at a regular pace, such as biweekly, in order to continuously assess parts of the system, building into a complete body of data about your product. This is a more “bite-sized” approach, intended to allow teams to focus on the health and usability of individual flows.
Progressive benchmarking is best suited when you’re iterating on or releasing components of a product or product suite independently of one another and want to test them individually. This approach can be especially useful when you want to integrate benchmarking into your regular research program and build a body of data to guide your product roadmap over time.
These two approaches can be thought of on a spectrum of sample size, qualitative vs. quantitative, time, and investment, and either approach can be customized to be smaller or larger in scale and scope. In either approach the key to successful benchmarking is creating a custom research design that is focused on highly valid measurements for your particular product.
Want to learn more? Check out Three Common Benchmarking Metrics to Ditch and What to Use Instead.