You can't make informed decisions if the information you're using can't answer your questions.
The report design working group sat around the table, sifting through the draft strategic performance report to suggest how to make it more useful. Measure by measure they chatted and suggested and critiqued and debated: "this one would look better if it was a bar chart", "yeah, I like the three-dimensional bar charts", "we should add another line to this chart because it would be interesting to show", "it's pretty easy to get Excel to turn this one into a stacked bar chart, that way we could get more information onto it". Then someone asked: "hang on, what questions are we trying to answer with these measures anyway?", and there was dead silence.
ARE YOU USING THIS KIND OF PERFORMANCE INFORMATION?
Performance reports are most commonly filled with a combination of information like the following:
- tables of comparisons of this month with last month, this month with the same month last year, year to date with target
- the default bar charts that Microsoft Excel formats for you
- fancy three-dimensional, stacked bar charts
- other charts that are crowded with information that looks visually exciting and at best, might be interesting
- occasionally some short term trends, consisting only of 5 or 6 or barely a few more points of data
- commentary about project or initiative progress or milestone completion
How many of these kinds of information are in your performance reports? How much of that information is read, valued, validly interpreted, understood and applied to inform decisions? How aligned to your organisation's strategic, tactical or operational priorities is this information?
WHAT'S WRONG WITH THIS KIND OF PERFORMANCE INFORMATION?
If you answered these three questions with "quite a lot", "not very much of it" and "not very well", then you'd be fairly normal. Most organisations' performance reports are created with only a basic awareness of good business statistics and even less of an awareness of the business questions the report should answer.
Most of the information provided in these reports is in the form of "limited comparisons", to borrow a phrase from Donald Wheeler (1). Limited comparisons can't really answer any business question, because they are not a representative picture of the performance results they monitor. This is due to the fact that there is always natural variation from month to month, week to week, day to day, and 2 points of data can never tell you what amount of variation is normal versus abnormal.
What makes matters worse, is that often it seems this information is designed this way purely because it always has been. Rarely is the design of the information questioned or challenged. And rarer still, is the design of the information reviewed in the context of the business questions it must answer. Information design, particularly the visual design of quantitative information (Edward Tufte has written some amazing books on this very topic - see references 2, 3 and 4), is a real body of knowledge, linking statistical theory with cognitive theory to provide insights into how we can make information more useful and usable in decision making.
WHAT ARE THE RIGHT BUSINESS QUESTIONS TO ANSWER?
What is business performance management really about? Ultimately it's about business success, and providing the information to make the decisions that increase that success, now and into the future. Performance management is about three specific things. Firstly, it's about monitoring your business' actual progress toward the outcomes (and targets) implied by your business strategy. If one of those outcomes is to increase customer loyalty, then it's about monitoring how much customer loyalty you have as time goes by (such as the average number of orders per customer per quarter), and comparing this actual level of customer loyalty with the targeted level (say 20 orders per customer per quarter).
Secondly, business performance management is about knowing which of your initiatives or projects are working and not working in making those outcomes happen. If you have an initiative around developing a customer relationship management system to improve customer loyalty, then you would expect to see that customer loyalty increases the more the customer relationship management system is implemented and used. If you don't see a change in customer loyalty despite implement customer relationship management, then how can you say the system was working? You can't. Thirdly, business performance management is about knowing why those things are working or not working so you can choose better things to do, or fix the things you are doing. Perhaps the customer relationship management system isn't impacting customer loyalty because customers are already happy with their relationships with you, but just feel your products or services aren't as relevant as they expected.
This description of business performance management suggests several specific questions are important in managing a business so it's success improves.
Have we achieved our target?
We monitor business performance so we can know when we are actually performing at the level we need to, or want to, perform at. Targets are the description of the "need to" or "want to".
Are we progressing toward our target?
Rather than just waiting to the end of the year, or the date we wanted to achieve the target by, monitoring continuously throughout that timeframe gives us more power to influence the end result.
Are their any unintended consequences of our actions?
Chaos theory, the butterfly effect, and system thinking all tell us that there will always be some kind of flow-on effect from our actions. These flow-on effects can be anywhere from small and insignificant, to shockingly dramatic.
Why are we getting the results we are getting?
This is a question that is seeking information about reasons. Of all the possible reasons that you are getting a particular performance result, which reasons are the main ones, those that have most of the impact? If you are getting a good result, then knowing these reasons helps you confirm what to celebrate and keep doing more of. If you are getting a bad result, then knowing these reasons helps you modify your course of action to turn the result around.
What is likely to happen in the future?
Predictive information is some of the most valuable information in business. While no information can really do the crystal ball thing, a really good understanding of drivers (or lead indicators) can certainly give some great clues about likely future results. Thus, you can prepare for the most likely outcomes, before they happen.
DESIGN YOUR INFORMATION TO ANSWER YOUR BUSINESS QUESTIONS
The types of information needed to answer these business questions are different for each question. And unless that information is designed with the question in mind, it's likely that you won't be able to answer the question, or if you do, it will suffer the risks of misjudgment.
One of the keys to designing the kinds of information that will support your judgment in answering these question is to start with identifying the type of comparison you are trying to make. What do you need to compare with what, in order to answer the question? Another key is to be familiar with the kinds of qualitative and quantitative analyses that can reliably make those comparisons for you. The following sections provide some examples of how these keys can be used to design information to answer the generic types of business questions above. HAVE WE ACHIEVED OUR TARGET?
This is basically a comparison between actual performance and targeted performance. But it's not quite as simple as that. Because business results are constantly affected day in, day out, as time goes by, this comparison can only be really valid if it takes into account the natural variation in results over time. This means that a simple comparison of actual performance for the year (such as the average number of orders per customer per month, rolled up into an annual statistic) compared to the target (of 20 orders per customer per month) is too simplistic. It doesn't take into account the very likely event that improvements may have happened within the last year, so it underestimates actual performance. A better analysis would be a run chart (1) that shows the real changes in the overall average as time goes by, and compares that latest overall average (the mean line in the run chart) with the targeted level.
ARE WE PROGRESSING TOWARD OUR TARGET?
This question requires a comparison quite similar to the question above: actual performance compared to targeted performance. But it's not just the comparison between the mean line and the target level that matters here, it's also a comparison of how the actual overall level (the mean line) is moving as time goes by. Is it moving closer to the target level, fast enough? If you relied just on monthly comparisons to target, you'd be mislead by the natural variation that happens from month to month. You need to see the big picture pattern or trend over time.
ARE THEIR ANY UNINTENDED CONSEQUENCES OF OUR ACTIONS?
A slightly more complex comparison is needed to answer this question. Answering this kind of business question means you need to have some idea of what kinds of unintended consequences you could have expected, and some information about the extent to which these consequences are occurring - before and after you take action to achieve the performance result you want.
When you have this information, it would ideally take a similar form to the information that answers the previous two questions: a run chart that shows real changes in the overall actual level as time goes by. You then can compare the patterns or trends over time of your performance result, with those of the unintended consequences, and if you see some kind of correlated pattern, there's a strong clue that by achieving your performance result, you are also getting some other kind of result as a consequence. This may be a good thing, but it also may be a bad thing. And by knowing, you can take action if it's needed.
WHY ARE WE GETTING THE RESULTS WE ARE GETTING?
The comparison type here is be able to see the relative size of impact of each of a range of possible reasons for the result you are getting. So step one is to have a good idea of what those reasons could be. The second step is to be able to source some data that lets you know how often, or to what extent, each reason has actually played out during the timeframe you monitored your performance result. A really useful analysis of such data is a Pareto chart. This shows the relative size of impact of each reason, from largest to smallest. It will highlight the 20% of reasons that are having 80% of the impact on your result. And viola, you have a place to start investigating further, to find how you can turn your result around.
WHAT IS LIKELY TO HAPPEN IN THE FUTURE?
The kind of comparison needed to answer questions like this is the comparison between or among a set of measures or factors or variables that you hypothesize have significant influence over the direction your performance result will head. These measures or factors or variables are your drivers, or lead indicators. You test these hypotheses through a scatter plot (entirely visual), correlation analysis (very visual, with a quantitative measure of strength of the relationship) or regression analysis (not visual, but with a quantitative model of the relationship) to determine the strongest of these lead indicators.
Then, using the run chart analysis described under previous questions, you can interpret the emerging trends in your lead indicators and estimate or calculate the impact this will have on your performance result.
DELIBERATELY DO TWO THINGS
Designing excellent information to inform decisions about business performance is not rocket science. But it usually does require some effort be applied to clearly articulating each business question that needs to be answered in order to understand and make the decisions, and applied to deliberately designing the kind of information that can adequately (even if not completely) answer those questions.
(1) Understanding Variation: The Key to Managing Chaos, Donald Wheeler, SPC Press, Inc., 1993
(2) The Visual Design of Quantitative Information, Edward Tufte, Graphics Press, 1983-1999
(3) Envisioning Information, Edward Tufte, Graphics Press, 1990
(4) Visual Explanations, Edward Tufte, Graphics Press, 1997
Stacey Barr is a specialist in organisational performance measurement, helping people get the kind of data and information that tells them how well their business is performing, and how to make it perform better. Sign up for Stacey's free 'mezhermnt Handy Hints' ezine at http://www.staceybarr.com
to receive your complimentary copy of her e-book "202 Tips for Performance Measurement".
Article Source: http://EzineArticles.com/?expert=Stacey_Barr