
Comparisons may be odious, but they’re essential for learning from data. Faced with performance metrics across outlets, rigs, factories, or service centres, we build league tables hoping poor performers can learn from the best.

The problem is that the manager of a unit in the worst-performing quarter of the database will immediately identify ten reasons why his unit is different from the best performers, along dimensions he cannot change. There will often be no observation exactly like his on all ten dimensions.
A common remedy for this is to cluster or categorize observations along each dimension (e.g. “large” vs “medium” vs “small”; or “urban” vs “suburban” vs “rural”; or “north” vs “midlands” vs “south”), and compare within cells of the resulting matrix. The problem with this approach is twofold:
PIMS has developed two methods over the years to solve this problem. The first step is to do research on the database using cross-tabulations and common sense to identify which dimensions of difference actually make a difference to performance. This naturally leads to multivariate statistical modelling using regression or similar methods to quantify which are the really powerful performance drivers, and enables a comparison of actual performance against “par performance” derived from the position of a this unit on the key drivers. The performance impact of each driver can be calculated as the effect of moving from the database average to the actual value of this observation on the driver, taking the drivers “most beyond management control” first. “PIMS method one” is to use the model directly to see where to position the observation on the key controllable drivers.
“PIMS method two” is to find look-alikes. This proceeds by calculating a distance from this observation to each other observation in the database along the drivers beyond management control (there are various ways of combining the drivers, e.g. a Euclidean distance is the square root of the sum of squares of the distances in standard deviations). Then we find say the nearest 4% of observations, and split them into “winners” (performing well) and “losers” (performing poorly). Where there is a significant difference between “winners” and “losers” – and this observation resembles the “losers” – we have identified either:
The look-alike methodology has several advantages over other approaches:
PIMS has wide practical experience of developing look-alikes in different situations; from cross-industry strategic business units (matching on competitive strength, market attractiveness, and capital/labour intensity) to various types of production plants or warehouses (matching on scale, complexity, product mix, automation and labour cost environment). We have developed different distance metrics (e.g. non-Euclidean and asymmetric), different ways of dealing with cross-sectional versus time-series comparisons, and different ways of showing results for different situations. If you want to compare against look-alikes – without being odious - talk to us first!