Site icon Kellblog

Using Triangulation Forecasts For Improved Forecast Accuracy and Better Conversations

Ever been in this meeting?

CEO:  What’s the forecast?
CRO:  Same as before, $3,400K.
Director 1:  How do you feel about it?
CRO:  Good.
Director 2:  Where will we really land?
CRO:  $3,400K.  That’s why that’s the forecast.
Director 1:  But best case, where do we land?
CRO:  Best case, $3,800K.
Director 2:  How do you define best case?
CRO:  If the stars align.

Not very productive, is it?

I’ve already blogged about one way to solve this problem:  encouraging your CRO think probabilistically about the forecast.  But that’s a big ask.  It’s not easy to change how sales leaders think, and it’s not always the right time to ask.  So, somewhat independent of that, in this series I’ll introduce three concepts that help ensure that we have better conversations about the forecast and ultimately forecast better as a result:  triangulation forecasts, to-go pipeline coverage, and this/next/all-quarter pipeline analysis.  In this post, we’ll cover triangulation forecasts.

Triangulation Forecasts

The simplest way to have better conversations about the forecast is to have more than one forecast to discuss.  Towards that end, much as we might take three or four bearings to triangulate our position when we’re lost in the backcountry, let’s look at three or four bearings to triangulate our position on the new annual recurring revenue (ARR) forecast for the quarter.

In this example [1] we track the forecast and its evolution along with some important context such as the plan and our actuals from the previous and year-ago quarters.  We’ve placed the New ARR forecast in its leaky bucket context [2], in bold so it stands out.  Just scanning across the New ARR row, we can see a few things:

I’m always impressed with how much information you can extract from that top block alone if you’re used to looking at it.  But can we make it better?  Can we enable much more interesting conversations?  Yes.  Look at the second block, which includes four rows:

Each of these tells you something different.

These triangulation forecasts provide different bearings that can help you understand your pipeline better, know where to focus your efforts, and improve the accuracy of predicting where you’ll land.

For example, if the rep- and manager-level forecasts are well below the CRO’s, it’s often because the CRO knows about some big deal they can pull forward to make up any gap.  Or, more sinisterly, because the CRO’s expense budget is automatically cut to preserve a target operating margin and thus they are choosing to be “upside down” rather face an immediate expense cut [11].

If the stage-weighted forecast is much lower than the others, it indicates that while we may have the right volume of pipeline that it’s not far enough along in its evolution, and ergo we should focus on velocity.

Now, looking at our sample data, let’s make some observations about the state of the quarter at SaaSCo.

Just by these numbers, if I were running SaaSCo I’d be thinking that we’re going to land between $2,800K and $3,200K [13].  But remember our goal here:  to have better conversations about the forecast.  What questions might I ask the CRO looking at this data?

In our next post in the series we’ll discuss to-go pipeline coverage.  A link to the spreadsheet used to the example is here.

# # #

Notes

[1] This is the top of the weekly sheet I recommend CEOs to start their weekly staff meeting.

[2] A SaaS company is conceptualized as a leaky bucket of ARR.

[3] I cheated and look one row down to see the churn forecast was $500K in weeks 1-6 and only started coming down (i.e., improving) as the CRO continued to cut their New ARR forecast.  This makes me suspicious, particularly if the VP of Customer Success reports to the CRO.

[4] I cheated and looked one row up to see starting ARR growing at 58% which is not going to sustain if New ARR is only growing at ~20%.  I also had to calculate planned growth (3900/2850 = 1.37) as it’s not done for me on the sheet.

[5] Assumes a world where managers do not forecast for their reps and/or otherwise cajole reps into forecasting what the manager thinks is appropriate, instead preferring for managers to make their own forecast, loosely coupling rep-level and the manager-level forecasts.

[6]  Typically, the sum of the forecasts from the CRO’s direct reports.  An equally, if perhaps not more, interesting measure would be the sum of the first-line managers’ forecasts.

[7] Expected value is math-speak for probability * value.  For example, if we had one $100K oppty with a 20% close probability, then its expected value would be $100K * 0.2 = $20K.

[8] A stage-weighted expected value of the (current quarter) pipeline is calculated by summing the expected value of each opportunity in the pipeline, using probabilities assigned to each stage.  For example, if we had only three stages (e.g., prospect, short-list, and vendor of choice) and assigned a probability to each (e.g., 10%, 30%, 70%) and then multiplied the new ARR value of each oppty by its corresponding probability and summed them, then we would have the stage-weighted expected value of the pipeline.  Note that in a more advanced world those probabilities are week-specific (and, due to quarterly seasonality, maybe week-within-quarter specific) but we’ll ignore that here for now.  Typically, one way I sidestep some of that hassle is to focus my quarterly analytics by snapshotting week 3, creating in effect week 3 conversion rates which I know will work better earlier in the quarter than later.  In the real world, these are often eyeballed initially and then calculated from regressions later on — i.e., in the last 8 quarters, what % of week 3, stage 2 oppties closed?

[9]  The forecast category-weighted expected value of the pipeline is the same the stage-weighted one, except instead of using stage we use forecast category as the basis for the calculation.  For example, if we have forecast categories of upside, forecast, commit we might assign probabilities of 0.3, 0.7, and 0.9 to each oppty in that respective category.

[10] Sometimes embarrassingly so for the CRO whose forecast thus ends up a mathematical negative value-add!

[11] This is not a great practice IMHO and thus CEOs should not inadvertently incent inflated forecasts by hard-coding expense cuts to the forecast.

[12] The point being there are two ways to fix this problem.  One is to revise the probabilities via regression.  The other is to apply a correction factor to the calculated result.  (Methods with consistent errors are good predictors that are just miscalibrated.)

[13]  In what I’d consider a 80% confidence interval — i.e., 10% chance we’re below $2,800K and 10% chance we’re above $3,200K.

Exit mobile version