Thoughts on Hiring:  Working for TBH

One of the most awkward situations in business is trying to recruit someone who will work for to-be-hired (TBH).   For example, say you’ve started a search for a director of product marketing, have a few great candidates in play, only to have your marketing VP suddenly quit the company to take care of a sick parent.   Boom, you’re in a working-for-TBH situation.

These are hard for many reasons:

  • Unknown boss effect. While your product marketing candidate may love the company, the market space, the would-be direct reports, and the rest of the marketing team, the fact is (as a good friend says) your boss is the company.  That is, 80% of your work experience is driven by your boss, and only 20% by the company.
  • Entourage effect. Your top product marketing candidate is probably worried that the new marketing VP has a favorite product marketing director, and that they’ve worked together through the past 10 years and 3 startups.  In which case, if there is an entourage effect in play, the candidate sees himself as having basically no chance of surviving it.
  • False veto effect. You may have tried to reassure product marketing candidates by telling them that they will “be part of the process” in recruiting the new boss, but the smart candidate will know that if everybody else says yes, then the real odds of stopping the train will be zero.

So who takes jobs working for TBH?  Someone who sees the net gain of taking job the job as exceeding the risk imposed by the unknown boss, entourage, and false veto effects.

That net gain might be:

  • The rare chance to switch industries. Switching industries is hard as most companies want to hire not only from within their industry (e.g., enterprise software) but ideally from within their category (e.g., BI).
  • The rare chance to get a cross-company promotion. Most companies promote from within but when they go outside for talent, they want to hire veterans who have done the job before.  For example, when LinkedIn needed a new CEO they promoted Jeff Weiner from within.  When ServiceNow needed a new CEO and didn’t find anyone internally who fit the bill, they didn’t hire a first-timer, they hired Frank Slootman, who had been CEO at Data Domain for six years and lead a spectacular exit to EMC.
  • The rare chance to get promoted into the TBH job. Sometimes this is explicitly pitched as a benefit to person working for TBH.  In reality, while this rarely happens, it’s always possible that the new hire does so well in the job – and it takes so long to hire TBH – that the person gets promoted up into the bigger job.  This is generally not a great sign for the company because it’s a straight-up admission that they viewed the working-for-TBH hire as not heavy enough for the TBH job, but eventually gave up because they were unable to attract someone in line with their original goals.

Who doesn’t take jobs working for TBH?  Veterans — who, by the way — are precisely the kind of people you want building your startup.  So, in general, I advise companies to avoid the working-for-TBH situation stalling the next-level search and hiring the boss first.

Making the working-for-TBH hire is particularly difficult when the CEO slot is open for two reasons:

  • E-staff direct reports are among the most sophisticated hires you will make, so they will be keenly aware of the risks associated with the unknown-boss, entourage, and false-veto effects. Thus the “win” for them personally needs to offset some serious downside risk.  And since that win generally means giving them opportunities they might not otherwise have, it means an almost certain downgrading in the talent that you can attract for any given position.
  • New CEO hires fail a large percentage of the time, particularly when they are “rock star” hires. For every Frank Slootman who has lined up consecutive major wins, there are about a dozen one-hit wonders, suggesting that CEO success is often as much about circumstance as it is about talent.  You need to look no farther than Carly Fiorina at HP, or any of the last 5 or so CEOs of Yahoo, for some poignant examples.  Enduring a failed new-hire CEO is painful for everyone — the company, the board — but no more group feels the pain more than the e-staff.  Frequently, they are terminated due to the entourage effect, but even if they survive their “prize” for doing so is to pull the slot-machine arm one more time and endure a second, new CEO.

A Missive to Human Resources (HR)

I built a very successful marketing career based on a three-word mission statement that I picked up from Chris Greendale, back when I was a product marketer at Ingres.  Chris always said that marketing exists to make sales easier.

I loved those three words.  I embraced that simple concept.  I made it my reductionist mission statement.  I taught it to every marketer I knew.  I loved it because it just made so much sense — if you derived a startup from scratch you’d first hire a developer and then salesperson.  Only after you had a bunch of salespeople would you then hire marketing, with the purpose of making the salespeople more effective.

Across my career, many people — ironically often from sales — challenged my “make sales easier” mission statement.  “It’s too tactical,” I’d hear.  Or, it “completely overlooks the strategic value of marketing.”  Not so, I’d counter.

  • Does picking a corporate strategy where we can win key market segments help make sales easier?  You bet it does.
  • Does designing better products for the target customer make sales easier?  You bet it does.

Simply put, while “make sales easier” might at first blush sound tactical in nature, the clever marketer can make sales easier in both tactical (e.g., lead generation) and very strategic ways.

Once, when I was thinking about human resources (HR), I wondered if I could come up with a similarly effective, reductionist mission statement.  I landed upon HR exists to help managers manage.

Like “make sales easier,” “help managers manage” often generates instant push-back.

  • Shouldn’t HR be focused on employee experience?  Yes, but don’t all employees work for managers, and isn’t it true if all our managers are doing a better job at managing, won’t our employees then have a better experience?
  • Doesn’t HR have an important legal and compliance role?  Yes, but don’t our managers want us in compliance?  I suppose it’s a bit like a police force motto of “to protect and serve,” but frankly I’d rather have the police define themselves as protecting and serving the community than any likely alternative.
  • Shouldn’t HR be focused on organizational development or talent management?  Yes and yes.  And helping managers to develop their teams and/or recruit talented new members is all part of helping managers manage.

Finally, it begs the question, shouldn’t HR represent the employee point of view, the vox populi, to the company?  My answer is no.  If I want to know what the average employee thinks about the company, I can run a survey.  In fact, I’ll probably ask HR to run that survey.  But I do not view chatting, jawboning,  or gossiping with employees as a core HR function.  Why?

  • Because when HR people enter the gossip chain, they are no longer an observer of the story, they are now part of the story.  Their opinion is just one in a sea of opinions and to assume that simply by virtue of having HR printed on their business card, that they can somehow be impartial aggregators of truth is not realistic.
  • Because I do not want to pay people to stir the pot.  Every company has issues, problems and challenges.  If you allow HR to define themselves as employee advocates or the keeper of the public-voice flame, you are, in effect, asking them to go stir the pot.  I greatly prefer chartering HR with a “help managers manage” mission which often translates to “help managers get stuff done” and then, when people/conflict/cultural/managerial issues come up they are not doing so in a vacuum, but in the specific context of what’s blocking progress on key organizational goals.

As I told marketers, “the more time a salesperson has to spend with you, the less you should care about his/her opinion” because the best people want to be out selling, not chatting with marketing.  I’d argue the same logic holds true for employees in general with HR.  As CEO, I want people focused on getting stuff done.  I care enormously about “soft issues” when they impede the organization’s progress on key goals.  When framed, however, against one individual employee’s views about how a company theoretically should work, well, I care less.

One sometimes difficult concept to grasp for support staff is that help is defined in the mind of the recipient.  HR staff, particularly those who come with a legal/compliance bent, may think that rejecting a poorly done performance improvement plan (PIP) is helping the company.  From the manager’s perspective, that’s not help:  showing them an example of  good one would be.

Alternatively, telling someone 10 reasons why they can’t terminate someone in the short-term isn’t help.  Sitting down and helping them understand the correct process and helping to make a PIP would be.

The more you are a cop who says no, the less you’re helping.  The more you are asking managers what they are trying to accomplish, devil’s advocating their viewpoints, and then helping them accomplish what they want to do, the more you are helping.

In the end, help is a way of approaching things.  So, HR, stop thinking about what people can’t or shouldn’t do and starting thinking about

  • How can I help managers hire great employees?
  • How can I help managers understand how their employees are doing?
  • How can I help managers — often in a very applied way — execute the annual review process and decide of annual raises for their team?
  • How can I help managers manage-out employees who aren’t succeeding?
  • How can I help managers develop their talent so they can move up within the organization?
  • How can I help managers become better managers overall?  Better interviewers?  Better feedback givers?  Better priortizers?  Better communicators?
  • How can I help managers live and communicate the core values?

Remember what’s sometimes called The Great Lie:  “we’re from corporate, we’re hear to help.”  How can you change this so that “we’re from HR, we’re here to help” instead becomes The Great Truth?

The Ultimate SaaS Metric: The Customer Lifetime Value to Customer Acquisition Cost Ratio (LTV/CAC)

I’m a big fan of software-as-a-service (SaaS) metrics.  I’ve authored very deep posts on SaaS renewals rates and customer acquisition costs.  I also routinely point readers to other great posts on the topic, including:

But in today’s post, I’m going to examine the question:  of the literally scores of SaaS metrics out there, if you could only pick one single metric, which one would it be?

Let’s consider some candidates:

  • Revenue is bad because it’s a lagging indicator in a SaaS business.
  • Bookings is good because it’s a leading indicator of both revenue and cash, but tells you nothing about the existing customer base.
  • ARR (annual recurring revenue) is good because it’s a leading indicator of revenue and includes the effects of both new sales and customer churn.  However, there are two ways to have slow ending ARR growth:  high sales and high churn or low sales and low churn — and they are very different.
  • Cashflow is good because it tends to net-out a lot of other effects, but can be misleading unless you understand the structure of a company’s bookings mix and payment terms.
  • Gross margin (GM) is nice because it gives you an indicator of how efficiently the service is run, but unfortunately tells you nothing else.
  • The churn rate is good because it helps you value the existing customer annuity, but tells you nothing about new sales.
  • Customer acquisition cost (CAC) is a great measure of sales and marketing efficiency, but by itself is not terribly meaningful because you don’t know what you’re buying:  are you paying, for example, $12K in sales and marketing (S&M) expense for a $1K/month customer who will renew for 3 months or 120?  There’s a big difference between the two.
  • Lifetime value (LTV) is good measure of the annuity value of your customer base, but says nothing about new sales.

Before revealing my single best-choice metric, let me make what might be an unfashionable and counter-intuitive statement.  While I love SaaS “unit economics” as much as anybody, to me there is nothing better than a realistic, four-statement, three-year financial model that factors everything into the mix.  I say this not only because my company makes tools to create such models, but more importantly because unit economics can be misleading in a complicated world of varying contract duration (e.g., 1 to 3+ years), payment terms (e.g., quarterly, annual, prepaid, non-prepaid), long sales cycles (typical CAC calculations assume prior-quarter S&M drives current-quarter sales), and renewals which may differ from the original contract in both duration and terms.

Remember that SaaS unit economics were born in an era of monthly recurring revenue (MRR), so the more your business runs monthly, the better those metrics work — and conversely.  For example, consider two companies:

  • Company A does month-to-month contracts charging $100/month and has a CAC ratio of 1.0.
  • Company B does annual contracts, does three-year prepaid deals, and has a CAC ratio of 2.0.

If both companies have 80% subscription gross margins (GM), then the CAC payback period is 15 months for company A and 30 months for company B.  (CAC payback period is months of subscription gross margin to recover CAC.)

This implies company B is much riskier than company A because company B’s payback period is twice as long and company B’s money is at risk for a full 30 months until it recovers payback.

But it’s completely wrong.  Note that because company B does pre-paid deals its actual, cash payback period is not 30 months, but 1 day.  Despite ostensibly having half the CAC payback period, company A is far riskier because it has to wait 15 months until recovering its S&M investment and each month presents an opportunity for non-renewal.  (Or, as I like to say, “is exposed to the churn rate.”)  Thus, while company B will recoup its S&M investment (and then some) every time, company A will only recoup it some percentage of the time as a function of its monthly churn rate.

Now this is not to say that three-year prepaid deals are a panacea and that everyone should do them.  From the vendor perspective, they are good for year 1 cashflow, but bad in years 2 and 3.  From the customer perspective, three-year deals make plenty of sense for “high consideration” purchases (where once you have completed your evaluation, you are pretty sure of your selection), but make almost no sense in try-and-buy scenarios.  So the point is not “long live the three-year deal,” but instead “examine unit economics, but do so with an awareness of both their origins and limitations.”

This is why I think nothing tells the story better than a full four-statement, three-year financial model.  Now I’m sure there are plenty of badly-built over-optimistic models out there.  But don’t throw the baby out with the bathwater.   It is just not that hard to model:

  • The mix of the different types of deals your company does by duration and prepayment terms — and how that changes over time.
  • The existing renewals base and the matrix of deals of one duration that renew as another.
  • The cashflow ramifications of prepaid and non-prepaid multi-year contracts.
  • The impact on ARR and cashflow of churn rates and renewals bookings.
  • The impact of upsell to the existing customer base

Now that I’ve disclaimed all that, let’s answer the central question posed by this post:  if you could know just one SaaS metric, which would it be?

The LTV/CAC ratio.

Why?  Because what you pay for something should be a function of what it’s worth.

Some people say, for example, that a CAC of 2.0 is bad.  Well, if you’re selling a month-to-month product where most customers discontinue by month 9, then a CAC of 2.0 is horrific.  However, if you’re selling sticky enterprise infrastructure, replacing systems that have been in place for a decade with applications that might well be in place for another decade, then a CAC is 2.0 is probably fine.  That’s the point:  there is no absolute right or wrong answer to what a company should be willing to pay for a customer.  What you are willing to pay for a customer should be a function of what they are worth.

The CAC ratio captures the cost of acquiring customers.  In plain English, the CAC ratio is the multiple you are willing to pay for $1 for annual recurring revenue (ARR).  With a CAC ratio of 1.5, you are paying $1.50 for a $1 of ARR, implying an 18 month payback period on a revenue basis and 18-months divided by subscription-GM on a gross margin basis.

Lifetime value (LTV) attempts to calculate what a customer is worth and is typically calculated using gross margin (the profit from a customer after paying the cost of operating the service) as opposed to simply revenue.  LTV is calculated first by inverting the annual churn rate (to get the average customer lifetime in years) and then multiplying by subscription-GM.

For example, with a churn rate is 10%, subscription GM of 75%, and a CAC ratio of 1.5, the LTV/CAC ratio is (1/10%) * 0.75 / 1.5 = 5.0.

The general rule of thumb is that LTV/CAC should be 3.0 or higher, with of course, the higher the better.

There are three limitations I am aware of in working with LTV/CAC as a metric.

  • Churn rate.  Picking the right churn rate isn’t easy and is made complicated in the presence of a mix of single- and multi-year deals.  All in, I think simple churn is the best rate to use as it reflects the “auto-renewal” of multi-year deals as well as the very real negative churn generated by upsell.
  • Statistics and distributions.  I’m not a hardcore stats geek, but I secretly worry that many different distributions can produce an average of 10%, and thus inverting a 10% churn rate to produce an average 10-year customer lifetime scares me a bit.  It’s the standard way to do things, but I do worry late at night that averages can be misleading.
  • Light from a distant star.  Remember that today’s churn rate is a function of yesterday’s deals.  The more you change who you sell to and how, the less reflective yesterday’s churn is of tomorrow’s.  It’s like light arriving from a star that’s three light-years away:  what you see today happened three years ago.  To the extent that LTV is a forward-looking metric, beware that it’s based on churn which is backward-looking.  In perfect world, you’d use predicted-churn in an LTV calculation but since calculating that would be difficult and controversial, we take the next best thing:  past churn.  But remember that the future doesn’t always look like the past.

 

You Can’t Analyze Churn by Analyzing Churn

One thing that amazes me is when I hear people talk about how they analyze churn in a cloud, software as a service (SaaS), or other recurring revenue business.

You hear things like:

  • “17% of our churn comes from emerging small business (ESB) segment, which is normal because small businesses are inherently unstable.”
  • “22% of our churn comes from companies in the $1B+ revenue range, indicating that we may have a problem meeting enterprise needs.”
  • “40% of the customers in the residential mortgage business churned, indicating there is something wrong our product for that vertical.”

There are three fallacies at work here.

The first is assumed causes.  If you that 17% of your churn comes from the ESB segment, you know one and only one thing:  that 17% of your churn comes from the ESB segment.  Asserting small business stability as the cause is pure speculation.  Maybe they did go out of business or get bought.  Or maybe they didn’t like your product.  Or maybe they did like your product, but decided it was overkill for their needs.  If you want to how much of your churn came from a given segment, ask a finance person.  If you want to know why a customer churned, ask them.  Companies with relatively small customer bases can do it via a phone.  Customers with big bases can use an online survey.  It’s not hard.  Use metrics to figure out where your churn comes from.  Use surveys to figure out why.

The second is not looking at propensities and the broader customer base. If I said that 22% of your annual recurring revenue (ARR) comes from $1B+ companies, then you shouldn’t be surprised that 22% of your churn comes from them as well.  If I said that 50% of your ARR comes from $1B+ companies (and they were your core target market), then you’d be thrilled that only 22% of your churn comes from them.  The point isn’t how much of your churn comes from a given segment:  it’s how much of your churn comes from a given segment relative to how much of your overall business comes from that segment.  Put differently, what is the propensity of someone to churn in one segment versus another.

And you can’t perform that analysis without getting a full data set — of both customers who did churn and customers who didn’t.  That’s why I say you can’t analyze churn by analyzing churn.  Too many people, when tasked with churn analysis:  say, “quick, get me a list of all the customers who churned in the past 6 months and we’ll look for patterns.”   At that instant you are doomed.  All you can do is decompose churn into buckets, but know nothing of propensities.

For example, if you noticed that in one country a surprising 89% of churn came from customers with blue eyes, you might be prompted to launch an immediate inquiry into how your product UI somehow fails for blue-eyed customers.  Unless, of course, the country was Estonia where 89% of the population has blue eyes and ergo 89% of your customers do.  Bucketing churn buys you nothing without knowing propensities.

The last is correlation vs. causation.  Knowing that a large percentage of customers in the residential mortgage segment churned (or even have higher propensity to churn) doesn’t tell you why they are churning.  Perhaps your product does lack functionality that is important in that segment.  Or perhaps it’s 2008, the real estate crisis is in full bloom, and those customers aren’t buying anything from anybody.  The root cause is the mortgage crisis, not your product.   Yes, there is a high correlation between customers in that vertical and their churn rate.  But the cause isn’t a poor product fit for that vertical, it’s that the vertical itself is imploding.

A better, and more fun, example comes from The Halo Effect, which tells the story that a famous statistician once showed a precise correlation between the increase in the number of Baptist preachers and the increase in arrests for public drunkenness during the 19th Century.  Do we assume that one caused the other?  No.  In fact, the underlying driver was the general increase in the population — with which both were correlated.

So, remember these two things before starting your next churn analysis

  • If you want to know why someone churned, ask them.
  • If you want to analyze churn, don’t just look at who churned — compare who churned to who didn’t

Why, as CEO, I Love Driver-Based Planning

While driver-based planning is a bit of an old buzzword (the first two Google hits date to 2009 and 2011 respectively), I am nevertheless a huge fan of driver-based planning not because the concept was sexy back in the day, but because it’s incredibly useful.  In this post, I’ll explain why.

When I talk to finance people, I tend to see two different definitions of driver-based planning:

  • Heavy in detail, one where you build a pretty complete bottom-up budget for an organization and play around with certain drivers, typically with a strong bias towards what they have historically been.  I would call this driver-based budgeting.
  • Light in detail where you struggle to find the minimum set of key drivers around which you can pretty accurately model the business and where drivers tend to be figures you can benchmark in the industry.  I call this driver-based modeling.

While driver-based budgeting can be an important step in building an operating plan, I am actually bigger fan of driver-based modeling.  Budgets are very important, no doubt.  We need them to run plan our business, align our team, hold ourselves accountable for spending, drive compensation, and make our targets for the year.  Yes, a good CEO cares about that as a sine qua non.

But a great CEO is really all about two things:

  • Financial outcomes (and how they create shareholder value)
  • The future (and not just next year, but the next few)

The ultimate purpose of driver-based models is to be able answer questions like what happens to key financial outcomes like revenue growth, operating margins, and cashflow given set of driver values.

I believe some CEOs are disappointed with driver-based planning because their finance team have been showing them driver-based budgets when they should have been showing them driver-based models.

The fun part of driver-based modeling is trying to figure out the minimum set of drivers you need to successfully build a complete P&L for a business.  As a concrete example I can build a complete, useful model of a SaaS software company off the following minimum set of drivers

  • Number and type of salesreps
  • Quota/productivity for each type
  • Hiring plans for each type
  • Deal bookings mix for each (e.g., duration, prepayments, services)
  • Intra-quarter bookings linearity
  • Services margins
  • Subscription margins
  • Sales employee types and ratios (e.g., 1 SE per 2 salesreps)
  • Marketing as % of sales or via a set of funnel conversion assumptions (e.g., responses, MQLs, oppties, win rate, ASP)
  • R&D as % of sales
  • G&A as % of sales
  • Renewal rate
  • AR and AP terms

With just those drivers, I believe I can model almost any SaaS company.  In fact, without the more detailed assumptions (rep types, marketing funnel), I can pretty accurately model most.

Finance types sometimes forget that the point of driver-based modeling is not to build a budget, so it doesn’t have to be perfect.  In fact, the more perfect you make it, the heavier and more complex it gets.  For example, intra-quarter bookings linearity (i.e., % of quarterly bookings by month) makes a model more accurate in terms of cash collections and monthly cash balances, but it also makes it heavier and more complex.

Like each link in Marley’s chains, each driver adds to the weight of the model, making it less suited to its ultimate purpose.  Thus, with the additional of each driver, you need to ask yourself — for the purposes of this model, does it add value?  If not, throw it out.

One of the most useful models I ever built assumed that all orders came in on the last day of quarter.  That made building the model much simpler and any sales before the last day of the quarter — of which we hope there are many — become upside to the conservative model.

Often you don’t know in advance how much impact a given driver will make.  For example, sticking with intra-quarter bookings linearity, it doesn’t actually change much when you’re looking at quarter granularity a few years out.  However, if your company has a low cash balance and you need to model months, then you should probably keep it in.  If not, throw it out.

This process makes model-building highly iterative.  Because the quest is not to build the most accurate model but the simplest, you should start out with a broad set of drivers, build the model, and then play with it.  If the financial outcomes with which you’re concerned (and it’s always a good idea to check with the CEO on which these are — you can be surprised) are relatively insensitive to a given driver, throw it out.

Finance people often hate this both because they tend to have “precision DNA” which runs counter to simplicity, and because they have to first write and then discard pieces of their model, which feels wasteful.  But if you remember the point — to find the minimum set of drivers that matter and to build the simplest possible model to show how those key drivers affect financial outcomes — then you should discard pieces of the model with joy, not regret.

The best driver-based models end up with drivers that are easily benchmarked in the industry.  Thus, the exercise becomes:  if we can converge to a value of X on industry benchmark Y over the next 3 years, what will it do to growth and margins?  And then you need to think about how realistic converging to X is — what about your specific business means you should converge to a value above or below the benchmark?

At Host Analytics we do a lot of driver-based modeling and planning internally.  I can say it helps me enormously as CEO think about industry benchmarks, future scenarios, and how we create value for the shareholders.  In fact, all my models don’t stop at P&L, they go onto implied valuation given growth/profit and ultimately calculate a range of share prices on the bottom line.

The other reason I love driver-based planning is more subtle.  Much as number theory helps you understand the guts of numbers in mathematics, so does driver-based modeling help you understand the guts of your business — which levers really matter, and how much.

And that knowledge is invaluable.