Category Archives: Entrepreneurship

A Look at the Tintri S-1

Every now and then I take a dive into an S-1 to see what clears the current, ever-changing bar for going public.  After a somewhat rocky IPO process, Tintri went public June 30 after cutting the IPO offering price and has traded flat thus far since then.

Let’s read an excerpt from this Business Insider story before taking a look at the numbers.

Before going public, Tintri had raised $260 million from venture investors and was valued at $800 million.

With the performance of this IPO, the company is now valued at about about $231 million, based on $7.50 a share and its roughly 31 million outstanding shares, (if the IPO’s bankers don’t buy their optional, additional roughly 1.3 million shares.)

In other words, this IPO killed a good $570 million of the company’s value.

In other words, Tintri looks like a “down-round IPO” (or an “IPO of last resort“) — something that frankly almost never happened before the recent mid/late stage private valuation bubble of the past 4 years.

Let’s look at some numbers.

tintri p+l

Of note:

  • $125M in FY2017 revenue.  (They have scale, but this is not a SaaS company so the revenue is mostly non-recurring, making it easier to get to grow quickly and making the revenue is worth less because only the support/maintenance component of it renews each year.)
  • 45% YoY total revenue growth.  (On the low side, especially given that they have a traditional license/maintenance model and recognize revenue on shipment.)
  • 65% gross margins  (Low, but they do seem to sell flash memory hardware as part of their storage solutions.)
  • 87% of revenue spent on S&M (High, again particularly for a non-SaaS company.)
  • 43% of revenue spent on R&D  (High, but usually seen as a good thing if you view the R&D money as well spent.)
  • -81% operating margins (Low, particularly for a non-SaaS company.)
  • -$70.4M in cashflow from operating activities in 2017 ($17M average quarterly cash burn from operations)
  • Incremental S&M / incremental product revenue = 73%, so they’re buying $1 worth of incremental (YoY) revenue for an incremental 73 cents in S&M.  Expensive but better than some.

Overall, my impression is of an on-premises (and to a lesser extent, hardware) company in SaaS clothing — i.e., Tintri’s metrics look like a SaaS company, but they aren’t so they should look better.  SaaS company metrics typically look worse than traditional software companies for two reasons:  (1) revenue growth is depressed by the need to amortize revenue over the course of the subscription and (2) subscriptions companies are willing to spend more on S&M to acquire a customer because of the recurring nature of a subscription.

Concretely, if you compare two 100-unit customers, the SaaS customer is worth twice the license/maintenance customer over 5 years.

saas compare

Moreover, even if Tintri were a SaaS company, it is quite out of compliance with the Rule of 40, that says growth rate + operating margin >= 40%.  In Tintri’s case, we get -35%, 45% growth plus -81% operating margin, so they’re 75 points off the rule.

Other Notes

  • 1250+ customers
  • 21 of the Fortune 100
  • 527 employees as of 1/31/17
  • CEO 2017 cash compensation $525K
  • CFO 2017 cash compensation $330K
  • Issued special retention stock grants in May 2017 that vest in the two years following an IPO
  • Did option repricing in May 2017 to $2.28/share down from weighted average exercise price of $4.05.
  • $260M in capital raised prior to IPO
  • Loans to CFO and CEO to exercise stock options at 1.6% to 1.9% interest in 2013
  • NEA 22.7% ownership prior to opening
  • Lightspeed 14.5% ownership
  • Insight Venture Partners 20.2% ownership
  • Silver Lake 20.4% ownership
  • CEO 3.8% ownership
  • CFO 0.7% ownership
  • $48.9M in long-term debt
  • $13.8M in 2017 stock-based compensation expense

Overall, and see my disclaimers, but this is one that I’ll be passing on.

 

The Strategy Compiler: How To Avoid the “Great” Strategy You Couldn’t Execute

Few phrases bother me more than this one:

“I know it didn’t work, but it was a great strategy.  We just didn’t have the resources to execute it.”

Huh.  Wait minute.  If you didn’t have the resources to execute it, then it wasn’t a great strategy.  Maybe it was a great strategy for some other company that could have applied the appropriate resources.  But it wasn’t a great strategy for you.  Ergo, it wasn’t a great strategy.  QED.

I learned my favorite definition of strategy at a Stanford executive program I attended a few years back.  Per Professor Robert Burgelman, author of Strategy is Destiny, strategy is simply “the plan to win.”  Which begets an important conversation about the definition of winning.  In my experience, defining winning is more important than making the plan, because if everyone is focused on taking different hills, any resultant strategy will be a mishmash of plans to support different objectives.

But, regardless of your company’s definition of winning, I can say that any strategy you can’t execute definitionally won’t succeed and is ergo a bad strategy.

It sounds obvious, but nevertheless a lot of companies fall into this trap.  Why?

  • A lack of focus.
  • A failure to “compile” strategy before executing it.

Focus:  Think Small to Grow Big

Big companies that compete in lots of broad markets almost invariably didn’t start out that way.

BusinessObjects started out focused on the Oracle financials installed base.  Facebook started out on Harvard students, then Ivy league students.  Amazon, it’s almost hard to remember at this point, started out in books.  Salesforce started out in SMB salesforce automation.  ServiceNow on IT ticket management.  This list goes on and on.

Despite the evidence and despite the fame Geoffrey Moore earned with Crossing the Chasm, focus just doesn’t come naturally to people.  The “if I could get 1% share of a $10B market, I’d be a $100M company” thought pattern is just far too common. (And investors often accidentally reinforce this.)

The fact is you will be more dominant, harder to dislodge, and probably more profitable if, as a $100M company, you control 30% of a $300M target as opposed to 1% of a $10B target.

So the first reason startups make strategies they can’t execute is because they forget to focus.  They aim too broadly. They sign up for too much.  The forget that strategy should be sequence of actions over time.  Let’s start with Harvard. Then go Ivy League.  Then go Universities in general.  Then go everyone.

Former big company executives often compound the problem.  They’re not used to working with scarce resources and are more accustomed to making “laundry list” strategies that check all the boxes than making focused strategies that achieve victory step by step.

A Failure to Compile Strategy Before Execution

The second reason companies make strategies they can’t execute is that they forget a critical step in the planning process that I call the strategy compiler.  Here’s what I think a good strategic planning process looks like.

  • Strategy offsite. The executive team spends a week offsite focused on situation assessment and strategy.  The output of this meeting should be (1) a list of strategic goals for the company for the following year and (2) a high-level financial model that concretizes what the team is trying to accomplish over the next three years.  (With an eye, at a startup, towards cash.)
  • First round budgeting. Finance issues top-down financial targets.  Executives who own the various objectives make strategic plans for how to attain them.  The output of this phase is (1) first-draft consolidated financials, (2) a set of written strategies along with proposed organizational structures and budgets for attaining each of the company’s ten strategic objectives.
  • Strategy compilation, resources. The team meets for a day to review the consolidated plans and financials. Invariably there are too many objectives, too much operating expense, and too many new hires. The right answer here is to start cutting strategic goals.  The wrong answer is to keep the original set of goals and slash the budget 20% across the board.  It’s better to do 100% of 8 strategic initiatives than do 80% of 10.
  • Strategy compilation, skills. The more subtle assessment that must happen is a sanity check on skills and talent.  Do your organization have the competencies and do your people have the skills to execute the strategic plans?  If a new engineering project requires the skills of 5 founder-level, Stanford computer science PHDs who each would want 5% of a company, you are simply not going to be able to hire that kind of talent as regular employees. (This is one reason companies do “acquihires”).  The output of this phase is a presumably-reduced set of strategic goals.
  • Second round budgeting. Executives to build new or revised plans to support the now-reduced set of strategic goals.
  • Strategy compilation. You run the strategy compiler again on the revised plan — and iterate until the strategic goals match the resources and the skills of the proposed organization.
  • Board socialization. As you start converging via the strategy compiler you need to start working with the board to socialize and eventually sell the proposed operating plan.  (This process could easily be the subject of another post.)

If you view strategy as the plan to win, then successful strategies include only those strategies that your organization can realistically execute from both a resources and skills perspective.  Instead of doing a single-pass process that moves from strategic objectives to budgets, use an iterative approach with a strategy compiler to ensure your strategic code compiles before you try to execute it.

If you do this, you’ll increase your odds of success and decrease the odds ending up in the crowded section of the corporate graveyard where the epitaphs all read:

Here Lies a Company that Had a “Great” Strategy  It Had No Chance of Executing

Blocking the End Run: Eleven Words to Reduce Politics in Your Organization

People are people.  Sometimes they’re conflict averse and just not comfortable saying certain things to their peers.  Sometimes they don’t like them and are actively trying to undermine them. Sometimes they’re in a completely functional relationship, but have been too darn busy to talk.

So when this happens, how do you — as a manager — respond?  What should you do?

“Hey Dave, I wanted to say that Sarah’s folks really messed up on the Acme call this morning.  They weren’t ready with the proposal and were completely not in line with my sales team.”

Do you pile on?

“Again?  Sarah’s folks are out of control, I’m going to go blast her.”  (The “Young Dave” response.)

Do you investigate?

“You know my friend Marcy always said there are three sides to every story:  yours, mine, and what actually happened.  So let me give Sarah a call and look into this.”

Do you defend?

“Well, that doesn’t sound like Sarah.  Her team’s usually buttoned up.”

In the first case, you’re going off half-cocked without sufficient information which, while emotionally satisfying in the short-term, often leads to a mess followed by several apologies in the mid-term.  In the second case, you’re being manipulated into investigating something when perhaps you were planning a better use of your time that day.  In the third case, you’re going off half-cocked again, but in the other direction.

In all three cases, you’re getting sucked into politics.  Politics?  Is it really politics?  Well, how do you think Sarah is going to feel in when you show up asking a dozen questions about the Acme call?  She’ll certainly consider it politics and, among other things, there’s about a 98% chance that she will say:

“Gosh, I wish Bill came and talked to me first.”

At which point, if you’re like me, you’re going to say:

“No, no, no.  I know what you’re thinking.  Don’t worry, this isn’t political.  It’s not like Bill was avoiding you on this one.  He just happened to be talking to me about another issue and he brought this up at the end.  It’s not political, no.”

But can you be sure?  Maybe it just did pop into Bill’s mind during the last minute of the other call.  Or maybe it didn’t.  Maybe the reason Bill called you was a masterfully political pretext.  Can you know the difference?

So what do you say to Bill when he drops the comment about Sarah’s team into your call?  The eleven words that reduce politics in any organization:

“What did Sarah say when you talked to her about this?”

[Mike Drop.]

# # #

(Props to Martin Cooke for teaching me the eleven words.)

Why has Standalone Cloud BI been such a Tough Slog?

I remember when I left Business Objects back in 2004 that it was early days in the cloud.  We were using Salesforce internally (and one of their larger customers at the time) so I was familiar with and a proponent of cloud-based applications, but never felt great about BI in the cloud.  Despite that, Business Objects and others were aggressively ramping on-demand offerings all of which amounted to pretty much nothing a few years later.

Startups were launched, too.  Specifically, I remember:

  • Birst, née Success Metrics, and founded in 2004 by Siebel BI veterans Brad Peters and Paul Staelin, which was originally supposed to be vertical industry analytic applications.
  • LucidEra, founded in 2005 by Salesforce and Siebel veteran Ken Rudin (et alia) whose original mission was to be to BI what Salesforce was to CRM.
  • PivotLink, which did their series A in 2007 (but was founded in 1998), positioned as on-demand BI and later moved into more vertically focused apps in retail.
  • GoodData, founded in 2007 by serial entrepreneur Roman Stanek, which early on focused on SaaS embedded BI and later moved to more of a high-end enterprise positioning.

These were great people — Brad, Ken, Roman, and others were brilliant, well educated veterans who knew the software business and their market space.

These were great investors — names like Andreessen Horowitz, Benchmark, Emergence, Matrix, Sequoia, StarVest, and Tenaya invested over $300M in those four companies alone.

This was theoretically a great, straightforward cloud-transformation play of a $10B+ market, a la Siebel to Salesforce.

But of the four companies named above only GoodData is doing well and still in the fight (with a high-end enterprise platform strategy that bears little resemblance to a straight cloud transformation play) and the three others all came to uneventful exits:

So, what the hell happened?

Meantime, recall that Tableau, founded in 2003, and armed in its early years with a measly $15M in venture capital, and with an exclusively on-premises business model, literally blew by all the cloud BI vendors, going public in May 2013 and despite the stock being cut by more than half since its July 2015 peak is still worth $4.2B today.

I can’t claim to have the definitive answer to the question I’ve posed in the title.  In the early days I thought it was related to technical issues like trust/security, trust/scale, and the complexities of cloud-based data integration.  But those aren’t issues today.  For a while back in the day I thought maybe the cloud was great for applications, but perhaps not for platforms or infrastructure.  While SaaS was the first cloud category to take off, we’ve obviously seen enormous success with both platforms (PaaS) and infrastructure (IaaS) in the cloud, so that can’t be it.

While some analysts lump EPM under BI, cloud-based EPM has not had similar troubles.  At Host, and our top competitors, we have never struggled with focus or positioning and we are all basically running slightly different variations on the standard cloud transformation play.  I’ve always believed that lumping EPM under BI is a mistake because while they use similar technologies, they are sold to different buyers (IT vs. finance) and the value proposition is totally different (tool vs. application).  While there’s plenty of technology in EPM, it is an applications play — you can’t sell it or implement it without domain knowledge in finance, sales, marketing or whatever domain for which you’re building the planning system.  So I’m not troubled to explain why cloud EPM hasn’t been a slog while cloud BI absolutely has been.

My latest belief is that the business model wasn’t the problem in BI.  The technology was.  Cloud transformation plays are all about business model transformation.  On-premises applications business models were badly broken:  the software cost $10s of millions to buy and $10s of millions more to implement (for large customers).  SMBs were often locked out of the market because they couldn’t afford the ante.  ERP and CRM were exposed because of this and the market wanted and needed a business model transformation.

With BI, I believe, the business model just wasn’t the problem.  By comparison to ERP and CRM, it was fraction of the cost to buy and implement.  A modest BusinessObjects license might have cost $150K and less than that to implement.  That problem was not that BI business model was broken, it was that the technology never delivered on the democratization promise that it made.  Despite shouting “BI for the masses” in 1995, BI never really made it beyond the analyst’s desk.

Just as RDBMS themselves failed to deliver information democracy with SQL (which, believe it or not, was part of the original pitch — end users could write SQL to answer their own queries!), BI tools — while they helped enable analysts — largely failed to help Joe User.  They weren’t easy enough to use.  They lacked information discovery.  They lacked, importantly, easy-yet-powerful visualization.

That’s why Tableau, and to a lesser extent Qlik, prospered while the cloud BI vendors struggled.  (It’s also why I find it profoundly ironic that Tableau is now in a massive rush to “go cloud” today.)  It’s also one reason why the world now needs companies like Alation — the information democracy brought by Tableau has turned into information anarchy and companies like Alation help rein that back in (see disclaimers).

So, I think that cloud BI proved to be such a slog because the cloud BI vendors solved the wrong problem. They fixed a business model that wasn’t fundamentally broken, all while missing the ease of use, data discovery, and visualization power that both required the horsepower of on-premises software and solved the real problems the users faced.

I suspect it’s simply another great, if simple, lesson is solving your customer’s problem.

Feel free to weigh in on this one as I know we have a lot of BI experts in the readership.

Do You Want to be Judged on Intentions or Results?

It was early in my career, maybe 8 years in, and I was director of product marketing at a startup.  One day, my peer, the directof of marketing programs hit me with this in an ops review meeting:

You want to be judged on intentions, not results.

I recall being dumbfounded at the time.  Holy cow, I thought.  Is he right?  Am I standing up arguing about mitigating factors and how things might have been when all the other people in the room were thinking only about black-and-white results?

It was one of those rare phrases that really stuck with me because, among other reasons, he was so right.  I wasn’t debating whether things happened or not.  I wasn’t making excuses or being defensive.  But I was very much judging our performance in the theoretical, hermetically sealed context of what might have been.

Kind of like sales saying a deal slipped instead of did not close.   Or marketing saying we got all the MQLs but didn’t get the requisite pipeline.  Or alliances saying that we signed up the 4 new partners, but didn’t get the new opportunities that were supposed to come with them.

Which phrase of the following sentence matters more — the first part or the second?

We did what we were supposed to, but it didn’t have the desired effect.

We would have gotten the 30 MQLS from the event if it hadn’t snowed in Boston.  But who decided to tempt fate by doing a live event in Boston in February?  People who want to be judged on intentions think about the snowstorm; people who want to be judged on results think about the MQLs.

People who want to judged on intentions build in what they see as “reasons” (which others typically see as “excuses”) for results not being achieved.

I’m six months late hiring the PR manager, but that’s because it’s hard to find great PR people right now.  (And you don’t want me to hire a bad one, do you?)

No, I don’t want you to hire a bad one.  I want you to hire a great one and I wanted you to hire them 6 months ago.  Do you think every other PR manager search in the valley took 6 months more than plan?  I don’t.

Fine lines exist here, no doubt.  Sometimes reasons are reasons and sometimes they are actually excuses.  The question isn’t about any one case.  It’s about, deep down, are you judging yourself by intentions or results?

You’d be surprised how many otherwise very solid people get this one thing wrong — and end up career-limited as a result.