Skip to main content

Platform Teams' Economics

After being involved in the Kubernetes project for about a year I have had some interesting reflections about the configuration and infrastructure space.  I have talked to a few Platform engineers and have gained some interesting insights into how these organizations operate at scale.  My observation is that in most organizations have a tough time measuring or financing their Platform Teams. A lot of the Platform teams are not sure how to evaluate whether or not their effectiveness.  This situation paired with the fact that the investment in Platform is generally removed from the business KPIs leads to underfunding in Platform resources.

Most of the Platform projects are evaluated based on activity instead of impact.  In the conversations and allocation of resources the head of Platform Engineering generally can come up with only a backlog of outstanding projects without a clear indicator of why speeding up delivery of those projects will improve the company bottom line.  The usual narratives are:

1. Deploying a new type of infrastructure

2. Helping launch a new application

3. Taking on a set of applications which were acquired through M&A

The connection between those activities and company goals usually stop around uptime and security.  Both of those are important but both of those are in the realm of “power grid” metrics - there is nothing to report until disaster strikes.

Given finite resource and ability for line of business (LOB) applications to have a more clear connection to revenue the Platform investments seem to get pushed down the stack.   I have talked to some Platform teams where half a dozen engineers are helping a multi thousand software development organization.

Platform teams are frequently tasked with creating DevOps infrastructure and often scramble to keep up with the demand.  The unique tooling like Terraform have a low “return on learning” for LOB application engineers and they try to push it off their plate.

One answer that teams come up with us using general purpose programming stacks for infrastructure projects, but that also leads to some hard problems.  End to end testing of a module that instantiates a load balancer for instance requires a non trivial test setup.

What is the way you measure your Platform Team’s success and if you are on one how do you know that you are making an impact?  We all know that number of bugs, lines of code, etc are not the right metrics.  Being business does not equal being important.  DM @mikebz on Twitter or leave a comment here.


Popular posts from this blog

SDET / QA Engineer Interview Checklist

After interviewing and hiring hundreds of engineers over the past 12+  years I have come up with a few checklists.  I wanted to share one of those with you so you could conduct comprehensive interviews of QA Engineers for your team. I use this checklist when I review incoming resumes and during the interview.  It keeps me from missing areas that ensure a good team and technology fit.  I hope you make good use of them.  If you think there are good questions or topics that I have missed - get in touch with me! SDE/T or QA Engineer interview checklist from Mike Borozdin If you like this checklist you might want to check out these posts: Emotional Intelligence in Software Teams   and  Good-bye manual tester, hello crowdsourcing!

Code versus Configuration

At Ethos we are building a distributed mortgage origination system and in mortgage there is a lot of different user types with processes that vary depending on geography.  One of our ongoing discussions is about how much of the logic resides in code vs. being in a workflow system or configuration.  After researching this topic for a bit, I have arrived at a conclusion that the logic should live outside of code very infrequently, which might come as a surprise to a lot of enterprise software engineers. Costs of configuration files and workflow engines First thing that I assume is true is that having any logic outside of the code has costs associated with it.  Debugging highly configurable system involves not only getting the appropriate branch from source control, you also need to make sure that the right configuration values or the database.  In most cases this is harder for programmers to deal with.  In many FinTech companies where the production data is not made readily acce

Intuitive Programming - Comments

Comments are a topic of vibrant discussion.  Ever since programmers could leave some text in the program that was ignored by the machine the debate started: “what’s a good comment, what’s a bad comment, why comment?” There are endless instructions to programmers that say many of the following things: 1) Describe your function! 2) Don’t write in the comment what you wrote in the code. 3) Tell people why you are doing what you are doing. What I think has been missing from this discourse is the audience for comments and through those audiences there is intent.  The code is being read, skimmed or analyzed by people and tools.  So what are the audiences and reading modes? 1) Maintaining and enhancing the code 2) Skimming through the entire module or file to figure out what the overall structure is 3) Reviewing the test files to check out the test coverage and edge cases 4) Seeing the docstrings of functions while being in a separate file altogether 5) Reviewi