I’m frequently getting requests for good Agile Metrics and I’m never quite sure how to respond. Courtesy of some time waiting at LGA, I’ve been giving this some more thought. For many organizations metrics are irrelevant, don’t bother collecting them as they will just waste your time (and money).
If you must collect metrics, here is what I would consider.
Running Tested Purchased Features – Ron Jeffries is famous the metric Running Tested Features (RTF). I suggest that you consider taking it one step further until you’ve sold the feature to the customer you don’t know if they value it or not. For most product organizations this is a bit of stretch to measure in which case stick with Ron’s advice.
Questions to ask:
- What would you like to change?
- If you had the information the metric provided what action would you take? Can you take that action now without the proof of the metric?
- Your key measure (i.e. RTF), should measure your widest span of control – Sold Features > Deployed > Automated Acceptance Tests > …
- Measure Cycle Time – i.e. how long it takes to get something done and not people.
- Other measures can be used: i.e. test coverage from Unit Tests, but be careful they might measure what you think they mean.
- Measures can be gamed/fooled (intentionally or otherwise): For example test coverage measures whether or not a line of code has been visited. It doesn’t measure if there its meaningfully tested. If you must use a measure like this pay attention to the trend and not the absolute number. In this case a large jump might indicate someone having written a not very useful test of the outside api of your application.
- Metrics have best before dates. Eventually you will stop getting real value from them. At that stage throw them away.
- Ask can I get this information by walking around, observing and asking questions.
Alright, you made it this far you deserve some options:
Martin Fowler says: CannotMeasureProductivity. Dave Nicolette presented on this at Agile 2009 (this article links to heaps of others). I wrote this for InfoQ last year: What is a Good Agile Metric?. InfoQ also has: Metrics in an Agile World.
The following tools will help you measure, but please remember they often have many bad measures (comments?) turned with the good ones. Think carefully when choosing your rulesets:
- Sonar – has a bunch of interesting measures: Cyclomatic Complexity, Duplicated code, … . While there are other plugins, its of most use in the Java world.
- JDepend – helps you spot good vs. bad dependencies.
- PMD, FindBugs, JLint – see a comparison of all three (pdf). Some of these tools check some pointless things: method name too short or too long? missing Javadoc comments? Please configure these with the help of a grown adult. But they can also be configured to spot methods (> 30-40 lines) and classes (>300-400 lines) that are too long.
- NDepend like JDepend and heaps more measures. Again please be careful configure only with an adults help 🙂 Caveat Emptor I’ve been given a free copy of NDepend (that I’ve never had a chance to use).
When paying attention to measures of the code, what matters is the trend and not the absolute numbers. Finally just because a tool can measure it doesn’t mean its worth measuring, conversely some of the best measures don’t have tools to measure them. In this case note that none of the above tools measure cycle time.
Updated to make clear the point that you shouldn’t measure people and the limitations of tools.
Mark Levison has been helping Scrum teams and organizations with Agile, Scrum and Kanban style approaches since 2001. From certified scrum master training to custom Agile courses, he has helped well over 8,000 individuals, earning him respect and top rated reviews as one of the pioneers within the industry, as well as a raft of certifications from the ScrumAlliance. Mark has been a speaker at various Agile Conferences for more than 20 years, and is a published Scrum author with eBooks as well as articles on InfoQ.com, ScrumAlliance.org an AgileAlliance.org.
Bob Marshall says
I approve of your modification of Ron’s origin metric – just one word of caution, though: “Purchased” often tells you little more about the value being earned by a feature than does “Tested”. Better to consider “in production” / “in use”, or even better still, take the Systems Thinking (Synergistic) perspective and consider to what extent the feature in use with a given customer actually elevates their constraint and increases the throughput of their “system”.
Cheers
– Bob
Mark Levison says
Bob – I agree with the point, although even deployed might not be meaningful since the users might not use it. In either case I think most of the value can be found by measuring just beyond your current span of control.
Oli says
Your URL to Martin Fowler’s article is missing /bliki/ in it:
https://www.martinfowler.com/bliki/CannotMeasureProductivity.html
Mark Levison says
Thank’s I’ve fixed it.
Mark Levison says
From a discussion on LinkedIn:
At the team level I would focus on things they control:
< xx lines per method,
Nigel Runnels-Moss says
Martin Fowler was right: it *is* impossible to measure output, and hence productivity. I discovered the reason why when doing research for Symbian on software quality metrics: Kolmogorov-Chaitin-Solomonoff complexity, an analogue to Gödel’s incompleteness theorem that tells us that it is fundamentally impossible to measure the complexity (and hence size) of any computer program over a certain trivial size (see J.P. Lewis 2001).
This little known academic work helps explain why so many software projects fail, and why we find it so hard to estimate software complexity, size, cost etc.
Even Tom DeMarco, who started the whole ‘you cannot control what you cannot measure’ malarkey now admits he was wrong, see https://bit.ly/pRrkd