The Plague of Entropy
Monday, September 14, 2009
By James Whittaker
Mathematically entropy is a measure of uncertainty. If there are, say, five events then maximum entropy occurs when those five events are equally likely and minimum entropy when one of those events is certain and the other four impossible.
Mathematically entropy is a measure of uncertainty. If there are, say, five events then maximum entropy occurs when those five events are equally likely and minimum entropy when one of those events is certain and the other four impossible.
The more uncertain events you have to consider, the higher measured entropy climbs. People often think of entropy as a measure of randomness: the more (uncertain) events one must consider, the more random the outcome becomes.
Testers introduce entropy into development by adding to the number of things a developer has to do. When developers are writing code, entropy is low. When we submit bugs, we increase entropy. Bugs divert their attention from coding. They must now progress in parallel on creating and fixing features. More bugs means more parallel tasks and raises entropy. This entropy is one reason that bugs foster more bugs ... the entropic principle ensures it. Entropy creates more entropy! Finally there is math to show what is intuitively appealing: that prevention beats a cure.
However, there is nothing we can do to completely prevent the plague of entropy other than create developers who never err. Since this is unlikely any time soon we must recognize how and when we are introducing entropy and do what we can to manage it. The more we can do during development the better. Helping out in code reviews, educating our developers about test plans, user scenarios and execution environments so they can code against them will reduce the number of bugs we have to report. Smoking out bugs early, submitting them in batches and making sure we submit only high quality bugs by triaging them ourselves will keep their mind on development. Writing good bug reports and quickly regressing fixes will keep their attention where it needs to be. In effect, it maximizes the certainty of the 'development event' and minimizes the number and impact of bugs. Entropy thus tends toward it's minimum.
We can't banish this plague but if we can recognize the introduction of entropy into development and understand its inevitable effect on code quality, we can keep it at bay.
Interesting post!
ReplyDeleteI had intended to submit an entropy post, but never finished it...
There are other ways that entropy in the test process can exist:
When we model a system (maybe to develop use cases or other test design purposes) then entropy is introduced.
Our model has "reduced" the system - ie we've thrown away information or created uncertainty.
We can help to alleviate this by other test design techniques such as exploratory testing.
But where we have made assumptions about the system (or really used a model for the test design) then we've increased the amount of entropy in the process.
There's a balance between trying to model the system accurately to achieve higher accuracy in the test design and using the model as a tool for the test design (good enough modelling.)
Not an easy balancing act!
It isn't till you start testing that you can identify some or all the events.
ReplyDeleteGood Post
I think that, ultimately, the vast majority of entropy is introduced by the developers themselves: some of it is due to bugs but a significant part is due to new/changed functionality that cross cuts multiple existing modules, etc.
ReplyDeleteEither way, we can't do much about it. Bugs are an inevitable by product of development, and so is adding functionality. Thus, automated testing is a mechanism for reducing the marginal entropy. A test suite will prevent the new code from inadvertently breaking existing code (increasing the entropy)
Bottom line: Programmers are mass producers of entropy. Testing helps us manage it.
Somewhat related, do you feel that if you discover more bugs early in the process that there is a higher chance that there are more.
ReplyDeleteOr do you feel that if 0 or few bugs are found that the risk is higher?
Another related question, do you find that the number of bugs stay close to an average over time.