Let us focus on assessing what changes and why, and use that information for decision making.
Imagine that you find your three-year-old with a proud grin on her face, wielding a marker pen. Her baby sister sits near by… her face carefully decorated, forehead to chin, with permanent ink! You think: What happened? Why? And then perhaps a few seconds later: What have I learned? What am I going to do?
What happened and why are exactly the questions we need to focus on when we are managing MSD programmes. Then we must determine what we have learned and how we will adapt our programme in response. No need for academic debates around attribution versus contribution.
“Can’t we just say that we contributed, attribution is really hard to do?”
This is the wrong question. What you really want to know is: Did you make a difference? If so, how? Sometimes that is pretty easy to figure out. Sometimes it takes a bit more effort.
It can be much more challenging than this. There can be so many direct and indirect factors that, together, caused a change. But if we stop investigating to understand why and how these changes occurred, we haven’t learned, and we don’t know what to do next. So we need to carry out our contribution analyses to understand causality. Contribution analysis helps us to structurally unpack cause and effect by gathering and analysing evidence that both supports and challenges a theory of change.
“If we can’t prove attribution, we can’t report anything.”
This is more than a missed opportunity. It is not taking into account the programme’s purpose. That means the push to report attributable impact numbers negatively affects some programmes. That hurts.
Let’s be realistic - systems are complex. Influencing them is an art in itself. Assessing system changes is often not straightforward. And investigating our contribution requires a culture of honest enquiry.
Let’s be pragmatic too. The guidance paper, A Pragmatic Approach to Assessing System Change, builds on what programmes are already doing to assess and understand system change. It provides guidance on how to sensibly assess system changes and how to use that information to adapt strategies and interventions.
It’s not about whether we can attribute changes to our interventions, or whether it’s better to assess our contribution to changes. These are not two different animals. They are part of a continuum. We should aim to assess causality.
But how does a programme report impact that is sometimes attributable and sometimes contributed to?
We need to have a discussion on what and how to report, given the different layers of change that programmes aim to influence.
We are curious to hear your suggestions on how donors and programmes should deal with the numbers issue. The floor is all yours.