Five lessons from my first year at DataStrategies4Change

Author

Dan Wilson

Published

March 24, 2026

Not long after the Brexit referendum in 2016, Dan left Sheffield for Freiburg. After several years studying and working at the University of Freiburg, he joined DataStrategies4Change in April 2025. This is a reflection on his first year working with data in the real world:

I spent years working on neural networks in academia. The most impactful thing I’ve built since? A decision tree.

I cut my teeth in machine-learning and AI in academia, designing and tuning complex neural networks - the kind of work where success means squeezing out a few extra percentage points on benchmark datasets1.

I’ve always enjoyed solving technical problems - something that took me from physics into neuroscience, and later into computer science research. Then around a year ago, I joined DataStrategies4Change as a Data Scientist & Analyst.

I didn’t expect NGOs to be using deep learning — they lack the resources and expertise. And even classical machine learning comes with real concerns around bias and misuse. So it’s understandable that many organisations are cautious.

But I went in assuming that at least the basics would already be in place — classical machine learning, statistical modelling and so on. In reality, resources are stretched far thinner than I expected, and even these methods are out of reach for most NGOs.

But the bigger surprise wasn’t what was missing. It was what actually mattered instead.

Impact hasn’t come from next-gen models. It came from making data usable, making results understandable, and building systems people could actually trust.

1: A decision tree can be revolutionary

Academia obsesses over the latest model; NGOs don’t have that luxury. A decades-old algorithm like the decision tree may be unimpressive in one - and invaluable in the other.

If an organisation is analysing donors using static, hand-written rules from years ago, a decision tree - a model that learns a set of yes/no questions from real data, like an automated flowchart - can make a real difference. Not because it’s complex, but because it’s data-driven.

It looks at all the data. It learns from real outcomes. It updates after every campaign.

In one project we worked on, our models sent nearly 20% fewer letters - while increasing net surplus by 6.3% (read more about this here). During my year at DS4C, I was able to improve this model further, helping this NGO which has been running all its mailing selections with our algorithms for several years now.

It’s the 80:20 rule2 in action: a relatively simple model that delivers most of the impact. Pushing beyond that - towards marginal gains - can require disproportionate effort.

And that was my biggest mindset shift: that we can make the biggest difference, not by building an incredibly complex or powerful model, but by building something that’s better than what came before - and works within the systems and resources available.

2: Your data is more valuable than you think

One organisation we worked with had email addresses for every contact — but didn’t see them as useful data.

A bit of string parsing later, we had useful patterns: like whether someone’s email came from a university or a company. Nothing fancy. But it improved the model.

This happens all the time. NGOs often don’t track engagement systematically, store useful data inconsistently, or only measure half the outcome — for example donations or donors, but not both together. Sometimes they don’t even know how much has been pledged to them in recurring donations for the coming year..

It’s not a capability problem. It’s a data culture problem.

There’s usually far more signal in their systems than anyone realises.

3: Trust nothing in the database

Real world data can be… messy in very human ways. Not because people are careless - but because systems evolve over time, teams change, and documentation rarely keeps up.

This leads to:

  • Silent failures
  • Data leakage
  • Definitions that changed over time without documentation
  • Bits of data that nobody can explain

The only way through is talking to the people behind the data - the ones who actually know what it means.

4: They want to see inside the box

I expected explainability to matter. Just not like this.

I thought it would be about and model refinement.

And it is — to a point. But more than that, it’s about trust. Because refining a model is pointless if people don’t trust it — or don’t trust you with their data in the first place.

In practice, it’s simpler than I expected. Organisations want to trace predictions, sense-check outcomes, and align results with domain knowledge. What builds trust isn’t accuracy scores — it’s transparency.

So we almost always use glassbox models: models whose logic you can fully inspect and follow.

And that trust doesn’t stop at the model. Many organisations are — rightly — cautious about AI systems, especially given how often they’re associated with bias, opacity, or misuse. But at the same time, they often rely heavily on external platforms and cloud providers, sometimes storing sensitive data outside their own jurisdiction.

So the caution is there. It’s just not always applied consistently. That’s not a criticism — it reflects how difficult it is to judge where the real risks are. It’s also why we’re deliberate about how we handle data: computation runs on our own infrastructure, under our control, in our office — not on external cloud services.

But it does highlight a broader opportunity: to build systems that are not just effective, but understandable — and to have more grounded conversations about where the real risks and benefits actually lie.

5: People’s intuitions about AI aren’t quite right

What people think is easy vs hard in AI is often flipped.

Finding articles where a politician promotes an idea without naming it — that sounds hard. It isn’t: the right RSS feed, semantic search and a good prompt, and it works well.

On the other hand: “Just compare this legal document to others” sounds simple — until you realise it references dozens of external laws that aren’t included.

Why it matters

I took this role because I enjoy working with data and solving problems - and it’s especially rewarding to apply those skills to causes I genuinely believe in.

These organisations are doing important work with limited resources. Even small improvements in how they use data can free up time, money, and attention for what actually matters.

The NGO sector doesn’t need bleeding-edge AI.

It needs people willing to meet it where it is - and help it take the next step.

Footnotes

  1. Daniel Wilson, Robin T. Schirrmeister, Lukas A. W. Gemein, Tonio Ball; Deep Riemannian Networks for end-to-end EEG decoding. Imaging Neuroscience 2025; 3 imag_a_00511. doi: https://doi.org/10.1162/imag_a_00511↩︎

  2. https://en.wikipedia.org/wiki/Pareto_principle↩︎