Publish What You Fund has undertaken some initial analysis of aid donors’ plans to publish to the IATI component of the agreed common standard for aid information. Here, Mark Brough explains the process they went through to take a series of Excel files, convert them into a format suitable for analysis, and come to some overall conclusions.
The short version: check out the Aid Transparency Tracker.
Why we did this
At the 4th High Level Forum on Aid Effectiveness in Busan, South Korea, in November 2011, the world’s aid donors agreed a “common, open standard” for publishing aid information. Donors also agreed in Busan to publish “implementation schedules” explaining in detail how and when they would meet this commitment.
The implementation schedules are like a forward-looking calendar, explaining when donors plan to publish specific pieces of data, like results, project documents, and conditions, as well as transaction-level spending data. They also explain whether the donor will be publishing under an open license (public domain or attribution-only) and whether they will be republishing every quarter as a minimum frequency of publication – both are required for IATI compliance. These implementation schedules were published on the OECD/DAC’s common standard website, but in several different formats, which required a detailed look at each donor’s schedule, as well as interpreting them when donors have completed them in different ways or understood the various options differently.
Pulling all the schedules into a single application allows us to assess donors’ overall ambition; compare fields across schedules; show the publication of fields over time from different donors; and provide CSV, JSON/JSONP) and iCal feeds for each donor.
The original implementation schedules
The implementation schedules were published in individual Excel files, containing three main sheets: general, agency, and activity-level information.
While there is a template for the common standard implementation schedules, several different templates exist, and donors further added to this complexity by modifying the templates, changing options, and adding and deleting rows – in all, we counted eleven different versions. Importing the schedules proved difficult because of this. In addition, some donors published their Excel-based templates in PDF format, which made it impossible to automatically parse them. In these cases, it was necessary to copy and paste the same data into new spreadsheets. While the data was thoroughly checked, it is possible that some human error will have resulted. Some interpretation was necessary to ensure consistency and comparability across the schedules.
The schedules were automatically parsed and imported into this application. Publish What You Fund staff then checked the resulting data to ensure that it had been parsed correctly and that it made sense.
Openness begets openness
None of this would have been possible without :
- A range of great open-source tools:
- iati-implementationxml (originally converted Excel IATI implementation schedules to XML)
- OKFN’s ReclineJS (for the graphs and timeline; which itself relies on a range of great open source software)
- Bootstrap, JQuery, and JQuery Tablesorter (for the front-end)
- Python, Flask, and a range of libraries (for the back-end)
- Implementation schedules released by donors as part of the common standard, transparently outlining in detail their specific commitments to publish more and better data.
But, we’ve also fed back: iati-implementationxml has been expanded to add compatibility with the common standard formats of implementation schedules; to detect the format of different schedules; and to function as a module that can be imported. We provided a couple of pull request for some small bugs with ReclineJS, and of course, we’re releasing all of our own code as well (although, it could definitely do with some tidying up…).
Now that we’ve looked at what information different donors are committing to publish, there are two main steps for us:
- Doing it again: we want to check the assumptions and interpretations we’ve made with donors to ensure that they’re accurate, as well as encouraging all donors to be more ambitious about the information they’re planning to publish. As part of this, we’ll also aim to iron out the problems in the schedule.
- Looking at publication: now that we have better data on what donors are committing to do, the question arises: what are they actually publishing, and are they meeting their commitments? We’ll be working on that throughout the year and will release this analysis in October. Until then, all of our code is of course being developed as an open-source project on Github.
We’d love your feedback, so please get in touch: