.plan-26-18: From tropical forest protection to oi swallowing its oxcaml tail

Our REDD+ over-crediting paper hits Nature Communications just as Microsoft retreats from removals, we talk responsible evidence synthesis while LLMs appear in UK planning, and oi grows a self-update bootstrap.

1 REDD+ over-crediting in Nature Communications

Our paper on learning lessons from over-crediting in REDD+ projects came out this week in Nature Communications, led by Thomas Swinfield. The reception to the paper has generally been encouragingly positive, especially with the framing that "bad credits are not the same as bad projects".

The timing of the paper hits the carbon market at a particular low time, since it's reeling from Microsoft's abrupt retreat from carbon removals meaning that the single largest buyer of removals is scaling back its commitments from just a few months ago. While this hits the whole stack of direct-air-capture pipelines through to nature-based projects, at least it can hopefully rebuild now with nature and technological removals/avoidance working in harmony rather than bickering about which specific method is 'the best'.

We need them all, and I think the necessary response should be to raise the floor on both at once, and not try to pick a winner. The Cambridge University and Computer Lab writers have done a great job carrying that point through, and I've written up my own full argument for those who want to dig in.

2 The inevitable rise of LLMs in government decision-making

Sadiq Jaffer and Sam Reynolds also gave a great talk at the Digital Statecraft Academy on our evidence synthesis work, on "The inevitable rise of Large Language Models in government decision making". Civil servants and policy folk in the room were asking practical questions about how to do the right thing. There were questions about best practices for reducing hallucinations with responses discussing retrieval grounding, structured outputs, human-in-the-loop checkpoints, and maintaining proper evaluation harnesses.

The inevitable rise of Large Language Models in government decision making (Sadiq and Sam)
The inevitable rise of Large Language Models in government decision making (Sadiq and Sam)

Then there were concerns about the computational and power crunch to keep all of this affordable as adoption scales across government. We discussed the use of smaller specialised models, on-prem inference for sensitive workloads, and the open question of whether the UK has the data-centre capacity to host serious sovereign deployments. The third was on whether quantum computing changes the picture (quick answer: no).

Just as this was all happening, the government announced Google had won a tender for planning-decision automation. English councils are trialling a Google AI tool to speed up planning, which is precisely the kind of black-box deployment my red-pill/blue-pill argument was cautioning against. Decisions affecting people's homes are now being filtered through opaque models with no public scrutiny of the reasoning chain.

%rc
Sadiq and I took a closer look at the tender notice, and spotted that bidders were required to integrate with the incumbent planning systems, which effectively freezes out smaller UK players. Last year about this time, I was looking at how the UK might benefit from an open-data substrate via a national data library. Without a credible open layer, every public-sector AI tender will keep collapsing onto the same handful of incumbent vendors.

If there's a bright spot, it's that the questions from the Statecraft audience suggest civil servants increasingly understand this, and the government itself is dispatching more contracts to UK firms in other sectors. We'll get there with the open AI story...

3 Hacking updates

3.1 Cyrus visits to talk Hazel

The week started with a fun visit from KC Sivaramakrishnan (over for a PaPOC keynote), Cyrus Omar, Andrew Blinn and Matthew Keenan (formerly an undergrad here at Cambridge, now doing a PhD over with Cyrus in UMich).

We all sat down with Ryan Gibb to brainstorm over ideas for how to combine recent advances in Hazel with the OxCaml work going on around here.

%rc
The two most exciting things were the emergence of a Hazel CLI (so we can not only integrate it more easily into MDX workflows but also agentic coding), and also Ryan's package calculus as the basis for a brand new approach to how we express dependencies (in a language that doesn't have any backwards compatibility baggage to worry about). More on this as we convene next at PROPL3 in June!

4 oi gains self-update

I've been making steady progress on oi, the uv-style binary distributor I started working on a few weeks ago, and have been dogfooding it with a few others. Mark Elvers and I have been cross checking that our tools are compatible while doing OCaml maintainance.

The big new feature this week is oi self update. Having distributed binaries via oi for a few weeks, the next obvious step was for oi itself to become one of the binaries it updates. This makes pushing fixes much less painful, and brings it closer to self-hosting itself.

The two features I really want oi to nail at this point are:

  • how to quickly run a binary in the uv style where you don't install anything. I've added oix for this now, so that oix utop just works — backed by source tracking and a local cache.
  • easy static-binary builds so you can ship a single binary that runs anywhere without thinking about which libc/arch the target is on. oi handles this by shelling out to Docker for the static-build pipeline. I'm still working on wiring the binary builds through so that it just works (and I need to investigate fat binaries to see if they're worth it, but I'm guessing not).
  • updates as a library, so that binaries can evolve

I've also been hacking with Thomas Gazagnaire to merge his monopampam tree back into the agentic-libraries trees from last year. Thomas has been doing an enormous amount of new coding for space protocols and has built a lovely CCSDS protocol stack. I've merged almost all of his changes back into my OCaml trees, and will look at OxCaml merges next. More on this mega monorepo as it stabilises in the next few weeks!

References

[1]Madhavapeddy (2026). Discussing effective conservation with all the UK Chief Scientists. 10.59350/qjrmv-38130
[2]Madhavapeddy (2025). Thoughts on the National Data Library and private research data. 10.59350/fk6vy-5q841
[3]Swinfield et al (2026). Learning lessons from over-crediting to ensure additionality in forest carbon credits. Nature Publishing Group. 10.1038/s41467-026-71552-3
[4]Iyer et al (2025). Careful design of Large Language Model pipelines enables expert-level retrieval of evidence-based information from syntheses and databases. 10.1371/journal.pone.0323563
[5]Gibb et al (2026). Package Managers à la Carte: A Formal Model of Dependency Resolution. arXiv. 10.48550/arXiv.2602.18602
[6]Pfeifer et al (2026). Sliding Into Silence? We Are Speaking 300 Daily Words Fewer Every Year. 10.1177/17456916261425131