Skip to content
Longterm Wiki
Back

AI progress and a Safeguarded AI pivot

web

ARIA's Safeguarded AI programme director David Dalrymple explains a strategic pivot away from developing specialized AI systems toward expanding a formal verification toolkit (TA1), driven by faster-than-expected frontier AI capability progress. Relevant to AI safety as it reflects how safety-focused research programs must adapt to rapid capability advances.

Metadata

Importance: 52/100interviewnews

Summary

ARIA's Safeguarded AI programme is pivoting away from developing dedicated ML R&D teams (TA2) toward broadening its TA1 formal verification toolkit, following faster-than-expected frontier AI capability growth. The expanded toolkit will target mathematical assurance and auditability across software/hardware verification, auditable multi-agent systems, and critical infrastructure security. A formally-verified firewall for critical infrastructure is cited as a leading proof-of-concept.

Key Points

  • Frontier AI capabilities have consistently exceeded expectations over the last 4-5 model generations, prompting a strategic programme pivot.
  • TA2 Phase 2 selection was aborted because expected ML R&D capabilities are now likely to emerge from frontier AI by default.
  • TA1 toolkit scope is expanding from cyber-physical system assurance to broader mathematical assurance, software/hardware verification, and multi-agent systems.
  • A formally-verified firewall for critical infrastructure (power grids, water systems, air traffic control) is the leading proof-of-concept application.
  • ARIA's flexible programme management structure allowed rapid strategic redirection before the full picture was complete.

Cited by 2 pages

2 FactBase facts citing this source

Cached Content Preview

HTTP 200Fetched Apr 7, 20265 KB
AI progress and a Safeguarded AI pivot Skip to main content 
 
 

 

 
 
 28 November 2025

 AI progress and a Safeguarded AI pivot 

 
We sat down with Programme Director David ‘davidad’ Dalrymple to unpack how his programme is pivoting and learn what lies ahead. 
 

 

 
While our conviction in the vision for the Safeguarded AI programme remains unchanged, the pace of frontier AI progress has fundamentally altered our path – instead of investing in specialised AI systems that can use our tools, it will be more catalytic to broaden the scope and power of the TA1 toolkit itself, making it a foundational component for the next generation of AI.

 
 What led to this pivot? 

 
The decision was driven by the speed of progress in frontier AI capabilities. This has outpaced expectations from when the programme was designed.

 If we look back over the last 12 months, every frontier model has more capability than I expected at that point in time. When METR, the nonprofit that measures capabilities, evaluated them, they quantitatively exceeded expectations in the same way. So, we’re not just updating based on one model; we’re updating based on a trend of the last four or five frontier models.

 

 Data from ‘Measuring AI Ability to Complete Long Tasks’ METR 

 

 How did you decide that a pivot was needed? 

 
Honestly, it didn’t occur to me that we could depart from the TA2 Phase 2 selection process so close to applications closing. But in a conversation with Nora, the Technical Specialist on the programme, and Ant, ARIA’s CTO, I mentioned that it was more likely than not we’d select zero successful teams for Phase 2 – the value of having a dedicated team pursuing ML R&D to develop the capabilities needed for the Safeguarded AI workflow seemed much lower than when we started, because we now expect many of these capabilities to come online soon by default.

 Ant said, “What if we just stop it right now?” Once I realised that was a real option, I quickly agreed. We then had a call with Ilan, ARIA’s CEO, and he said, “Great, let’s do it.”

 Did this experience of programme management surprise you? 

 One of the things I appreciate about ARIA is that they back Programme Directors to steer into new directions before the picture is complete. I’m grateful we were able to move so quickly on aborting the TA2 selection, while also having space to develop the new vision thoughtfully.

 As part of the pivot, you indicated wanting to expand the scope and ambition of TA1… 

 
Yes, instead of narrowly targeting assurance for cyber-physical system control, we are thinking about TA1 as a toolkit for mathematical assurance and auditability across a wider range of areas, including software and hardware verification, auditable multi-agent systems, and more informal knowledge.

 Can you give us an example of one of those challenges? 

 
My current leading hypothesis for a proof point for the platform we’re building is formally-verified firewall. Specifically, securing point

... (truncated, 5 KB total)
Resource ID: kb-1ad0fbdc383198b6