How promising is compute usage regulation for AI governance?

Recent AI progress has been characterized as relying on the "AI triad": compute, data, and algorithms. AI development cannot proceed without all three, making regulation of these components a candidate strategy for AI governance. Regulating compute may be especially viable, because:

  • Compute cannot be used simultaneously by more than one actor, making it possible to restrict it more effectively than data and algorithms, which can be copied and used by anyone once they have been made public.

  • Compute takes up substantial physical space and requires physical materials to build and maintain.

  • Amounts of compute are easily quantifiable (e.g., in terms of FLOPS or number of CPUs).

It’s also easier to govern compute because of some related features of the computing hardware industry today:

  • Training the best AI models today requires especially large amounts of compute. This means they require a lot of physical infrastructure. This can involve “football-field-sized supercomputer centers'', which have extremely high energy and water demands, making them easy to track and detect.

ML training runs are more expensive than they used to be, so much so that Sevilla et al. call the current era of ML the “Large-Scale Era”1

However, there are a number of limitations to compute governance. For example:

  • Improvements in algorithms might mean that less compute will be necessary to make dangerous models.

  • Improvements in hardware efficiency might make it possible to train dangerous models more cheaply and in a way that is more difficult to track.

  • Scaffolding can allow AI systems to accomplish a broader range of tasks, for de facto increased capabilities, even without new training runs. Post-training optimization such as RLHF and supervised fine-tuning also increase capabilities of base models without substantially increasing the number of FLOPs used in their training.

  • Compared to other interventions like supporting alignment research, compute usage regulation is more potentially adversarial and poses the risk of increasing political or geopolitical tensions. For example, some have identified this as an effect of the early 2023 US export controls for advanced microchips to China.


  1. Sevilla et al. identified three trends: An 18-month doubling time, between 1952 and 2010, is in line with Moore’s law (according to which transistor density doubles roughly every two years) a 6-month doubling time between 2010 and 2022, and a new trend starting in 2015 with a 10-month doubling time, but it started 2 to 3 orders of magnitude over the previous trend. Increases in compute requirements exceeding Moore’s law require increasing monetary and space investment. ↩︎

  2. 92% global market share of commercial production of chips below 10nm (Netherlands Innovation Network report, June 2022) ↩︎



AISafety.info

We’re a global team of specialists and volunteers from various backgrounds who want to ensure that the effects of future AI are beneficial rather than catastrophic.