1 Introduction
Deep learning is continuing to deliver groundbreaking results on the path towards human-level intelligence. This path is character-ized by growing model size, in just six years, the compute requirements for model training grew by 300,000 times [3]. Transform-ers [54] are a typical representative in this trend. As the model size grows, Transformer based models have proven their success in in the field of natural language processing [13, 43, 43, 54]. Recent work [8-10, 14] shows that Transformers also achieve promising re-sults in computer vision tasks, i.e., on par or better than other types of models such as convolutional [31] and recurrent [21] networks. These growing models must be trained on distributed accelera-tor supercomputers. Even today's models are too big to be stored on a single accelerator-for example, GPT-3's 175 billion parame-ters [7] require 350 GiB main memory if stored with 16 bits precision. Switch transformers [17] have in their largest configuration 1.6 tril-lion parameters, a 6.4 TiB storage requirements. Furthermore, the necessary memory for activations, gradients, and optimizer state during training at least triples these memory requirements.