I. Introduction
Neural architecture search (NAS) [1], aiming to replace human experts with machines in designing neural architectures, is widely anticipated. Typical works included reinforcement learning approaches [2], [3], evolutionary algorithms [4], [5], and Bayesian methods [6], [7]. These methods require multiple trials (i.e., training many architectures separately to assess their quality), which is computationally unaffordable for many researchers. Recent weight-sharing NAS solutions encoded a search space into a weight-sharing supernet and trained all architectures in the supernet at once, significantly improving search efficiency.