1 Introduction
Distributed optimization has received extensive attention in recent years [1]–[3], where the nodes in networks collaboratively minimize the global cost function through local information transformation. This technology has been researched and developed in various fields, such as machine learning [4], game theory [5], and resource allocation [6]. However, the actual environment can be dynamic and full of uncertainty, and the setting of fixed local cost functions has limitations. To overcome this bottleneck, online optimization is widely studied, where the cost function is revealed only after the individual chooses its action. The delay in the exposure of the cost function produce a cost, so the concept of regret is introduced to evaluate the performance of the class of online optimization algorithms. Regret measures the gap between accumulated cost and the cost caused by the optimal solution with all cost functions known in advance. If an online algorithm can achieve a regret bound that grows sub-linearly with the num of iterations T, then it is called effective. Initially, Zinkevich proposed a centralized online optimization [7] and the sublinear regret bound of is obtained. Benefit from the advantages of distributed algorithms in processing big data, it was later developed into distributed online optimization [8]–[10].