I. Introduction
The worldwide rollout of 5G technology has given rise to an unprecedented surge in intelligent devices and data traffic [1]. As a result, we are witnessing the emergence of groundbreaking applications and services like the Internet of Vehicles (IoV), Virtual Reality (VR), and Expanded Reality (ER) [2]. These advancements necessitate ultra-high data rates, ultra-low transmission latency, and seamless connectivity on a massive scale [3]. In response to the pressing demands stemming from these breakthroughs, both academia and industry have begun efforts to delve deeper into next-generation networks (6G) [4], [5]. Envisioned as a cornerstone of the future communications infrastructure, 6G is poised to play a crucial role in the forthcoming decade and beyond. In the impending era of 6G, a large number of Industrial Internet of Things (IIoT) end devices will be connected to the network, while the integration of artificial intelligence and 6G networks opens the door to intelligent manufacturing. Nevertheless, end devices in IIoT generate huge amounts of data during operation, but their computational and storage capacities are limited [6], and how to efficiently handle large-scale computationally intensive and delay-sensitive tasks has become an imperative problem.