I. Introduction
The rapid advancement of 6 G IoT networks marks a new era in distributed computing, enabling unprecedented levels of connectivity and communication across various sectors, including smart manufacturing, intelligent transportation, healthcare, and defense [1], [2], [3], [4]. A key technology supporting these advancements is Low Power Wide Area Networks (LPWANs), known for their cost-efficiency and wide coverage. Among LPWANs, Long Range Wide Area Networks (LoRaWAN) have gained significant attention in IoT applications like smart agriculture, healthcare, and smart metering. LoRaWAN primarily enables one-hop communication between network nodes and gateways, which simplifies deployment but can also strain gateway capacity, especially as nodes independently transmit packets via direct links. This can increase power consumption and can lead to interference from other wireless networks, causing packet loss and reducing overall network throughput. To address these issues, distributed computing in 6G IoT networks enables neighboring nodes to act as relays, adopting a multi-hop approach that reduces congestion, minimizes power use, and improves reliability and performance. Various studies have focused on reducing energy consumption and network latency in multihop data transmission in LoRaWAN technology. The authors in [5] compiled a comprehensive list of works on multi-hop transmission in LoRaWAN technology. In [6], the authors discussed the impact of IoT in modern healthcare systems, such as fitness assistance, sleep monitoring, and daily diet tracking. The survey includes studying intelligent health monitoring systems by categorizing their sensor components using device-based and device-free techniques. In 6G IoT multi-hop networks, faulty nodes may reduce the network performance and lifetime. Therefore, predicting the faulty nodes before the data transmission may further enhance network lifetime and QoS. The authors in [7] have introduced a novel Q-learning framework that successfully predicts faulty nodes and avoids data transmission, thus improving network performance. The reference [8] discusses the problem of optimal relay selection in dynamic IoT networks in which the topology is not static. A new methodology named Optimal Relay IoD Selection Using Q-Learning (OptRISQL) is used in this work, aiming to improve network performance by dynamically selecting the best relay in the network for EE and QoS.