Loading [MathJax]/extensions/MathZoom.js
A Hybrid Framework of Reinforcement Learning and Convex Optimization for UAV-Based Autonomous Metaverse Data Collection | IEEE Journals & Magazine | IEEE Xplore

A Hybrid Framework of Reinforcement Learning and Convex Optimization for UAV-Based Autonomous Metaverse Data Collection


Abstract:

Unmanned aerial vehicles (UAVs) are promising for providing communication services due to their advantages in cost and mobility, especially in the context of the emerging...Show More

Abstract:

Unmanned aerial vehicles (UAVs) are promising for providing communication services due to their advantages in cost and mobility, especially in the context of the emerging Metaverse and Internet of Things (IoT). This article considers a UAV-assisted Metaverse network, in which UAVs extend the coverage of the base station (BS) to collect the Metaverse data generated at roadside units (RSUs). Specifically, to improve the data collection efficiency, resource allocation and trajectory control are integrated into the system model. The time-dependent nature of the optimization problem makes it non-trivial to be solved by traditional convex optimization methods. Based on the proposed UAV-assisted Metaverse network system model, we design a hybrid framework with reinforcement learning and convex optimization to cooperatively solve the time-sequential optimization problem. Simulation results show that the proposed framework is able to reduce the mission completion time with a given transmission power resource.
Published in: IEEE Network ( Volume: 37, Issue: 4, July/August 2023)
Page(s): 248 - 254
Date of Publication: 24 October 2023

ISSN Information:

References is not available for this document.

Introduction

Metaverse is an emerging concept that aims to generate the digital duality of the physical world and provide users with immersive experience through augmented reality (AR), virtual reality (VR), and other novel techniques [1]. To ensure the quality of service (QoS), 5G or future 6G communication technologies are needed to support the required transmission rate of AR/VR service. However, the improvement of transmission rate is at the cost of the coverage area of one single base station (BS) due to severe attenuation of signals, which makes it more expensive to provide wireless communication coverage. On the other hand, the maintenance and update of the virtual world in Metaverse is based on massive data collected by individual devices, roadside units (RSUs), and other Internet of Things (IoT) devices. Thus, it is prominent to extend the coverage of Metaverse data collection and synchronization to provide more users with seamless service.

Select All
1.
R. Cheng et al., "Will Metaverse be NextG Internet? Vision Hype and Reality", IEEE Network, vol. 36, no. 5, pp. 197-204, 2022.
2.
M. Xu et al., "A Full Dive into Realizing the Edge-Enabled Metaverse: Visions Enabling Technologies and Challenges", IEEE Commun. Surveys & Tutorials, 2022.
3.
S. Wan et al., "Toward Big Data Processing in IoT: Path Planning and Resource Management of UAV Base Stations in Mobile-Edge Computing System", IEEE Internet of Things J., vol. 7, no. 7, pp. 5995-6009, 2019.
4.
Z. Wei et al., "UAV Assisted Data Collection for Internet of Things: A Survey", IEEE Internet of Things J., vol. 9, no. 17, pp. 460-83, 2022.
5.
Y. Han et al., "A Dynamic Resource Allocation Framework for Synchronizing Metaverse with IoT Service and Data", ICC 2022-IEEE Int'I. Conf. Commun., 2022.
6.
L. Shen et al., "Energy-Aware Dynamic Trajectory Planning for UAV-enabled Data Collection in mMTC Networks", IEEE Trans. Green Commun. Networking, vol. 6, no. 4, pp. 1957-71, 2022.
7.
K. Liu and J. Zheng, "UAV Trajectory Optimization for Time-constrained Data Collection in UAV-Enabled Environmental Monitoring Systems", IEEE Internet of Things J., vol. 9, no. 23, pp. 24,300-14, 2022.
8.
S. Burer and A. N. Letchford, "Non-Convex Mixed-integer Nonlinear Programming: A Survey", Surveys in Operations Research and Management Science, vol. 17, no. 2, pp. 97-106, 2012.
9.
H. Ye, G. Y. Li and B.-H. F. Juang, "Deep Reinforcement Learning based Resource Allocation for V2V Communications", IEEE Trans. Vehicular Technology, vol. 68, no. 4, pp. 3163-73, 2019.
10.
J. Cui, Y. Liu and A. Nallanathan, "Multi-Agent Reinforcement Learning-based Resource Allocation for UAV Networks", IEEE Trans. Wireless Commun., vol. 19, no. 2, pp. 729-43, 2019.
11.
I. Grondman et al., "A Survey of Cctor-critic Reinforcement Learning: Standard and Natural Policy Gradients", IEEE Trans. Systems Man and Cybernetics Part C (Applications and Reviews), vol. 42, no. 6, pp. 1291-1307, 2012.
12.
J. Schulman et al., "Proximal Policy Optimization Algorithms", arXiv preprint, 2017.
13.
J. Fan et al., "A Theoretical Analysis of Deep Q-learning", Learning for Dynamics and Control, pp. 486-89, 2020.
14.
L. Pinto et al., "Robust Adversarial Reinforcement Learning", Proc. Int'l. Conf. Machine Learning, pp. 2817-26, 2017.
15.
Y. Arjevani et al., "Oracle Complexity of Secondorder Methods for Smooth Convex Optimization", Mathematical Programming, vol. 178, no. 1, pp. 327-60, 2019.
Contact IEEE to Subscribe

References

References is not available for this document.