Reinforcement Learning-Based Optimization of Quality-of-Service and Path Planning for Multi-UAV Mobile Edge Computing

Authors

  • Xiang Li
  • Qi Liu
  • Zhuocheng Yang

DOI:

https://doi.org/10.54097/vw4ykh27

Keywords:

Multi-UAV, Mobile Edge Computing, Reinforcement Learning, Quality-of-Service, Path Planning.

Abstract

Unmanned Aerial Vehicles (UAVs) have traditionally served as network processors in mobile networks, and more recently, as mobile servers in Mobile Edge Computing (MEC). However, deploying UAVs in dynamic and obstacle-rich environments while ensuring efficient coordination among multiple UAVs presents significant challenges. To tackle these challenges, we propose a unified reinforcement learning framework for a multi-UAV MEC platform that enhances Quality-of-Service (QoS) and optimizes path planning. Specifically, our contributions include: (1) Integrating QoS optimization and UAV path planning into a unified reinforcement learning framework; (2) Modeling terminal users' demands with a sigmoid-like function to enhance service quality; (3) Incorporating terminal users' demand, risk factors, and geometric distance into the reinforcement learning reward matrix for balancing service quality, risk mitigation, and cost efficiency. Experimental results demonstrate the framework’s effectiveness: simulations in 3D environments show a 20–35% reduction in collision incidents compared to greedy algorithms and a 15–25% improvement in demand fulfillment rates under dynamic user distributions. Further analysis highlights the impact of sigmoid parameter tuning on risk-QoS trade-offs, validating the adaptability of the proposed approach in complex scenarios.

Downloads

Download data is not yet available.

References

[1] Alma Saeid M H. UAV-assisted mobile edge computing model for cognitive radio-based IoT networks [J]. Computer Communications, 2025, 233108071 - 108071.

[2] Muwafaq L, Noordin K N, Othman M, et al. A Survey on Cloudlet Computation Optimization in the Mobile Edge Computing Environment [J]. International Journal of Advanced Computer Science and Applications (IJACSA), 2023, 14 (1).

[3] Asiful S H, Sangman M. Survey on computation offloading in UAV-Enabled mobile edge computing [J]. Journal of Network and Computer Applications, 2022, 201.

[4] Technology - Vehicle Technology; Report Summarizes Vehicle Technology Study Findings from University of Electronic Science and Technology of China (Joint Resources and Workflow Scheduling in Uav-enabled Wirelessly-powered Mec for Iot Systems) [J]. Telecommunications Weekly, 2020, Y. Du, K.

[5] Margot D. Editorial: Special Issue "Unmanned Aerial Vehicle (UAV)-Enabled Wireless Communications and Networking”. [J]. Sensors (Basel, Switzerland), 2022, 22 (12): 4458 - 4458.

[6] Lav G, Raj J, Gabor V. Survey of Important Issues in UAV Communication Networks [J]. IEEE Communications Surveys & Tutorials, 2016, 18 (2): 1123 - 1152.

[7] Internet and World Wide Web - Internet of Things; New Findings Reported from Beijing University of Posts and Telecommunications Describe Advances in Internet of Things (Multi-uav-enabled Load-balance Mobile-edge Computing for Iot Networks) [J]. Information Technology Newsweekly,2020,442-.

[8] LUO Y, DING W, ZHANG B, et al. Optimization of bits allocation and path planning with trajectory constraint in UAV-enabled mobile edge computing system [J]. Chinese Journal of Aeronautics, 2020, (republish).

[9] Wang F, Zhang Z, Zhou L, et al. Robust Multi-UAV Cooperative Trajectory Planning and Power Control for Reliable Communication in the Presence of Uncertain Jammers [J]. Drones, 2024, 8 (10): 558 - 558.

[10] Qian L, Long S, Linlin S, et al. Path Planning for UAV-Mounted Mobile Edge Computing with Deep Reinforcement Learning[J]. IEEE Transactions on Vehicular Technology, 2020, 69 (5): 1 - 1.

[11] Faraci G, Grasso C, Schembra G. Design of a 5G Network Slice Extension with MEC UAVs Managed with Reinforcement Learning [J]. IEEE Journal on Selected Areas in Communications,2020, PP (99): 1 - 1.

[12] Lu H, He X, Zhang D. Security-Aware Task Offloading Using Deep Reinforcement Learning in Mobile Edge Computing Systems [J]. Electronics, 2024, 13 (15): 2933 - 2933.

[13] Aviation - Unmanned Aerial Vehicle; Investigators at Beijing University of Posts and Telecommunications Detail Findings in Unmanned Aerial Vehicle (Multi-uav Dynamic Wireless Networking with Deep Reinforcement Learning) [J]. Defense & Aerospace Week, 2020,

[14] Maria A, Muhammad S, Almas A, et al. Distributed application execution in fog computing: A taxonomy, challenges and future directions [J]. Journal of King Saud University - Computer and Information Sciences, 2022, 34 (7): 3887 - 3909.

[15] S. Kim, H. Oh, J. Suk, and A. Tsourdos. Coordinated trajectory planning for efficient communication relay using multiple uavs [J]. Control Engineering Practice, vol. 29, pp. 42 – 49, 2014.

[16] Baochang Z, Wanquan L, Zhili M, et al. Cooperative and Geometric Learning Algorithm (CGLA) for path planning of UAVs with limited information [J]. Automatica,2013, 50 (3): 809 - 820.

[17] LUO Y, DING W, ZHANG B, et al. Optimization of bits allocation and path planning with trajectory constraint in UAV-enabled mobile edge computing system [J]. Chinese Journal of Aeronautics, 2020, (republish).

[18] Cao X, Xu J, and Zhang R. Mobile edge computing for cellular- connected uav: Computation offloading and trajectory optimization [J]. In 2018 IEEE 19th International Workshop on Signal Processing Advances in Wireless Communications (SPAWC). IEEE, 2018, pp.1 – 5.

[19] Cheng N, Lyu F, Quan W, Zhou C, He H, Shi W, and Shen X. Space/aerial-assisted computing offloading for iot applications: A learning-based approach [J]. IEEE Journal on Selected Areas in Commu- nications, vol. 37, no. 5, pp. 1117 – 1129, 2019.

[20] Pengshuo J, Jie J, Jian C, et al. Reinforcement learning based joint trajectory design and resource allocation for RIS-aided UAV multicast networks [J]. Computer Networks, 2023, 227.

[21] Lee J.-W, Mazumdar R, and Shroff N. B. Non-convex optimization and rate control for multi-class services in the internet [J]. IEEE/ACM transactions on networking, vol. 13, no. 4, pp. 827 – 840, 2005.

Downloads

Published

11-05-2025

How to Cite

Li , X., Liu, Q., & Yang, Z. (2025). Reinforcement Learning-Based Optimization of Quality-of-Service and Path Planning for Multi-UAV Mobile Edge Computing. Highlights in Science, Engineering and Technology, 138, 246-254. https://doi.org/10.54097/vw4ykh27