iRAF: A Deep Reinforcement Learning Approach for Collaborative Mobile Edge Computing IoT Networks

Abstract
Recently, as the development of artificial intelligence (AI), data-driven AI methods have shown amazing performance in solving complex problems to support the Internet of Things (IoT) world with massive resource-consuming and delay-sensitive services. In this paper, we propose an intelligent Resource Allocation Framework (iRAF) to solve the complex resource allocation problem for the Collaborative Mobile Edge Computing (CoMEC) network. The core of iRAF is a multi-task deep reinforcement learning algorithm for making resource allocation decisions based on network states and task characteristics, such as the computing capability of edge servers and devices, communication channel quality, resource utilization, and latency requirement of the services, etc. The proposed iRAF can automatically learn the network environment and generate resource allocation decision to maximize the performance over latency and power consumption with self-play training. iRAF becomes its own teacher: a Deep Neural Network (DNN) is trained to predict iRAF’s resource allocation action in a self-supervised learning manner, where the training data is generated from the searching process of Monte Carlo Tree Search (MCTS) algorithm. A major advantage of MCTS is that it will simulate trajectories into the future, starting from a root state, to obtain a best action by evaluating the reward value. Numerical results show that our proposed iRAF achieves 59.27% and 51.71% improvement on service latency performance compared with the greedy-search and the deep Q-learning based methods, respectively.
Funding Information
  • National Natural Science Foundation of China (61701059)

This publication has 44 references indexed in Scilit: