MEC网络中基于强化学习的任务卸载和资源分配  

Reinforcement learning-based task offloading and resource allocation in MEC networks

在线阅读下载全文

作  者:陈雷[1] CHEN Lei(College of Public Security Information Technology and Intelligence,Criminal Investigation Police University of China,Shenyang 110854,China)

机构地区:[1]中国刑事警察学院公安信息技术与情报学院,辽宁沈阳110854

出  处:《武汉大学学报(工学版)》2024年第3期363-371,共9页Engineering Journal of Wuhan University

基  金:中国刑事警察学院重大培育项目(编号:D2023002);沈阳市社会治理科技专项(编号:22-322-3-35);辽宁省教育厅科学研究项目(编号:ZGXJ2020005);中国刑事警察学院校级项目(编号:D2022045)。

摘  要:针对基于移动边缘计算(mobile edge computing, MEC)的双层蜂窝网络中由于移动设备的任务迁移而产生额外开销的问题,在移动感知下通过联合任务卸载和资源分配来减少任务迁移概率,进而最大化用户总收益。首先,提出了最大化用户总收益的最优化问题;其次,在考虑时变的计算任务和资源分配下,将最优化问题描述为一个马尔科夫决策过程(Markov decision process, MDP),同时,提出了一个新颖的采用基于Q-学习的强化学习算法(reinforcement learning-based algorithm with Q-learning method, RLAQM)进行求解;最后,仿真验证了所提出的算法与其他算法相比能明显提高用户总收益。In order to solve the problem of extra cost caused by task migration of mobile devices in two-layer cellular networks based on mobile edge computing(MEC),a joint mobile awareness task offloading and resource allocation strategy is proposed to reduce the probability of task migration and maximize the total user revenue.Firstly,the problem of maximizing the total user revenue is presented.Secondly,considering time-varying computation tasks and resource allocation,the optimization problem is described as a Markov decision process(MDP).At the same time,a novel reinforcement learning-based algorithm with Q-learning method(RLAQM)is proposed to solve the MDP.Finally,simulation results show that compared with other algorithms,the proposed algorithm can significantly improve the total revenue of users.

关 键 词:移动边缘计算 任务卸载 移动感知 马尔科夫决策过程 强化学习 

分 类 号:TN929.53[电子电信—通信与信息系统]

 

参考文献:

正在载入数据...

 

二级参考文献:

正在载入数据...

 

耦合文献:

正在载入数据...

 

引证文献:

正在载入数据...

 

二级引证文献:

正在载入数据...

 

同被引文献:

正在载入数据...

 

相关期刊文献:

正在载入数据...

相关的主题
相关的作者对象
相关的机构对象