+高级检索
D3DQN-CAA:一种基于DRL的自适应边缘计算任务调度方法
DOI:
作者:
作者单位:

作者简介:

通讯作者:

基金项目:


D3DQN-CAA:a DRL-based Adaptive Edge Computing Task Scheduling Method
Author:
Affiliation:

Fund Project:

  • 摘要
  • |
  • 图/表
  • |
  • 访问统计
  • |
  • 参考文献
  • |
  • 相似文献
  • |
  • 引证文献
  • |
  • 资源附件
    摘要:

    为解决已有基于深度强化学习的边缘计算任务调度面临的动作空间探索度固定不变、样本效率低、内存需求量大、稳定性差等问题,更好地在计算资源相对有限的边缘计算系统中进行有效的任务调度,在改进深度强化学习模型D3DQN(Dueling Double DQN)的基础上,提出了自适应边缘计算任务调度方法D3DQN-CAA.在任务卸载决策时,将任务与处理器的对应关系看作一个多维背包问题,根据当前调度任务与计算节点的状态信息,为任务选择与其匹配度最高的计算节点进行任务处理;为提高评估网络的参数更新效率,降低过估计的影响,提出一种综合性Q值计算方法;为进一步加快神经网络的收敛速度,提出了一种自适应动作空间动态探索度调整策略;为减少系统所需的存储资源,提高样本效率,提出一种自适应轻量优先级回放机制.实验结果表明,和多种基准算法相比,D3DQN-CAA方法能够有效地降低深度强化学习网络的训练步数,能充分利用边缘计算资源提升任务处理的实时性,降低系统能耗.

    Abstract:

    To solve the problems faced by the existing edge computing task scheduling based on deep reinforcement learning, such as fixed action space exploration, low sample efficiency, large memory demand and poor stability and to better carry out effective task scheduling in the edge computing system with relatively limited computing resources, an adaptive edge computing task scheduling method D3DQN-CAA is proposed based on the improved deep reinforcement learning model D3DQN (Dueling Double DQN). In the task offloading decision, the corresponding relationship between the task and processor is regarded as a multidimensional knapsack problem, and the computing node with the highest matching degree is selected for task processing according to the state information of the current scheduled task and the computing node; For improving the parameters updating efficiency of the evaluation network and reducing the influence of overestimation, a comprehensive Q-value calculation method is proposed; For accelerating the convergence speed of neural networks, an adaptive dynamic exploration degree of action space adjustment strategy is proposed; For reducing the storage resources required and improving the sample efficiency, an adaptive lightweight prioritized playback mechanism is proposed. Experimental results show that compared with multiple benchmark algorithms, the D3DQN-CAA algorithm can effectively reduce the number of training steps of deep reinforcement learning networks and make full use of edge computing resources to improve the real-time performance of task processing and reduce the system energy consumption.

    参考文献
    相似文献
    引证文献
文章指标
  • PDF下载次数:
  • HTML阅读次数:
  • 摘要点击次数:
  • 引用次数:
引用本文

巨涛 ?,王志强 ,刘帅 ,火久元 ,李启南. D3DQN-CAA:一种基于DRL的自适应边缘计算任务调度方法[J].湖南大学学报:自然科学版,2024,(6):73~85

复制
历史
  • 收稿日期:
  • 最后修改日期:
  • 录用日期:
  • 在线发布日期: 2024-07-05
  • 出版日期:
作者稿件一经被我刊录用,如无特别声明,即视作同意授予我刊论文整体的全部复制传播的权利,包括但不限于复制权、发行权、信息网络传播权、广播权、表演权、翻译权、汇编权、改编权等著作使用权转让给我刊,我刊有权根据工作需要,允许合作的数据库、新媒体平台及其他数字平台进行数字传播和国际传播等。特此声明。
关闭