检索规则说明:AND代表“并且”;OR代表“或者”;NOT代表“不包含”;(注意必须大写,运算符两边需空一格)
检 索 范 例 :范例一: (K=图书馆学 OR K=情报学) AND A=范并思 范例二:J=计算机应用与软件 AND (U=C++ OR U=Basic) NOT M=Visual
作 者:Zaipeng Xie Cheng Ji Chentai Qiao WenZhan Song Zewen Li Yufeng Zhang Yujing Zhang
机构地区:[1]Key Laboratory of Water Big Data Technology of Ministry of Water Resources,Hohai University,Nanjing,China [2]College of Computer and Information,Hohai University,Nanjing,China [3]Center for Cyber‐Physical Systems,University of Georgia,Athens,Georgia,USA [4]Information Networking Institute,Carnegie Mellon University,Pittsburgh,Pennsylvania,USA [5]Department of Electrical and Systems Engineering,University of Pennsylvania,Philadelphia,Pennsylvania,USA
出 处:《CAAI Transactions on Intelligence Technology》2024年第4期1014-1030,共17页智能技术学报(英文)
基 金:National Natural Science Foundation of China,Grant/Award Number:61872171;The Belt and Road Special Foundation of the State Key Laboratory of Hydrology‐Water Resources and Hydraulic Engineering,Grant/Award Number:2021490811。
摘 要:Multi‐agent reinforcement learning relies on reward signals to guide the policy networks of individual agents.However,in high‐dimensional continuous spaces,the non‐stationary environment can provide outdated experiences that hinder convergence,resulting in ineffective training performance for multi‐agent systems.To tackle this issue,a novel reinforcement learning scheme,Mutual Information Oriented Deep Skill Chaining(MioDSC),is proposed that generates an optimised cooperative policy by incorporating intrinsic rewards based on mutual information to improve exploration efficiency.These rewards encourage agents to diversify their learning process by engaging in actions that increase the mutual information between their actions and the environment state.In addition,MioDSC can generate cooperative policies using the options framework,allowing agents to learn and reuse complex action sequences and accelerating the convergence speed of multi‐agent learning.MioDSC was evaluated in the multi‐agent particle environment and the StarCraft multi‐agent challenge at varying difficulty levels.The experimental results demonstrate that MioDSC outperforms state‐of‐the‐art methods and is robust across various multi‐agent system tasks with high stability.
关 键 词:artificial intelligence techniques decision making intelligent multi‐agent systems
分 类 号:TP18[自动化与计算机技术—控制理论与控制工程]
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在链接到云南高校图书馆文献保障联盟下载...
云南高校图书馆联盟文献共享服务平台 版权所有©
您的IP:216.73.216.49