检索规则说明:AND代表“并且”;OR代表“或者”;NOT代表“不包含”;(注意必须大写,运算符两边需空一格)
检 索 范 例 :范例一: (K=图书馆学 OR K=情报学) AND A=范并思 范例二:J=计算机应用与软件 AND (U=C++ OR U=Basic) NOT M=Visual
作 者:李赟波 王士同 LI Yunbo;WANG Shitong(School of Artificial Intelligence and Computer Science,Jiangnan University,Wuxi,Jiangsu 214122,China;Jiangsu Key Construction Laboratory of IoT Application Technology,Jiangnan University,Wuxi,Jiangsu 214122,China)
机构地区:[1]江南大学人工智能与计算机学院,江苏无锡214122 [2]江南大学江苏省物联网应用技术重点建设实验室,江苏无锡214122
出 处:《计算机科学与探索》2023年第6期1441-1452,共12页Journal of Frontiers of Computer Science and Technology
基 金:国家自然科学基金(61972181)。
摘 要:深度决策树迁移学习Boosting方法(DTrBoost)仅能适应一个源域与一个目标域的训练数据,无法适应多个不同分布的源域的样本。此外,DTrBoost方法同步地从源域中学习数据至目标域模型,并没有根据重要程度量化学习知识的权重。在实践中,对于某数据集的数据按照某一或某些特征划分出来的数据往往分布不一致,并且这些不同分布的数据对于最终模型的重要性也不一致,知识迁移的权重也因此不平等。针对这一问题,提出了多源域优化权重的迁移学习方法,主要思想是根据不同分布的源域空间计算出到目标域的KL距离,利用KL距离的比值计算出不同分布的源域样本的学习权重比例,从而优化整体梯度函数,使学习方向朝着梯度下降最快的方向进行。使用梯度下降算法能使模型较快收敛,在确保迁移学习效果的同时,也能确保学习的速度。实验结果表明,提出的算法在整体上实现了更好的性能并且对于不同的训练数据能够实现自适应效果,分类错误率平均下降0.013,在效果最好的OCR数据集上下降0.030。The deep decision tree transfer learning Boosting method(DTrBoost)can only adapt to the training data of one source domain and one target domain,and can not adapt to the samples of several different distribution source domains.In addition,the DTrBoost method synchronously learns data from the source domain to the target domain model,and does not quantify the weight of the learned knowledge according to the degree of importance.In practice,the distribution of data divided according to one or some characteristics of a certain dataset is often inconsistent,the importance of these different distributions to the final model is also inconsistent,and the weight of knowledge transfer is therefore not equal.To solve this problem,a transfer learning method of multi-source domain weight optimization is proposed.The main idea is to calculate the KL divergence distance to the target domain according to the source domain space of different distributions,and calculate the learning weight proportion parameters of the source domain samples of different distributions by using the ratio of KL divergence,so as to optimize the overall gradient function and make the learning direction towards the direction of the fastest gradient decline.The gradient descent algorithm can make the model converge quickly,and ensure the learning speed as well as the transfer learning effect.Experimental results show that the algorithm proposed in this paper adaptively achieves better average performance on the whole.The average classification error rate on all the adopted datasets decreases by 0.013 and even 0.030 on OCR dataset.
关 键 词:深度决策树迁移学习Boosting方法(DTrBoost) 多源域迁移学习 KL距离 决策树
分 类 号:TP181[自动化与计算机技术—控制理论与控制工程]
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在链接到云南高校图书馆文献保障联盟下载...
云南高校图书馆联盟文献共享服务平台 版权所有©
您的IP:18.221.161.189