基于弱点增强的LLM知识蒸馏算法  

A Knowledge Distillation Algorithm for LLM Based on Weakness Enhancement

在线阅读下载全文

作  者:李崭 

机构地区:[1]中国电子科技集团公司第十研究所,四川 成都

出  处:《人工智能与机器人研究》2025年第2期304-312,共9页Artificial Intelligence and Robotics Research

基  金:国家自然科学基金资助项目(项目编号:62406263)。

摘  要:得益于大语言模型的技术发展,自然语言处理领域的知识蒸馏范式发生了颠覆式的改变。大模型提示知识获取方式,使得知识蒸馏的方式向着更通用的知识获取以及数据增强的方式发展。面向大模型时代知识提炼模式转变以及小模型小样本学习挑战,本文提出了一种基于弱点增强的LLM知识蒸馏算法,结合LLM的语义理解以及文本生成能力,实现小样本情况下的学生模型弱点分析,通过LLM教师模型针对学生模型弱点进行增强样本构建迭代训练强化,增强学生模型的能力。通过在多种自然语言处理实验结果表明,本文提出的方法在少样本标注需求下,通过知识蒸馏,可以大幅度提升模型的训练效果,充分证明了方法的有效性。Thanks to the technological advancements in large language models, the knowledge distillation paradigm in the field of natural language processing has undergone a revolutionary change. The knowledge acquisition method prompted by LLM has led to a shift in knowledge distillation towards more universal knowledge acquisition and data augmentation approaches. In response to the transformation of knowledge extraction patterns in the era of LLM and the challenges of small-model, few-sample learning, this paper proposes a knowledge distillation algorithm for LLM based on weakness enhancement. By leveraging the semantic understanding and text generation capabilities of LLM, this algorithm enables the analysis of student model weaknesses under few-sample conditions. The LLM teacher model enhanced samples to train and strengthen the student model to enhance student model’s capabilities. Experimental results in various NLP tasks demonstrate that the proposed method, under the requirement of few labeled samples, can significantly improve the training effectiveness of models through knowledge distillation, fully proving the effectiveness of the method.

关 键 词:知识蒸馏 小样本弱点分析 大模型数据增强 

分 类 号:TP391[自动化与计算机技术—计算机应用技术]

 

参考文献:

正在载入数据...

 

二级参考文献:

正在载入数据...

 

耦合文献:

正在载入数据...

 

引证文献:

正在载入数据...

 

二级引证文献:

正在载入数据...

 

同被引文献:

正在载入数据...

 

相关期刊文献:

正在载入数据...

相关的主题
相关的作者对象
相关的机构对象