检索规则说明:AND代表“并且”;OR代表“或者”;NOT代表“不包含”;(注意必须大写,运算符两边需空一格)
检 索 范 例 :范例一: (K=图书馆学 OR K=情报学) AND A=范并思 范例二:J=计算机应用与软件 AND (U=C++ OR U=Basic) NOT M=Visual
出 处:《计算机科学与应用》2025年第1期220-229,共10页Computer Science and Application
摘 要:信息抽取技术随着自然语言处理技术的发展,已经取得了较好的发展,但在实际应用中,由于算法标注数据需求高、训练代价大,上下文理解难,私有化领域落地一直存在较高瓶颈。本文提出了一种基于提示增强的LLM信息抽取算法(LLM-IE Base on Prompt Enhance),通过将文本信息抽取任务转化为文本生成任务,并基于生成文本进行结构化解析,形成文本信息抽取结果。该方法在实体、关系、事件三类自建数据集上进行测试验证,面对少样本困境,该方法通过提示增强激发模型信息提取任务能力,可以近似达成模型微调的效果,同时相较于其他主流信息抽取模型在准确率与召回率上都有提升。With the development of natural language processing technology, information extraction techniques have made significant progress. However, in practical applications, due to high algorithmic annotation data requirements, large training costs, and challenges in understanding context, private domain implementations have consistently faced high barriers. This paper proposes an information extraction algorithm for LLMs based on prompt enhance (LLM-IE Based on Prompt Enhance). This method transforms text information extraction tasks into text generation tasks and performs structured parsing based on the generated text to form the results of information extraction. The method was tested and validated on three self-built datasets for entities, relationships, and events. In addressing the challenge of limited sample data, this approach can approximate the effect of model fine-tuning by stimulating the model’s information extraction task capabilities through prompt Enhancement. Additionally, compared to other mainstream information extraction models, this method shows improvements in both accuracy and recall rates.
分 类 号:TP3[自动化与计算机技术—计算机科学与技术]
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在载入数据...
正在链接到云南高校图书馆文献保障联盟下载...
云南高校图书馆联盟文献共享服务平台 版权所有©
您的IP:3.133.113.227