一种基于BERT的自动文本摘要模型构建方法  被引量:4

An Automatic Text Summarization Model Construction Method Based on BERT Embedding

在线阅读下载全文

作  者:岳一峰 黄蔚[1] 任祥辉[1] YUE Yi-feng;HUANG Wei;REN Xiang-hui(North China Institute of Computing Technology,Beijing 100083,China)

机构地区:[1]华北计算技术研究所

出  处:《计算机与现代化》2020年第1期63-68,共6页Computer and Modernization

基  金:国家重点研发计划资助项目(2016YFB0801400)

摘  要:针对传统词向量在自动文本摘要过程中因无法对多义词进行有效表征而降低文本摘要准确度和可读性的问题,提出一种基于BERT(Bidirectional Encoder Representations from Transformers)的自动文本摘要模型构建方法。该方法引入BERT预训练语言模型用于增强词向量的语义表示,将生成的词向量输入Seq2Seq模型中进行训练并形成自动文本摘要模型,实现对文本摘要的快速生成。实验结果表明,该模型在Gigaword数据集上能有效地提高生成摘要的准确率和可读性,可用于文本摘要自动生成任务。Aiming at the problem that the traditional word vector can not effectively represent polysemous words in text summarization,which reduces the accuracy and readability of summarization,this paper proposes an automatic text summarization model construction method based on BERT( Bidirectional Encoder Representations from Transformers) Embedding. This method introduces the BERT pre-training language model to enhance the semantic representation of word vector. The generated word vectors are input into the Seq2 Seq model for training to form an automatic text summarization model,which realizes the rapid generation of text summarization. The experimental results show that the model can effectively improve the accuracy and readability of the generated summarization on Gigaword dataset,and can be used for automatic text summarization generation tasks.

关 键 词:文本摘要 BERT模型 注意力机制 Sequence-to-Sequence模型 

分 类 号:TP391.1[自动化与计算机技术—计算机应用技术]

 

参考文献:

正在载入数据...

 

二级参考文献:

正在载入数据...

 

耦合文献:

正在载入数据...

 

引证文献:

正在载入数据...

 

二级引证文献:

正在载入数据...

 

同被引文献:

正在载入数据...

 

相关期刊文献:

正在载入数据...

相关的主题
相关的作者对象
相关的机构对象