本文已被:浏览 1052次 下载 1932次
Received:February 24, 2021 Revised:March 19, 2021
Received:February 24, 2021 Revised:March 19, 2021
中文摘要: 短文本匹配是自然语言处理领域中的一个核心问题, 可应用于信息检索、问答系统、复述问题等任务. 过去的工作大多在提取文本特征时只考虑文本内部信息, 忽略了两个文本之间的交互信息, 或者仅进行单层次交互. 针对以上问题, 提出一种基于Transformer改进的短文本匹配模型ISTM. ISTM模型以DSSM为基本架构, 利用BERT模型对文本进行向量化表示, 解决Word2Vec一词多义的问题, 使用Transformer编码器对文本进行特征提取, 获取文本内部信息, 并考虑两个文本之间的多层次交互信息, 最后由拼接向量推理计算出两个文本之间的语义匹配度. 实验表明, 相比经典深度短文本匹配模型, 本文提出的ISTM模型在LCQMC中文数据集上表现出了更好的效果.
Abstract:Short text matching is a core problem in the field of natural language processing, which can be applied to tasks such as information retrieval, question answering systems, and question paraphrase. Most of the past work only considered the internal information of the text when extracting text features, ignoring the interactive information between two texts, or only performed single-level interaction. Given the above problems, an Improved Short Text Matching model (ISTM) based on Transformer is constructed. The ISTM model takes DSSM as the basic architecture and uses the BERT model to vectorize the text to solve the ambiguity of Word2Vec. It relies on the Transformer encoder to extract features of the text and obtain its internal information. It considers the multi-level interactive information between the two texts and finally infers and computes the degree of semantic matching between two texts by the concatenated vector. Experiments show that compared with the classic deep short text matching model, the ISTM model proposed in this study shows better results on the LCQMC Chinese dataset.
文章编号: 中图分类号: 文献标志码:
基金项目:智能化公共法律服务关键技术湖南省重点研发项目(2022SK2106)
引用文本:
蔡林杰,刘新,刘龙,唐朝.基于Transformer的改进短文本匹配模型.计算机系统应用,2021,30(12):268-272
CAI Lin-Jie,LIU Xin,LIU Long,TANG Chao.Improved Short Text Matching Model Based on Transformer.COMPUTER SYSTEMS APPLICATIONS,2021,30(12):268-272
蔡林杰,刘新,刘龙,唐朝.基于Transformer的改进短文本匹配模型.计算机系统应用,2021,30(12):268-272
CAI Lin-Jie,LIU Xin,LIU Long,TANG Chao.Improved Short Text Matching Model Based on Transformer.COMPUTER SYSTEMS APPLICATIONS,2021,30(12):268-272