###
DOI:
计算机系统应用英文版:2014,23(5):101-106
本文二维码信息
码上扫一扫!
联合无监督词聚类的递归神经网络语言模型
(中国科学技术大学 计算机科学与技术学院, 合肥 230027)
Using Word Clustering to Improve Recurrent Neural Network Language Model
(School of Computer Science and Technology, University of Science and Technology of China, Hefei 230027, China)
摘要
图/表
参考文献
相似文献
本文已被:浏览 1637次   下载 3283
Received:September 12, 2013    Revised:November 11, 2013
中文摘要: 研究表明,在递归神经网络语言模型的输入层加入词性标注信息,可以显著提高模型的效果. 但使用词性标注需要手工标注的数据训练,耗费大量的人力物力,并且额外的标注器增加了模型的复杂性. 为了解决上述问题,本文尝试将布朗词聚类的结果代替词性标注信息加入到递归神经网络语言模型输入层. 实验显示,在Penn Treebank语料上,加入布朗词类信息的递归神经网络语言模型相比原递归神经网络语言模型困惑度下降8~9%.
Abstract:Previous studies proved that, adding part of speech tag information to the input layer of neural language model, can improve the performance significantly. But part of speech tag need hand-annotated data to train the tag model, which consumes a lot and the extra tagger also makes the model more complicated. To solve the problem, this article propose adding the results of brown clustering, instead of part of speech tag information to the input layer of the recurrent network language model. In the Penn Treebank corpus, the relative improvement over the original recurrent neural network language model reaches 8%~9%.
文章编号:     中图分类号:    文献标志码:
基金项目:
引用文本:
刘章,陈小平.联合无监督词聚类的递归神经网络语言模型.计算机系统应用,2014,23(5):101-106
LIU Zhang,CHEN Xiao-Ping.Using Word Clustering to Improve Recurrent Neural Network Language Model.COMPUTER SYSTEMS APPLICATIONS,2014,23(5):101-106