Mongolian-Chinese Neural Machine Translation Based on Adversarial Learning
CSTR:
Author:
  • Article
  • | |
  • Metrics
  • |
  • Reference [25]
  • |
  • Related [20]
  • | | |
  • Comments
    Abstract:

    In the construction and training stage of the machine translation model, the maximum likelihood estimation principle used in the end-to-end machine translation framework training can lead to the low quality of the translation model. To alleviate the problem, this study uses the adversarial learning strategy to train generative adversarial networks and improves the translation quality of the generator with the assistance of discriminators. Through experiments, the machine translation framework, Transformer, is chosen for its better performance with generators, and the convolution neural network for its better performance with discriminators. The experimental results verify that adversarial training can improve the naturalness, fluency, and accuracy of the translation. In the model optimization stage, the Mongolian-Chinese machine translation quality is still unsatisfactory due to the lack of Mongolian-Chinese parallel data sets. For improvement, the dual-generative adversarial networks (Dual-GAN) algorithm is introduced to the Mongolian-Chinese machine translation. Through the effective use of a large number of Mongolian-Chinese monolingual data, the dual learning strategy is adopted to further improve the quality of the Mongolian-Chinese machine translation model based on adversarial learning.

    Reference
    [1] Allen RB. Several studies on natural language and back-propagation. Proceedings of the IEEE 1st International Conference on Neural Networks. San Diego: IEEE, 1987. 341.
    [2] Chrisman L. Learning recursive distributed representations for holistic computation. Connection Science, 1991, 3(4): 345–366. [doi: 10.1080/09540099108946592
    [3] Kalchbrenner N, Grefenstette E, Blunsom P. A convolutional neural network for modelling sentences. Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics. Baltimore: ACL, 2014. 655–665.
    [4] Sutskever I, Vinyals O, Le QV. Sequence to sequence learning with neural networks. Proceedings of the 27th International Conference on Neural Information Processing Systems. Cambridge: ACM, 2014. 3104–3112.
    [5] Cho K, van Merri?nboer B, Gulcehre C, et al. Learning phrase representations using RNN encoder-decoder for statistical machine translation. Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP). Doha: ACL, 2014. 1724–1734.
    [6] Cho K, van Merri?nboer B, Bahdanau D, et al. On the properties of neural machine translation: Encoder-decoder approaches. Proceedings of the 8th Workshop on Syntax, Semantics and Structure in Statistical Translation. Doha: ACL, 2014. 103–111.
    [7] Bahdanau D, Cho K, Bengio Y. Neural machine translation by jointly learning to align and translate. International Conference on Learning Representations 2015. San Diego: ICLR, 2015. 1–15.
    [8] Luong M T, Pham H, Manning CD. Effective approaches to attention-based neural machine translation. Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing. Lisbon: ACL, 2015. 1412–1421.
    [9] Jean S, Cho K, Memisevic R, et al. On using very large target vocabulary for neural machine translation. Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). Beijing: ACL, 2015. 1–10.
    [10] Jean S, Firat O, Cho K, et al. Montreal neural machine translation systems for WMT’15. Proceedings of the 10th Workshop on Statistical Machine Translation. Lisbon: ACL, 2015. 134–140.
    [11] Bojar O, Buck C, Federmann C, et al. Findings of the 2014 workshop on statistical machine translation. Proceedings of the 9th Workshop on Statistical Machine Translation. Baltimore: ACL, 2014. 12–58.
    [12] Cettolo M, Niehues J, Stüker S, et al. The IWSLT 2016 evaluation campaign. International Workshop on Spoken Language Translation 2016. Seattle: Association for Computational Linguistics, 2016.
    [13] Junczys-Dowmunt M, Dwojak T, Hoang H. Is neural machine translation ready for deployment? a case study on 30 translation directions. Proceedings of the 9th International Workshop on Spoken Language Translation 2016. Seattle: Association for Computational Linguistics, 2016.
    [14] Bentivogli L, Bisazza A, Cettolo M, et al. Neural versus phrase-based machine translation quality: A case study. Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing. Austin: ACL, 2016. 257–267.
    [15] Wu YH, Schuster M, Chen ZF, et al. Google’s neural machine translation system: Bridging the gap between human and machine translation. arXiv: 1609.08144, 2016.
    [16] Hochreiter S, Schmidhuber J. Long short-term memory. Neural Computation, 1997, 9(8): 1735–1780. [doi: 10.1162/neco.1997.9.8.1735
    [17] Gehring J, Auli M, Grangier D, et al. Convolutional sequence to sequence learning. Proceedings of the 34th International Conference on Machine Learning. Sydney, 2017. 1243–1252.
    [18] Goodfellow I, Pouget-Abadie J, Mirza M, et al. Generative adversarial networks. Communications of the ACM, 2020, 63(11): 139–144. [doi: 10.1145/3422622
    [19] 刘爽, 赵景秀, 杨红亚, 等. 文本情感分析综述. 软件导刊, 2018, 17(6): 1–4, 21
    [20] 陈肇雄, 高庆狮. 自然语言处理. 计算机研究与发展, 1989, (11): 1–16
    [21] Sennrich R, Haddow B, Birch A. Improving neural machine translation models with monolingual data. Computer Science, 2015, (1): 86–96
    [22] He D, Xia YC, Qin T, et al. Dual learning for machine translation. Proceedings of the 30th International Conference on Neural Information Processing Systems. Red Hook: ACM, 2016. 820–828.
    [23] Yi ZL, Zhang H, Tan P, et al. DualGAN: Unsupervised dual learning for image-to-image translation. 2017 IEEE International Conference on Computer Vision (ICCV). Venice: IEEE, 2017. 2868–2876.
    [24] Wu LJ, Xia YC, Tian F, et al. Adversarial neural machine translation. Proceedings of the 10th Asian Conference on Machine Learning. Beijing: PMLR, 2018. 534–549.
    [25] Yang Z, Chen W, Wang F, et al. Improving neural machine translation with conditional sequence generative adversarial nets. Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers). New Orleans: ACL, 2018. 1346–1355.
    Cited by
    Comments
    Comments
    分享到微博
    Submit
Get Citation

苏依拉,王昊,贺玉玺,孙晓骞,仁庆道尔吉,吉亚图.基于对抗学习的蒙汉神经机器翻译.计算机系统应用,2022,31(1):249-258

Copy
Share
Article Metrics
  • Abstract:947
  • PDF: 1728
  • HTML: 1270
  • Cited by: 0
History
  • Received:March 30,2021
  • Revised:April 29,2021
  • Online: December 17,2021
Article QR Code
You are the first990402Visitors
Copyright: Institute of Software, Chinese Academy of Sciences Beijing ICP No. 05046678-3
Address:4# South Fourth Street, Zhongguancun,Haidian, Beijing,Postal Code:100190
Phone:010-62661041 Fax: Email:csa (a) iscas.ac.cn
Technical Support:Beijing Qinyun Technology Development Co., Ltd.

Beijing Public Network Security No. 11040202500063