Automatic Function Naming Based on Graph Convolutional Network
CSTR:
Author:
  • Article
  • | |
  • Metrics
  • |
  • Reference [28]
  • |
  • Related [20]
  • | | |
  • Comments
    Abstract:

    Automatic method naming, as an important task in software engineering, aims to generate the target function name for an input source code to enhance the readability of program codes and accelerate software development. Existing automatic method naming approaches based on machine learning mainly encode the source code through sequence models to automatically generate the function name. However, these approaches are confronted with problems of long-term dependency and code structural encoding. To better extract structural and semantic information from programs, we propose a automatic function naming method called TrGCN based on Transformer and Graph Convolutional Network (GCN). In this method, the self-attention mechanism in Transformer is used to alleviate the long-term dependency and the Character-word attention mechanism to extract the semantic information of codes. The TrGCN introduces a GCN-based AST Encoder that enriches the eigenvector information at AST nodes and models the structural information of the source code well. Empirical studies are conducted on three Java datasets. The results show that TrGCN outperforms conventional approaches, namely code2seq and Sequence-GNNs, in automatic method naming as its F1-score is 5.2% and 2.1% higher than the values of the two approaches, respectively.

    Reference
    [1] Høst EW, Østvold BM. Debugging method names. Proceedings of the 23rd European Conference on Object-Oriented Programming. Genoa, Italy. 2009. 294-317.
    [2] Sutskever I, Vinyals O, Le QV. Sequence to sequence learning with neural networks. Proceedings of the 27th International Conference on Neural Information Processing Systems. Montreal, QC, Canada. 2014. 3104-3112.
    [3] Cho K, Van Merriënboer B, Gulcehre C, et al. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv:1406.1078, 2014.
    [4] Bahdanau D, Cho K, Bengio Y. Neural machine translation by jointly learning to align and translate. arXiv:1409.0473, 2014.
    [5] Luong MT, Pham H, Manning CD. Effective approaches to attention-based neural machine translation. arXiv:1508.04025, 2015.
    [6] Allamanis M, Peng H, Sutton C. A convolutional attention network for extreme summarization of source code. Proceedings of the 33rd International Conference on Machine Learning. New York City, NY, USA. 2016. 2091-2100.
    [7] Alon U, Zilberstein M, Levy O, et al. A general path-based representation for predicting program properties. ACM SIGPLAN Notices, 2018, 53(4):404-419.[doi:10.1145/3296979.3192412
    [8] Alon U, Zilberstein M, Levy O, et al. Code2vec:Learning distributed representations of code. Proceedings of the ACM on Programming Languages, 2019, 3(POPL):40
    [9] Alon U, Brody S, Levy O, et al. Code2seq:Generating sequences from structured representations of code. arXiv:1808.01400, 2018.
    [10] Fernandes P, Allamanis M, Brockschmidt M. Structured neural summarization. arXiv:1811.01824, 2018.
    [11] Bengio Y, Simard P, Frasconi P. Learning long-term dependencies with gradient descent is difficult. IEEE Transactions on Neural Networks, 1994, 5(2):157-166.[doi:10.1109/72.279181
    [12] Vaswani A, Shazeer N, Parmar N, et al. Attention is all you need. Proceedings of the 31st International Conference on Neural Information Processing Systems. Long Beach, CA, USA. 2017. 5998-6008.
    [13] Niepert M, Ahmed M, Kutzkov K. Learning convolutional neural networks for graphs. Proceedings of the 33rd International Conference on International Conference on Machine Learning. New York City, NY, USA. 2016.2014-2023.
    [14] Allamanis M, Barr ET, Bird C, et al. Suggesting accurate method and class names. Proceedings of the 10th Joint Meeting on Foundations of Software Engineering. Bergamo, Italy. 2015. 38-49.
    [15] Iyer S, Konstas I, Cheung A, et al. Summarizing source code using a neural attention model. Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1:Long Papers). Berlin, Germany. 2016. 2073-2083.
    [16] Hu X, Li G, Xia X, et al. Summarizing source code with transferred api knowledge. Proceedings of the 27th International Joint Conference on Artificial Intelligence. Stockholm, Sweden. 2018. 2269-2275.
    [17] Li J, Wang Y, Lyu MR, et al. Code completion with neural attention and pointer networks. arXiv:1711.09573, 2017.
    [18] Liu C, Wang X, Shin R, et al. Neural code completion. Proceedings of International Conference on Learning Representations 2017 Conference Submission. Toulon, France. 2017. 1-14.
    [19] Sun ZY, Zhu QH, Xiong YF, et al. TreeGen:A tree-based transformer architecture for code generation. The 34th AAAI Conference on Artificial Intelligence (AAAI 2020), the 32nd Innovative Applications of Artificial Intelligence Conference (IAAI 2020), the 10th AAAI Symposium on Educational Advances in Artificial Intelligence (EAAI 2020). New York, NY, USA. 2020. 8984-8991.
    [20] Sun ZY, Zhu QH, Mou LL, et al. A grammar-based structural CNN decoder for code generation. Proceedings of the 33rd AAAI Conference on Artificial Intelligence (AAAI 2019), the Thirty-First Innovative Applications of Artificial Intelligence Conference (IAAI 2019), the Ninth AAAI Symposium on Educational Advances in Artificial Intelligence (EAAI 2019). Honolulu, HI, USA. 2019.7055-7062.
    [21] Allamanis M, Brockschmidt M, Khademi M. Learning to represent programs with graphs. arXiv:1711.00740, 2017.
    [22] Beck D, Haffari G, Cohn T. Graph-to-sequence learning using gated graph neural networks. Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics. 2018. 273-283.
    [23] Ba JL, Kiros JR, Hinton GE. Layer normalization. arXiv:1607.06450, 2016.
    [24] He KM, Zhang XY, Ren SQ, et al. Deep residual learning for image recognition. Proceedings of 2016 IEEE Conference on Computer Vision and Pattern Recognition. Las Vegas, NV, USA. 2016. 770-778.
    [25] See A, Liu PJ, Manning CD. Get to the point:Summarization with pointer-generator networks. Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics. Vanconver, BC, Canada. 2017. 1073-1083.
    [26] Lin CY. Rouge:A package for automatic evaluation of summaries. Proceedings of 2004 Workshop on Text Summarization Branches Out. Barcelona, Spain. 2004. 74-81.
    [27] Kingma DP, Ba J. Adam:A method for stochastic optimization. Proceedings of the 3rd International Conference on Learning Representations. San Diego, CA, USA. 2015.
    [28] Tai KS, Socher R, Manning CD. Improved semantic representations from tree-structured long short-term memory networks. Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing of the Asian Federation of Natural Language Processing. Beijing, China. 2015. 1556-1566.
    Cited by
    Comments
    Comments
    分享到微博
    Submit
Get Citation

王堃,李征,刘勇.基于图卷积神经网络的函数自动命名.计算机系统应用,2021,30(8):256-265

Copy
Share
Article Metrics
  • Abstract:999
  • PDF: 1737
  • HTML: 1522
  • Cited by: 0
History
  • Received:November 23,2020
  • Revised:December 22,2020
  • Online: August 03,2021
Article QR Code
You are the first990360Visitors
Copyright: Institute of Software, Chinese Academy of Sciences Beijing ICP No. 05046678-3
Address:4# South Fourth Street, Zhongguancun,Haidian, Beijing,Postal Code:100190
Phone:010-62661041 Fax: Email:csa (a) iscas.ac.cn
Technical Support:Beijing Qinyun Technology Development Co., Ltd.

Beijing Public Network Security No. 11040202500063