Overview on Graph-based Zero-shot Learning
CSTR:
Author:
  • Article
  • | |
  • Metrics
  • |
  • Reference [123]
  • |
  • Related [20]
  • | | |
  • Comments
    Abstract:

    Although the deep learning method has made a huge breakthrough in machine learning, it requires a large amount of manual work for data annotation. Limited by labor costs, however, many applications are expected to reason and judge the instance labels that have never been encountered before. For this reason, zero-shot learning (ZSL) came into being. As a natural data structure that represents the connection between things, the graph is currently drawing more and more attention in ZSL. Therefore, this study reviews the methods of graph-based ZSL systematically. Firstly, the definitions of ZSL and graph learning are outlined, and the ideas of existing solutions for ZSL are summarized. Secondly, the current ZSL methods are classified according to different utilization ways of graphs. Thirdly, the evaluation criteria and datasets concerning graph-based ZSL are discussed. Finally, this study also specifies the problems to be solved in further research on graph-based ZSL and predicts the possible directions of its future development.

    Reference
    [1] Wang W, Zheng VW, Yu H, et al. A survey of zero-shot learning: Settings, methods, and applications. ACM Transactions on Intelligent Systems and Technology, 2019, 10(2): 13
    [2] Farhadi A, Endres I, Hoiem D, et al. Describing objects by their attributes. 2009 IEEE Conference on Computer Vision and Pattern Recognition. Miami: IEEE, 2009. 1778–1785.
    [3] Lampert CH, Nickisch H, Harmeling S. Learning to detect unseen object classes by between-class attribute transfer. 2009 IEEE Conference on Computer Vision and Pattern Recognition. Miami: IEEE, 2009. 951–958.
    [4] Li X, Fang M, Chen B. Generalized zero-shot classification via iteratively generating and selecting unseen samples. Signal Processing: Image Communication, 2021, 92: 116115. [doi: 10.1016/j.image.2020.116115
    [5] Gan C, Lin M, Yang Y, et al. Exploring semantic inter-class relationships (SIR) for zero-shot action recognition. Proceedings of the 29th AAAI Conference on Artificial Intelligence. Austin: AAAI Press, 2015. 3769–3775.
    [6] Gan C, Yang Y, Zhu LC, et al. Recognizing an action using its name: A knowledge-based approach. International Journal of Computer Vision, 2016, 120(1): 61–77. [doi: 10.1007/s11263-016-0893-6
    [7] Liu JG, Kuipers B, Savarese S. Recognizing human actions by attributes. CVPR 2011. Colorado: IEEE, 2011. 3337–3344.
    [8] Su Y, Xing M, An SM, et al. VDARN: Video disentangling attentive relation network for few-shot and zero-shot action recognition. Ad Hoc Networks, 2021, 113: 102380. [doi: 10.1016/j.adhoc.2020.102380
    [9] Firat O, Sankaran B, Al-Onaizan Y, et al. Zero-resource translation with multi-lingual neural machine translation. arXiv: 1606.04164, 2016.
    [10] Johnson M, Schuster M, Le QV, et al. Google’s multilingual neural machine translation system: Enabling zero-shot translation. Transactions of the Association for Computational Linguistics, 2017, 5: 339–351. [doi: 10.1162/tacl_a_00065
    [11] Xian YQ, Schiele B, Akata Z. Zero-shot learning—The good, the bad and the ugly. Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Honolulu: IEEE, 2017. 3077–3086.
    [12] Fu YW, Xiang T, Jiang YG, et al. Recent advances in zero-shot recognition: Toward data-efficient understanding of visual content. IEEE Signal Processing Magazine, 2018, 35(1): 112–125. [doi: 10.1109/MSP.2017.2763441
    [13] 冀中, 汪浩然, 于云龙, 等. 零样本图像分类综述: 十年进展. 中国科学: 信息科学, 2019, 49(10): 1299–1320
    [14] Fu YW, Sigal L. Semi-supervised vocabulary-informed Learning. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Las Vegas: IEEE, 2016. 5337–5346.
    [15] Palatucci M, Pomerleau D, Hinton GE, et al. Zero-shot learning with semantic output codes. Proceedings of the 22nd International Conference on Neural Information Processing Systems. New York: Curran Associates Inc., 2009. 1410–1418.
    [16] Lampert CH, Nickisch H, Harmeling S. Attribute-based classification for zero-shot visual object categorization. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2014, 36(3): 453–465. [doi: 10.1109/Tpami.2013.140
    [17] Al-Halah Z, Tapaswi M, Stiefelhagen R. Recovering the missing link: Predicting class-attribute associations for unsupervised zero-shot learning. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Las Vegas: IEEE, 2016. 5975–5984.
    [18] Norouzi M, Mikolov T, Bengio S, et al. Zero-shot learning by convex combination of semantic embeddings. arXiv: 1312.5650, 2013.
    [19] Jayaraman D, Grauman K. Zero-shot recognition with unreliable attributes. Proceedings of the 27th International Conference on Neural Information Processing Systems. Cambridge: MIT Press, 2014. 3464–3472.
    [20] Kankuekul P, Kawewong A, Tangruamsub S, et al. Online incremental attribute-based zero-shot learning. 2012 IEEE Conference on Computer Vision and Pattern Recognition. Providence: IEEE, 2012. 3657–3664.
    [21] Larochelle H, Erhan D, Bengio Y. Zero-data learning of new tasks. Proceedings of the 23rd National Conference on Artificial Intelligence. Chicago: AAAI Press, 2008. 646–651.
    [22] Parikh D, Grauman K. Relative attributes. 2011 International Conference on Computer Vision. Barcelona: IEEE, 2011. 503–510.
    [23] Yu FX, Cao LL, Feris RS, et al. Designing category-level attributes for discriminative visual recognition. 2013 IEEE Conference on Computer Vision and Pattern Recognition. Portland: IEEE, 2013. 771–778.
    [24] Fu YW, Hospedales TM, Xiang T, et al. Transductive multi-view zero-shot learning. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2015, 37(11): 2332–2345. [doi: 10.1109/TPAMI.2015.2408354
    [25] Akata Z, Reed S, Walter D, et al. Evaluation of output embeddings for fine-grained image classification. 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Boston: IEEE, 2015. 2927–2936.
    [26] Ba JL, Swersky K, Fidler S, et al. Predicting deep zero-shot convolutional neural networks using textual descriptions. 2015 IEEE International Conference on Computer Vision (ICCV). Santiago: IEEE, 2015. 4247–4255.
    [27] Elhoseiny M, Saleh B, Elgammal A. Write a classifier: Zero-shot learning using purely textual descriptions. Proceedings of the 2013 IEEE International Conference on Computer Vision. Sydney: IEEE, 2013. 2584–2591.
    [28] Qiao RZ, Liu LQ, Shen CH, et al. Less is more: Zero-shot learning from online textual documents with noise suppression. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Las Vegas: IEEE, 2016. 2249–2257.
    [29] Mikolov T, Sutskever I, Chen K, et al. Distributed representations of words and phrases and their compositionality. Proceedings of the 26th International Conference on Neural Information Processing Systems. New York: Curran Associates Inc., 2013. 3111–3119.
    [30] Pennington J, Socher R, Manning CD. Glove: Global vectors for word representation. Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP). Doha: Association for Computational Linguistics, 2014. 1532–1543.
    [31] Frome A, Corrado GS, Shlens J, et al. DeViSE: A deep visual-semantic embedding model. Proceedings of the 26th International Conference on Neural Information Processing Systems. New York: Curran Associates Inc., 2013. 2121–2129.
    [32] Socher R, Ganjoo M, Manning CD, et al. Zero-shot learning through cross-modal transfer. Proceedings of the 26th International Conference on Neural Information Processing Systems. New York: Curran Associates Inc., 2013. 935–943.
    [33] Lazaridou A, Dinu G, Baroni M. Hubness and pollution: Delving into cross-space mapping for zero-shot learning. Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). Beijing: Association for Computational Linguistics, 2015. 270–280.
    [34] Shigeto Y, Suzuki I, Hara K, et al. Ridge regression, hubness, and zero-shot learning. Joint European Conference on Machine Learning and Knowledge Discovery in Databases. Porto: Springer, 2015. 135–151.
    [35] Zhang L, Xiang T, Gong SG. Learning a deep embedding model for zero-shot learning. 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Honolulu: IEEE, 2017. 3010–3019.
    [36] Changpinyo S, Chao WL, Sha F. Predicting visual exemplars of unseen classes for zero-shot learning. 2017 IEEE International Conference on Computer Vision (ICCV). Venice: IEEE, 2017. 3496–3505.
    [37] Zhang ZM, Saligrama V. Zero-shot learning via semantic similarity embedding. 2015 IEEE International Conference on Computer Vision (ICCV). Santiago: IEEE, 2015. 4166–4174.
    [38] Zhang ZM, Saligrama V. Zero-shot learning via joint latent similarity embedding. Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Las Vegas: IEEE, 2016. 6034–6042.
    [39] Xian YQ, Lorenz T, Schiele B, et al. Feature generating networks for zero-shot learning. Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Salt Lake City: IEEE, 2018. 5542–5551.
    [40] Gong MM, Zhang K, Liu TL, et al. Domain adaptation with conditional transferable components. Proceedings of the 33rd International Conference on International Conference on Machine Learning. New York: JMLR, 2016. 2839–2848.
    [41] Atzmon Y, Kreuk F, Shalit U, et al. A causal view of compositional zero-shot recognition. arXiv: 2006.14610, 2020.
    [42] Gilmer J, Schoenholz SS, Riley PF, et al. Neural message passing for quantum chemistry. arXiv: 1704.01212, 2017.
    [43] Li YJ, Vinyals O, Dyer C, et al. Learning deep generative models of graphs. arXiv: 1803.03324, 2018.
    [44] De Cao N, Kipf T. MolGAN: An implicit generative model for small molecular graphs. arXiv: 1805.11973, 2018.
    [45] Van Den Berg R, Kipf TN, Welling M. Graph convolutional matrix completion. arXiv: 1706.02263, 2017.
    [46] Ying R, He RN, Chen KF, et al. Graph convolutional neural networks for web-scale recommender systems. Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. London: ACM, 2018. 974–983.
    [47] Monti F, Bronstein MM, Bresson X. Geometric matrix completion with recurrent multi-graph neural networks. arXiv: 1704.06803, 2017.
    [48] Sze V, Chen YH, Yang TJ, et al. Efficient processing of deep neural networks: A tutorial and survey. Proceedings of the IEEE, 2017, 105(12): 2295–2329. [doi: 10.1109/JPROC.2017.2761740
    [49] Chami I, Abu-El-Haija S, Perozzi B, et al. Machine learning on graphs: A model and comprehensive taxonomy. arXiv: 2005.03675, 2020.
    [50] Deng J, Ding J, Jia YQ, et al. Large-scale object classification using label relation graphs. Proceedings of the 13th European Conference on Computer Vision. Zurich: Springer, 2014. 48–64.
    [51] Bordes A, Usunier N, Garcia-Durán A, et al. Translating embeddings for modeling multi-relational data. Proceedings of the 26th International Conference on Neural Information Processing Systems. New York: Curran Associates Inc., 2013. 2787–2795.
    [52] Wang HW, Zhang FZ, Xie X, et al. DKN: Deep knowledge-aware network for news recommendation. Proceedings of the 2018 World Wide Web Conference. Lyon: International World Wide Web Conferences Steering Committee, 2018. 1835–1844.
    [53] Wang HW, Zhang FZ, Wang JL, et al. RippleNet: Propagating user preferences on the knowledge graph for recommender systems. Proceedings of the 27th ACM International Conference on Information and Knowledge Management. Virtual Event: ACM, 2018. 417–426.
    [54] Yoon S, Dernoncourt F, Kim DS, et al. A compare-aggregate model with latent clustering for answer selection. Proceedings of the 28th ACM International Conference on Information and Knowledge Management. Virtual Event: ACM, 2019. 2093–2096.
    [55] Ma XY, Zhu QL, Zhou YL, et al. Improving question generation with sentence-level semantic matching and answer position inferring. Proceedings of the 35th AAAI Conference on Artificial Intelligence. Palo Alto: AAAI, 2020. 8464–8471.
    [56] Zhao Y, Gao S, Gallinari P, et al. Zero-shot embedding for unseen entities in knowledge graph. IEICE Transactions on Information and Systems, 2017, 100(7): 1440–1447. [doi: 10.1587/transinf.2016EDP7446
    [57] Wang ZK, Li LJ, Li QD, et al. Multimodal data enhanced representation learning for knowledge graphs. 2019 International Joint Conference on Neural Networks (IJCNN). Budapest: IEEE, 2019. 1–8.
    [58] Li CH, Xian XF, Ai XS, et al. Representation learning of knowledge graphs with embedding subspaces. Scientific Programming, 2020, 2020: 4741963. [doi: 10.1155/2020/4741963
    [59] Xie RB, Liu ZY, Jia J, et al. Representation learning of knowledge graphs with entity descriptions. Proceedings of the 30th AAAI Conference on Artificial Intelligence. Phoenix: AAAI Press, 2016. 2659–2665.
    [60] Ding JH, Ma SH, Jia WJ, et al. Jointly modeling structural and textual representation for knowledge graph completion in zero-shot scenario. Proceedings of the 2nd Asia-Pacific Web (APWeb) and Web-age Information Management (WAIM) Joint International Conference on Web and Big Data. Macao: Springer, 2018. 369–384.
    [61] Imrattanatrai W, Kato MP, Yoshikawa M. Identifying entity properties from text with zero-shot learning. Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval. Paris: ACM, 2019. 195–204.
    [62] Kumar S, Jat S, Saxena K, et al. Zero-shot word sense disambiguation using sense definition embeddings. Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Florence: Association for Computational Linguistics, 2019. 5670–5681.
    [63] Gao LL, Song JK, Shao JM, et al. Zero-shot image categorization by image correlation exploration. Proceedings of the 5th ACM on International Conference on Multimedia Retrieval. Lisboa: ACM, 2015. 487–490.
    [64] Li AX, Lu ZW, Wang LW, et al. Zero-shot scene classification for high spatial resolution remote sensing images. IEEE Transactions on Geoscience and Remote Sensing, 2017, 55(7): 4157–4167. [doi: 10.1109/Tgrs.2017.2689071
    [65] Fu ZY, Xiang T, Kodirov E, et al. Zero-shot learning on semantic class prototype graph. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2018, 40(8): 2009–2022. [doi: 10.1109/Tpami.2017.2737007
    [66] Castanon G, Chen YT, Zhang ZM, et al. Efficient activity retrieval through semantic graph queries. Proceedings of the 23rd ACM International Conference on Multimedia. Lisboa: ACM, 2015. 391–400.
    [67] Shen YM, Liu L, Shen FM, et al. Zero-shot sketch-image hashing. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Salt Lake City: IEEE, 2018. 3598–3607.
    [68] Cai HY, Zheng VW, Chang KCC. A comprehensive survey of graph embedding: Problems, techniques, and applications. IEEE Transactions on Knowledge and Data Engineering, 2018, 30(9): 1616–1637. [doi: 10.1109/Tkde.2018.2807452
    [69] Ullmann JR. An algorithm for subgraph isomorphism. Journal of the ACM, 1976, 23(1): 31–42. [doi: 10.1145/321921.321925
    [70] Boykov YY, Jolly MP. Interactive graph cuts for optimal boundary & region segmentation of objects in N-D images. Proceedings Eighth IEEE International Conference on Computer Vision. ICCV 2001. Vancouver: IEEE, 2001. 105–112.
    [71] Chang DS, Cho GH, Choi YS. Similarity-based calibration method for zero-shot recognition in multi-object scenes. Proceedings of the 35th Annual ACM Symposium on Applied Computing. Brno: ACM, 2020. 1096–1103.
    [72] Chang DS, Cho GH, Choi YS. Zero-shot recognition enhancement by distance-weighted contextual inference. Applied Sciences, 2020, 10(20): 7234. [doi: 10.3390/app10207234
    [73] Ding N, Deng J, Murphy KP, et al. Probabilistic label relation graphs with Ising models. 2015 IEEE International Conference on Computer Vision (ICCV). Santiago: IEEE, 2015. 1161–1169.
    [74] Kadowaki T, Nishimori H. Quantum annealing in the transverse Ising model. Physical Review E, 1998, 58(5): 5355–5363. [doi: 10.1103/PhysRevE.58.5355
    [75] Kordumova S, Mensink T, Snoek CGM. Pooling objects for recognizing scenes without examples. Proceedings of the 2016 ACM on International Conference on Multimedia Retrieval. Lisboa: ACM, 2016. 143–150.
    [76] Deutsch S, Kolouri S, Kim K, et al. Zero shot learning via multi-scale manifold regularization. 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Honolulu: IEEE, 2017. 5292–5299.
    [77] Hammond DK, Vandergheynst P, Gribonval R. Wavelets on graphs via spectral graph theory. Applied and Computational Harmonic Analysis, 2011, 30(2): 129–150. [doi: 10.1016/j.acha.2010.04.005
    [78] Zhong FM, Chen ZK, Min GY. An exploration of cross-modal retrieval for unseen concepts. Proceedings of the 24th International Conference on Database Systems for Advanced Applications. Chiang Mai: Springer, 2019. 20–35.
    [79] Long Y, Guan Y, Shao L. Generic compact representation through visual-semantic ambiguity removal. Pattern Recognition Letters, 2019, 117: 186–192. [doi: 10.1016/j.patrec.2018.04.024
    [80] Ding ZM, Liu HF. Marginalized latent semantic encoder for zero-shot learning. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Long Beach: IEEE, 2019. 6184–6192.
    [81] K?nig D. Gráfok és mátrixok. Matematikai és Fizikai Lapok, 1931, 38: 116–119
    [82] Changpinyo S, Chao WL, Gong BQ, et al. Synthesized classifiers for zero-shot learning. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Las Vegas: IEEE, 2016. 5327–5336.
    [83] Chen Y, Xiong YH, Gao X, et al. Structurally constrained correlation transfer for zero-shot learning. 2018 IEEE Visual Communications and Image Processing (VCIP). Taichung: IEEE, 2018. 1–4.
    [84] Chen BH, Deng WH. Hybrid-attention based decoupled metric learning for zero-shot image retrieval. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Long Beach: IEEE, 2019. 2745–2754.
    [85] Chen BH, Deng WH. Energy confused adversarial metric learning for zero-shot image retrieval and clustering. Proceedings of the 33rd AAAI Conference on Artificial Intelligence. Palo Alto: AAAI, 2019. 8134–8141.
    [86] Chen LH, Qu YR, Wang ZH, et al. Sampled in pairs and driven by text: A new graph embedding framework. The World Wide Web Conference. San Francisco: ACM, 2019. 2644–2651.
    [87] Li YN, Hu HH, Wang DH. Learning visually aligned semantic graph for cross-modal manifold matching. 2019 IEEE International Conference on Image Processing (ICIP). Taipei: IEEE, 2019. 3412–3416.
    [88] Ayyalasomayajula KR, Brun A. Historical document binarization combining semantic labeling and graph cuts. Proceedings of the 20th Scandinavian Conference on Image Analysis. Troms: Springer, 2017. 386–396.
    [89] Long Y, Shao L. Describing unseen classes by exemplars: Zero-shot learning using grouped simile ensemble. 2017 IEEE Winter Conference on Applications of Computer Vision (WACV). Santa Rosa: IEEE, 2017. 907–915.
    [90] Huang S, Elhoseiny M, Elgammal A, et al. Learning hypergraph-regularized attribute predictors. 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Boston: IEEE, 2015. 409–417.
    [91] Kampffmeyer M, Chen YB, Liang XD, et al. Rethinking knowledge graph propagation for zero-shot learning. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Long Beach: IEEE, 2019. 11479–11488.
    [92] Gao JY, Zhang TZ, Xu CS. I know the relationships: Zero-shot action recognition via two-stream graph convolutional networks and knowledge graphs. Proceedings of the 33rd AAAI Conference on Artificial Intelligence. Palo Alto: AAAI, 2019. 8303–8311.
    [93] Wang FX, Liu J, Zhang SW, et al. Inductive zero-shot image annotation via embedding graph. IEEE Access, 2019, 7: 107816–107830. [doi: 10.1109/Access.2019.2925383
    [94] Bucher M, Vu TH, Cord M, et al. Zero-shot semantic segmentation. Advances in Neural Information Processing Systems, 2019, 32: 468–479.
    [95] Xie GS, Liu L, Zhu F, et al. Region graph embedding network for zero-shot learning. Proceedings of the 16th European Conference on Computer Vision. Glasgow: Springer, 2020. 562–580.
    [96] Wang XL, Ye YF, Gupta A. Zero-shot recognition via semantic embeddings and knowledge graphs. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Salt Lake City: IEEE, 2018. 6857–6866.
    [97] Yan CX, Zheng QH, Chang XJ, et al. Semantics-preserving graph propagation for zero-shot object detection. IEEE Transactions on Image Processing, 2020, 29: 8163–8176. [doi: 10.1109/Tip.2020.3011807
    [98] Lee CW, Fang W, Yeh CK, et al. Multi-label zero-shot learning with structured knowledge graphs. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Salt Lake City: IEEE, 2018. 1576–1585.
    [99] Yedidia JS, Freeman WT, Weiss Y. Understanding belief propagation and its generalizations. Exploring Artificial Intelligence in the New Millennium. San Francisco: Morgan Kaufmann Publishers Inc., 2003. 239–269.
    [100] Xiao B, Du YJ, Wu QMJ, et al. A fast hybrid model for large-scale zero-shot image recognition based on knowledge graphs. IEEE Access, 2019, 7: 119309–119318. [doi: 10.1109/Access.2019.2935175
    [101] Zhang CR, Lyu XQ, Tang Z. TGG: Transferable graph generation for zero-shot and few-shot learning. Proceedings of the 27th ACM International Conference on Multimedia. Lisboa: ACM, 2019. 1641–1649.
    [102] Wang WG, Lu XK, Shen JB, et al. Zero-shot video object segmentation via attentive graph neural networks. 2019 IEEE/CVF International Conference on Computer Vision (ICCV). Seoul: IEEE, 2019. 9235–9244.
    [103] Zhang CY, Tian YL, Guo XJ, et al. DAAL: Deep activation-based attribute learning for action recognition in depth videos. Computer Vision and Image Understanding, 2018, 167: 37–49. [doi: 10.1016/j.cviu.2017.11.008
    [104] Bapna A, Tür G, Hakkani-Tür D, et al. Towards zero-shot frame semantic parsing for domain scaling. 18th Annual Conference of the International Speech Communication Association. Stockholm: ISCA, 2017. 2476–2480.
    [105] Deng J, Dong W, Socher R, et al. ImageNet: A large-scale hierarchical image database. 2009 IEEE Conference on Computer Vision and Pattern Recognition. Miami: IEEE, 2009. 248–255.
    [106] Everingham M, van Gool L, Williams CKI, et al. The PASCAL visual object classes (VOC) challenge. International Journal of Computer Vision, 2010, 88(2): 303–338. [doi: 10.1007/s11263-009-0275-4
    [107] Russakovsky O, Deng J, Su H, et al. ImageNet large scale visual recognition challenge. International Journal of Computer Vision, 2015, 115(3): 211–252. [doi: 10.1007/s11263-015-0816-y
    [108] Lin TY, Maire M, Belongie S, et al. Microsoft COCO: Common objects in context. Proceedings of the 13th European Conference on Computer Vision. Zurich: Springer, 2014. 740–755.
    [109] Lazaridou A, Bruni E, Baroni M. Is this a wampimuk? Cross-modal mapping between distributional semantics and the visual world. Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics. Baltimore: Association for Computational Linguistics, 2014. 1403–1414.
    [110] Wah C, Branson S, Welinder P, et al. The Caltech-UCSD birds-200-2011 dataset. 2011.
    [111] Patterson G, Hays J. SUN attribute database: Discovering, annotating, and recognizing scene attributes. 2012 IEEE Conference on Computer Vision and Pattern Recognition. Providence: IEEE, 2012. 2751–2758.
    [112] Zhou B, Khosla A, Lapedriza A, et al. Places2: A large-scale database for scene understanding. http://places2. csail. mit. edu. (2017-06-21)[2021-07-22].
    [113] Chua TS, Tang JH, Hong RC, et al. NUS-WIDE: A real-world web image database from National University of Singapore. Proceedings of the ACM International Conference on Image and Video Retrieval. Santorini: ACM, 2009. 48.
    [114] Grubinger M, Clough P, Müller H, et al. The IAPR TC-12 benchmark: A new evaluation resource for visual information systems. International Workshop ontoImage. Genoa: OntoImage 2006, 2006. 13–23.
    [115] Duygulu P, Barnard K, de Freitas JFG, et al. Object recognition as machine translation: Learning a lexicon for a fixed image vocabulary. Proceedings of the 7th European Conference on Computer Vision. Copenhagen: Springer, 2002. 97–112.
    [116] Niebles JC, Chen CW, Li FF. Modeling temporal structure of decomposable motion segments for activity classification. Proceedings of the 11th European Conference on Computer Vision. Heraklion: Springer, 2010. 392–405.
    [117] Kuehne H, Jhuang H, Garrote E, et al. HMDB: A large video database for human motion recognition. 2011 International Conference on Computer Vision. Barcelona: IEEE, 2011. 2556–2563.
    [118] Soomro K, Zamir AR, Shah M. UCF101: A dataset of 101 human actions classes from videos in the wild. arXiv: 1212.0402, 2012.
    [119] Palmer M, Fellbaum C, Cotton S, et al. English tasks: All-words and verb lexical sample. Proceedings of SENSEVAL-2 2nd International Workshop on Evaluating Word Sense Disambiguation Systems. Toulouse: Association for Computational Linguistics, 2001. 21–24.
    [120] Snyder B, Palmer M. The English all-words task. Proceedings of SENSEVAL-3, the 3rd International Workshop on the Evaluation of Systems for the Semantic Analysis of Text. Barcelona: Association for Computational Linguistics, 2004. 41–43.
    [121] Navigli R, Jurgens D, Vannella D. Semeval-2013 task 12: Multilingual word sense disambiguation. 2nd Joint Conference on Lexical and Computational Semantics (*SEM), Volume 2: Proceedings of the 7th International Workshop on Semantic Evaluation (SemEval 2013). Atlanta: Association for Computational Linguistics, 2013. 222–231.
    [122] Moro A, Navigli R. Semeval-2015 task 13: Multilingual all-words sense disambiguation and entity linking. Proceedings of the 9th International Workshop on Semantic Evaluation (SemEval 2015). Denver: Association for Computational Linguistics, 2015. 288–297.
    [123] Riedel S, Yao LM, McCallum A. Modeling relations and their mentions without labeled text. Joint European Conference on Machine Learning and Knowledge Discovery in Databases. Barcelona: Springer, 2010. 148–163.
    Cited by
    Comments
    Comments
    分享到微博
    Submit
Get Citation

支瑞聪,万菲,张德政.零样本图学习综述.计算机系统应用,2022,31(5):1-20

Copy
Share
Article Metrics
  • Abstract:7422
  • PDF: 7918
  • HTML: 4835
  • Cited by: 0
History
  • Received:July 14,2021
  • Revised:August 18,2021
  • Online: February 21,2022
Article QR Code
You are the first992235Visitors
Copyright: Institute of Software, Chinese Academy of Sciences Beijing ICP No. 05046678-3
Address:4# South Fourth Street, Zhongguancun,Haidian, Beijing,Postal Code:100190
Phone:010-62661041 Fax: Email:csa (a) iscas.ac.cn
Technical Support:Beijing Qinyun Technology Development Co., Ltd.

Beijing Public Network Security No. 11040202500063