Discrimination Method of Terrorism Audio Based on Transfer Learning
CSTR:
Author:
  • Article
  • | |
  • Metrics
  • |
  • Reference [10]
  • |
  • Related [20]
  • |
  • Cited by
  • | |
  • Comments
    Abstract:

    This article intercepts the horror audio clips from the network and movies to build terrorism audio dataset. However, the source of the horror audio is limited, whereas the convolutional neural network depends on a large amount of data. To this end, the transfer learning technology is performed in the discrimination of the terrorism audio. Firstly, pre-train the network by using the public TUT acoustic scenes dataset, and then retain the model weight and transfer the neural network to the discrimination of terrorism audio. Finally, add more layers after the fine-tune network to utilize more audio information, the structure of the added layers is similar to the residual network. The experimental results indicate that the average discriminant rate of the transfer learning method is 3.97% higher than that of the non-transfer learning method, which effectively solves the training problem caused by small audio dataset in the study of terrorism audio discrimination, and the average discriminant rate of the improved transfer learning network has increased by 1.01%, finally reaches the discriminant rate of 96.97%.

    Reference
    [1] Giannakopoulos T, Kosmopoulos D, Aristidou A, et al. Violence content classification using audio features. Proceedings of the 4th Helenic Conference on Artificial Intelligence. Crete, Greece. 2006. 502-507.
    [2] 冯佳军. 暴力音频场景分类技术研究与系统实现[硕士学位论文]. 哈尔滨:哈尔滨工业大学, 2016.
    [3] Pikrakis A, Giannakopoulos T, Theodoridis S. Gunshot detection in audio streams from movies by means of dynamic programming and Bayesian networks. Proceedings of 2008 IEEE International Conference on Acoustics, Speech and Signal Processing. Las Vegas, NV, USA. 2008. 21-24.
    [4] 梁家欣, 李海峰, 马琳. 基于张量模型的暴力音频检测研究. 智能计算机与应用, 2016, 6(1):108-111.[doi:10.3969/j.issn.2095-2163.2016.01.030
    [5] 王红斌, 沈强, 线岩团. 融合迁移学习的中文命名实体识别. 小型微型计算机系统, 2017, 38(2):346-351
    [6] 孙超, 吕俊伟, 刘峰, 等. 基于迁移学习的红外图像超分辨率方法研究. 激光与红外, 2017, 47(12):1559-1564.[doi:10.3969/j.issn.1001-5078.2017.12.019
    [7] Piczak K J. The details that matter:Frequency resolution of spectrograms in acoustic scene classification. Proceedings of the Detection and Classification of Acoustic Scenes and Events. Munich, Germany. 2017. 1-5.
    [8] 周成豪. 基于概率潜在语义分析的音频场景识别方法[硕士学位论文]. 哈尔滨:哈尔滨工业大学, 2013.
    [9] 关胤. 基于残差网络迁移学习的花卉识别系统. 计算机工程与应用, 2019, 55(1):174-179
    [10] 张振焕, 周彩兰, 梁媛. 基于残差的优化卷积神经网络服装分类算法. 计算机工程与科学, 2018, 40(2):354-360.[doi:10.3969/j.issn.1007-130X.2018.02.023
    Cited by
    Comments
    Comments
    分享到微博
    Submit
Get Citation

胡鑫旭,周欣,何小海,熊淑华,王正勇.基于迁移学习的暴恐音频判别方法.计算机系统应用,2019,28(11):147-152

Copy
Share
Article Metrics
  • Abstract:1949
  • PDF: 2121
  • HTML: 1822
  • Cited by: 0
History
  • Received:April 11,2019
  • Revised:May 08,2019
  • Online: November 08,2019
  • Published: November 15,2019
Article QR Code
You are the first992180Visitors
Copyright: Institute of Software, Chinese Academy of Sciences Beijing ICP No. 05046678-3
Address:4# South Fourth Street, Zhongguancun,Haidian, Beijing,Postal Code:100190
Phone:010-62661041 Fax: Email:csa (a) iscas.ac.cn
Technical Support:Beijing Qinyun Technology Development Co., Ltd.

Beijing Public Network Security No. 11040202500063