###
计算机系统应用英文版:2019,28(11):101-106
本文二维码信息
码上扫一扫!
基于ZYNQ和CNN模型的服装识别系统
(中国科学技术大学 信息科学技术学院, 合肥 230027)
Fashion Recognition System Based on ZYNQ and CNN Model
(School of Information Science and Technology, University of Science and Technology of China, Hefei 230027, China)
摘要
图/表
参考文献
相似文献
本文已被:浏览 1250次   下载 2435
Received:April 19, 2019    Revised:May 16, 2019
中文摘要: 商品检索是电商行业智能化发展的一个重要的问题.本设计实现了基于ZYNQ和CNN模型的服装识别系统.利用TensorFlow训练自定义网络,定点化处理权重参数.利用ZYNQ器件的ARM+FPGA软硬件协同的特点搭建系统,使用ARM端OpenCV进行图像预处理,FPGA端CNN IP进行实时识别.ARM与FPGA之间实现了权重可重加载结构,无需修改FPGA硬件而实现在线升级.系统采用fashion-minist数据集作为网络训练样本,根据系统资源配置CNN IP的加速引擎的数量来提高卷积运算的并行性.实验表明,本系统针对电商平台下的图片能够实时准确识别和显示,准确率达92.39%.在100 MHz工作频率下,图像处理速度每帧可达到1.361 ms,功耗仅为0.53 W.
中文关键词: ZYNQ  CNN  服装识别  软硬件协同
Abstract:Commodity retrieval is fundamental to the intelligent development of the e-commerce industry. This study is concerned with a fashion recognition system which bases on ZYNQ and CNN model, trains the custom network with TensorFlow and processes the weighs by using fixed-point calculations. This system applies ARM + FPGA software and hardware coordination method of the ZYNQ device to construct its framework, ARM to preprocess image by OpenCV and the CNN IP of FPGA to synchronously recognize the image. A weight-reloadable structure is implemented between ARM and FPGA, so online upgrade could be realized without modifying the FPGA hardware. The system uses fashion-minist datasets as the sample in the network training and improves the parallelism of convolution operations by increasing the acceleration engines of CNN IP. According to the experiment, the system realizes accurate and real-time identification and display for actual pictures of the e-commerce platform and its accuracy reaches 92.39%. The image processing speed can reach 1.261 ms per frame and the power consumption is only 0.53 W at 100 MHz.
文章编号:     中图分类号:    文献标志码:
基金项目:国家自然科学基金(61874102)
引用文本:
熊伟,黄鲁.基于ZYNQ和CNN模型的服装识别系统.计算机系统应用,2019,28(11):101-106
XIONG Wei,HUANG Lu.Fashion Recognition System Based on ZYNQ and CNN Model.COMPUTER SYSTEMS APPLICATIONS,2019,28(11):101-106