Home > Papers

 
 
An attention-enhanced neural network with distillation training for barcode detection
Wang Zijian,Zhou Xiaoguang *
Beijing University of Posts and Telecommunications,Schoold of Modern Posts, Beijing 100876;Beijing University of Posts and Telecommunications,Schoold of Modern Posts, Beijing 100876
*Correspondence author
#Submitted by
Subject:
Funding: none
Opened online:12 April 2023
Accepted by: none
Citation: Wang Zijian,Zhou Xiaoguang.An attention-enhanced neural network with distillation training for barcode detection[OL]. [12 April 2023] http://en.paper.edu.cn/en_releasepaper/content/4759920
 
 
Barcodes have played an essential role in our daily life. Localizing, or detecting them in real scenes in a fast and robust way has many practical applications. Recently, some deep learning-based methods have shown great potential in object detection. However, because barcodes are placed at any angle, vertical bounding boxes cannot sufficiently capture accurate orientation and scale information. In this paper, we propose a barcode detector that performs dense prediction to accurately locate the position of pixels belonging to the barcode region. For better detection performance, we design a spatial attention module to integrate global information adaptively, which can be easily plugged into the prediction backbone. Meanwhile, we employ the knowledge distillation training strategy to train a small student network with the help of a heavy teacher network. Extensive experiment results demonstrate that our method can perform real-time speed on CPU environments and locate barcodes in images with complex scenes.
Keywords: Computer vision; Image process; Barcode detection
 
 
 

For this paper

  • PDF (0B)
  • ● Revision 0   
  • ● Print this paper
  • ● Recommend this paper to a friend
  • ● Add to my favorite list

    Saved Papers

    Please enter a name for this paper to be shown in your personalized Saved Papers list

Tags

Add yours

Related Papers

Statistics

PDF Downloaded 7
Bookmarked 0
Recommend 0
Comments Array
Submit your papers