|Table of Contents|

Multi-Attention Mechanism of Mask Wearing Detection Network(PDF)

南京师范大学学报(工程技术版)[ISSN:1006-6977/CN:61-1281/TN]

Issue:
2021年01期
Page:
23-29
Research Field:
计算机科学与技术
Publishing date:

Info

Title:
Multi-Attention Mechanism of Mask Wearing Detection Network
Author(s):
Yu Axiang1Li Chengrun1Yu Shuyi1Li Hongjun12
(1.School of Information Science and Technology,Nantong University,Nantong 226019,China)(2.State Key Laboratory for Novel Software Technology,Nanjing University,Nanjing 210023,China)
Keywords:
mask wearing testmulti-attention mechanismfeature of the miningsoft-NMS
PACS:
TP391
DOI:
10.3969/j.issn.1672-1292.2021.01.004
Abstract:
A mask in public places can effectively control the transmission of the coronavirus. To this end,a mask wearing detection model is proposed. The model introduces a multi-attention mechanism to improve the network feature mining ability and uses soft-NMS methods to eliminate redundant target detection boxes. A simulation experiment is conducted on a public database. The average accuracy of the proposed face mask wearing detection reaches 93.81%,and the frame rate reaches 11.8 fps. The experimental results show that the model can effectively detect the face mask wearing.

References:

[1] 赵文明,宋述慧,陈梅丽,等. 2019 新型冠状病毒信息库[J]. 遗传,2020,42(2):212-221.
[2]白浪,王铭,唐小琼,等. 对新型冠状病毒肺炎诊疗中的热点问题的思考[J]. 华西医学,2020,35(2):125-131.
[3]GIRSHICK R,DONAHUE J,DARRELL T,et al. Rich feature hierarchies for accurate object detection and semantic segmentation[C]//IEEE Conference on Computer Vision and Pattern Recognition(CVPR). Columbus,USA:IEEE Computer Society,2014:580-587.
[4]GIRSHICK R. Fast R-CNN[C]//IEEE International Conference on Computer Vision(CVPR). Santiago,Chile:IEEE Computer Society,2015:1440-1448.
[5]REN S Q,HE K M,GIRSHICK R,et al. Faster R-CNN:towards real-time object detection with region proposal networks[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence,2017,39(6):1137-1149.
[6]REDMON J,DIVVALA S,GIRSHICK R,et al. You only look once:unified,real-time object detection[C]//IEEE Conference on Computer Vision and Pattern Recognition(CVPR). Las Vegas,USA:IEEE Computer Society,2016:779-788.
[7]LIU W,ANGUELOV D,ERHAN D,et al. SSD:single shot MultiBox detector[C]//European Conference on Computer Vision(ECCV). Amsterdam,Netherlands:ECCV,2016:21-37.
[8]REDMON J,FARHADI A. YOLO9000:better,faster,stronger[C]//IEEE Conference on Computer Vision and Pattern Recognition(CVPR). Honolulu,USA:IEEE Computer Society,2017:6517-6525.
[9]REDMON J,FARHAD A. YOLOv3:An incremental improvement[EB/OL]. [2020-08-08]. https://arxiv.org/abs/1804.02767.
[10]LIN T,GOYAL P,GIRSHICK R,et al. Focal loss for dense object detection[C]//IEEE Conference on Computer Vision and Pattern Recognition(CVPR). Honolulu,USA:IEEE Computer Society,2017:2999-3007.
[11]TAN M,LE Q V. EfficientNet:rethinking model scaling for convolutional neural networks[C]//International Conference on Machine Learning(ICML). California,USA:IMLS,2019:6105-6114.
[12]TAN M,PANG R,LE Q V,et al. EfficientDet:scalableand efficient object detection[C]//IEEE Conference on Computer Vision and Pattern Recognition(CVPR). Seattle,USA:IEEE Computer Society,2020:10781-10790.
[13]HE K M,ZHANG X Y,REN S Q,et al. Deep residual learning for image recognition[C]//IEEE Conference on Computer Vision and Pattern Recognition(CVPR). Las Vegas,USA:IEEE Computer Society,2016:770-778.
[14]SIMONYAN K,ZISSERMAN A. Very deep convolutional networks for large-scale image recognition[C]//International Conference of Learning Representation. San Diego,USA,2015.
[15]SANDLER M,HOWARD A,ZHU M,et al. MobileNetV2:inverted residuals and linear bottlenecks[C]//IEEE Conference on Computer Vision and Pattern Recognition(CVPR). Salt Lake City,USA:IEEE Computer Society,2018:4510-4520.
[16]LIU S,QI L,QI H,et al. Path aggregation network for instance segmentation[C]//IEEE Conference on Computer Vision and Pattern Recognition(CVPR). Salt Lake City,USA:IEEE Computer Society,2018:8759-8768.
[17]GHIASI G,LIN T,LE Q V,et al. NAS-FPN:learning scalable feature pyramid architecture for object detection[C]//IEEE Conference on Computer Vision and Pattern Recognition(CVPR). Long Beach,USA:IEEE Computer Society,2019:7036-7045.
[18]石磊,王毅,成颖,等. 自然语言处理中的注意力机制研究综述[J]. 数据分析与知识发现,2020,4(5):1-14.
[19]王文冠,沈建冰,贾云得. 视觉注意力检测综述[J]. 软件学报,2019,30(2):416-439.
[20]HU J,SHEN L,SUN G. Squeeze-and-excitation networks[C]//IEEE Conference on Computer Vision and Pattern Recognition(CVPR). Salt Lake City,USA:IEEE Computer Society,2018:7132-7141.
[21]BODLA N,SINGH B,CHELLAPPA R,et al. Soft-NMS—Improving object detection with one line of code[C]//IEEE International Conference on Computer Vision(ICCV). Venice,Italy:IEEE Computer Society,2017:5562-5570.
[22]GE S,LI J,YE Q,et al. Detecting masked faces in the wild with LLE-CNNs[C]//IEEE Conference on Computer Vision and Pattern Recognition(CVPR). Honolulu,USA:IEEE Computer Society,2017:426-434.
[23]YANG S,LUO P,LOY C C,et al. WIDER FACE:a face detection benchmark[C]//IEEE Conference on Computer Vision and Pattern Recognition(CVPR). Las Vegas,USA:IEEE Computer Society,2016:5525-5533.
[24]KINGMA D P,BA J. Adam:a method for stochastic optimization[C]//The 3rd International Conference for Learning Representations. San Diego,USA,2015.
[25]GLOYOT X,BENGIO Y. Understanding the difficulty of training deep feed forward neural networks[J]. Journal of Machine Learning Research,2010,9:249-250.

Memo

Memo:
-
Last Update: 2021-03-15