Computer Science ›› 2020, Vol. 47 ›› Issue (11A): 258-265.doi: 10.11896/jsjkx.191200115

• Computer Graphics & Multimedia • Previous Articles     Next Articles

MACTEN:Novel Large Scale Cloth Texture Classification Architecture

LI Hao-xiang, LI Hao-jun   

  1. College of Education Science and Technology,Zhejiang University of Technology,Hangzhou 310023,China
  • Online:2020-11-15 Published:2020-11-17
  • About author:LI Hao-xiang,born in 1995,master.His main research interests include computervision and deep learning.
    LI Hao-jun,born in 1977,doctor,professor.His main research interests include intelligent computing and intelligent learning.

Abstract: The miscellany of fabric and the complexity of its texture have been viewed as enormous challenges to distinguish artificially.The merging multi-scale attention co-occurrence representation's residual texture encoding network(MACTEN) has been proposed with the introduction of the deep learning technology.And based on that,the large-scale fabric classification system on the web has been carried out.The MACTEN mainly composed of attention co-occurrence representation module (ACM) and improved residual coding module (REM),as well as multi-scale texture coding fusion module (MTEM).In this work,the mechanism of attention has been implemented into ACM to deal with different types of clothes,which adaptively adjusts the weight of texture co-occurrence features,and optimizes the joint distribution of co-occurrence features by expanding the co-occurrence domain to form more refined texture co-occurrence features.Moreover,the improved residual coding,including global texture information of spatial invariance,has been obtained with introduction of dictionary learning method into REM,which can solve the problem of disordered representation of cloth texture effectively.Finally,MTEM combined multiple scale attention texture co-occurrence features and cascaded residual texture coding as descriptors,can represent different shape and size of disordered fabric texture.On self-building cloth dataset,MACTEN has exhibited better performance than other baseline algorithms.Furthermore,the experimental results of KTHTIPS,FMD and DTD datasets show that MACTEN can be generalized as a general texture classification algorithm.

Key words: Cloth classification, Co-occurrence feature, Deep learning, Feature fusion, Texture descriptors

CLC Number: 

  • TP391
[1] KRIZHEVSKY A,SUTSKEVER I,HINTON G.Imagenetclassification with deep convolutional neural networks [J].Communications of the ACM,2017,60(6):84-90.
[2] JEGOU H,DOUZE M,SCHMID C,et al.Aggregating local descriptors into a compact image representation[C]//IEEE Conference on Computer Vision and Pattern Recognition.San Francisco:IEEE,2010:3304-3311.
[3] LI H X.Application of machine learning in classification ofstudent course evaluation [J].Journal of Zhejiang Shu Ren University,2019,19(3):5-11.
[4] HARALICK R,SHANMUGAM K,DINSTEIN I.Textural features for image classification [J].IEEE Trans on Systems,Man,and Cybernetics,1973,6:610-621.
[5] PICHLER O,TEUNER A,HOSTICKA B.A comparison oftexture feature extraction using adaptive Gabor filtering,pyramidal and tree structured wavelet transforms [J].Pattern Recognition,1996,29(5):733-742.
[6] OJALA T,PIETIKAINEN M,HARWOOD D.A comparative study of texture measures with classification based on featured distributions [J].Pattern Recognition,1996,29(1):51-59.
[7] OJALA T,PIETIKAINEN M,MAENPAA T.Multiresolution gray-scale and rotation invariant texture classification with local binary patterns [J].IEEE Transactions on Pattern Analysis and Machine Intelligence,2002,24(7):971-987.
[8] KIAPOUR M H,XU F H,SVETLANA L,et al.Where to buy it:matching street clothing photos in online shops [C]//IEEE International Conference on Computer Vision.Santiago:IEEE,2015:3343-3351.
[9] LIU Z,LUO P,QIU S,et al.DeepFashion:powering robustclothes recognition and retrieval with rich annotations [C]//IEEE Conference on Computer Vision and Pattern Recognition.Seattle:CVPR,2016:1096-1104.
[10] WANG Z,GU Y,ZHANG Y,et al.Clothing retrieval with visual attention model [C]//IEEE International Conference on Visual Communications and Image Processing.Petersburg:VCIP,2017:1-4.
[11] CORBIEREC,BENYOUNES H,RAME A,et al.Leveragingweakly annotated data for fashion image retrieval and label prediction [C]//IEEE International Conference on Computer Vision Workshops.Venice:ICCVW,2017:2268-2274.
[12] GODI M,JOPPI C,GIACHETTI A,et al.Texel-att:representing and classifying element-based textures by attributes[C]//British Machine Vision Conference.British:BMVC,2019.
[13] JOPPI C,GODI M,GIACHETTI A,et al.Texture retrieval in the wild through detection-based attributes [C]//International Conference on Image Analysis and Processing.Trento:ICIAP,2019:522-533.
[14] KAIMING H,GEORGIA G,PIOTR D,et al.Mask r-cnn [C]//International Conference on Computer Vision.Venice:ICCV,2017:2980-2988.
[15] FURKAN K,BARIŞ O,FURKAN K.Fashion Image retrievalwith capsule networks [C]//IEEE International Conference on Computer Vision.Seoul:ICCV,2019.
[16] SABOUR S,FROSST N,HINTON G E.Dynamic routing between capsules [C]//Advances in Neural Information Processing Systems.Long Beach:NIPS,2017.
[17] CIMPOI M,MAJI S,KOKKINOS I,et al.Deep Filter Banks for texture recognition,description,and segmentation [J].International Journal of Computer Vision,2016,118(1):65-94.
[18] CIMPOI M,MAJI S,KOKKINOS I,et al.Describing textures in the wild [C]//IEEE Conference on Computer Vision and Pattern Recognition.Columbus:IEEE,2014:3606-3613.
[19] GONG Y,WANG L,GUO R,et al.Multi-scale orderless pooling of deep convolutional activation features[C]//European Conference on Computer Vision.Zurich:ECCV,2014:392-407.
[20] ZHANG H,XUE J,DANA K.Deep TEN:texture encoding network [C]//IEEE Conference on Computer Vision and Pattern Recognition.Honolulu:IEEE,2017:2896-2905.
[21] ZHU K,ZHAI W,ZHA Z J,et al.One-shot texture retrieval with global context metric [C]//International Joint Conference on Artificial Intelligence.Macao:IJCAI,2019:4461-4467.
[22] FREEMAN W T,TENENBAUM J B.Learning bilinear models for two-factor problems in vision [C]//IEEE Conference on Computer Vision and Pattern Recognition.San Juan:CVPR,1997:554-560.
[23] GATYS L A,ECKER A S,BETHGE M.Texture Synthesis Using Convolutional Neural Networks [C]//Conference on Neural Information Processing Systems.Montreal:NIPS,2015:262-270.
[24] LIN T Y,MAJI S.Visualizing and understanding deep texture representations [C]//IEEE Conference on Computer Vision and Pattern Recognition.Seattle:IEEE,2016:2791-2799.
[25] GATYS L A,ROYCHOWDHURY A,MAJI S.Bilinear cnnmodels for fine-grained visual recognition [C]//IEEE International Conference on Computer Vision.Santiago:ICCV,2015:1449-1457.
[26] LIN T,ROYCHOWDHURY A,MAJI S.Bilinear convolutional neural networks for fine-grained visual recognition [J].IEEE Transactions on Pattern Analysis and Machine Intelligence,2018,40(6):1309-1322.
[27] MENGRAN G,FEI X,OCTAVIA I,et al.MoNet:moments embedding network [C]//IEEE Conference on Computer Vision and Pattern Recognition.Salt Lake City:CVPR,2018:3175-3183.
[28] MNIH V,HEESS N,GRAVES A,et al.Recurrent models ofvisual attention [C]//Conference on Neural Information Processing Systems.Montreal:NIPS,2014.
[29] JADERBERG M,SIMONYAN K,ZISSERMAN A,et al.Spatial transformer networks [C]//Conference on Neural Information Processing Systems.Montreal:NIPS,2015.
[30] HU J,SHEN L,ALBANIE S,et al.Squeeze-and-excitation networks[C]//IEEE Conference on Computer Vision and Pattern Recognition.Salt Lake City:CVPR,2018:7132-7141.
[31] XUE J,ZHANG H,DANA K,et al.Differential angular imaging for material recognition [C]//IEEE Conference on Computer Vision and Pattern Recognition.Honolulu:CVPR,2017:6940-6949.
[32] LIU L,CHEN J,FIEGUTH P,et al.From bow to cnn:two decades of texture representation for texture classification [J].International Journal of Computer Vision,2019,127(1):74-109.
[33] HE K,ZHANG X,REN S,et al.Deep residual learning for image recognition [C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.Las Vegas:IEEE,2016:770-778.
[34] IOFFE S,SZEGEDY C.Batch normalization:accelerating deep network training by reducing internal covariate shift [C]//International Conference on Machine Learning.Lille:ICML,2015:448-456.
[35] Hayman E,Caputo B,Fritz M,et al.On the signi?cance of real world conditions for material classification [C]//European Conference on Computer Vision.Prague:ECCV,2004:253-266.
[36] SHARAN L,ROSENHOLTZ R,ADELSON E.Material per-ception:what can you see in a brief glance [J].Journal of Vision,2009,9(8):784-784.
[1] RAO Zhi-shuang, JIA Zhen, ZHANG Fan, LI Tian-rui. Key-Value Relational Memory Networks for Question Answering over Knowledge Graph [J]. Computer Science, 2022, 49(9): 202-207.
[2] TANG Ling-tao, WANG Di, ZHANG Lu-fei, LIU Sheng-yun. Federated Learning Scheme Based on Secure Multi-party Computation and Differential Privacy [J]. Computer Science, 2022, 49(9): 297-305.
[3] XU Yong-xin, ZHAO Jun-feng, WANG Ya-sha, XIE Bing, YANG Kai. Temporal Knowledge Graph Representation Learning [J]. Computer Science, 2022, 49(9): 162-171.
[4] WANG Jian, PENG Yu-qi, ZHAO Yu-fei, YANG Jian. Survey of Social Network Public Opinion Information Extraction Based on Deep Learning [J]. Computer Science, 2022, 49(8): 279-293.
[5] HAO Zhi-rong, CHEN Long, HUANG Jia-cheng. Class Discriminative Universal Adversarial Attack for Text Classification [J]. Computer Science, 2022, 49(8): 323-329.
[6] JIANG Meng-han, LI Shao-mei, ZHENG Hong-hao, ZHANG Jian-peng. Rumor Detection Model Based on Improved Position Embedding [J]. Computer Science, 2022, 49(8): 330-335.
[7] SUN Qi, JI Gen-lin, ZHANG Jie. Non-local Attention Based Generative Adversarial Network for Video Abnormal Event Detection [J]. Computer Science, 2022, 49(8): 172-177.
[8] HU Yan-yu, ZHAO Long, DONG Xiang-jun. Two-stage Deep Feature Selection Extraction Algorithm for Cancer Classification [J]. Computer Science, 2022, 49(7): 73-78.
[9] ZHANG Ying-tao, ZHANG Jie, ZHANG Rui, ZHANG Wen-qiang. Photorealistic Style Transfer Guided by Global Information [J]. Computer Science, 2022, 49(7): 100-105.
[10] CHENG Cheng, JIANG Ai-lian. Real-time Semantic Segmentation Method Based on Multi-path Feature Extraction [J]. Computer Science, 2022, 49(7): 120-126.
[11] HOU Yu-tao, ABULIZI Abudukelimu, ABUDUKELIMU Halidanmu. Advances in Chinese Pre-training Models [J]. Computer Science, 2022, 49(7): 148-163.
[12] ZHOU Hui, SHI Hao-chen, TU Yao-feng, HUANG Sheng-jun. Robust Deep Neural Network Learning Based on Active Sampling [J]. Computer Science, 2022, 49(7): 164-169.
[13] SU Dan-ning, CAO Gui-tao, WANG Yan-nan, WANG Hong, REN He. Survey of Deep Learning for Radar Emitter Identification Based on Small Sample [J]. Computer Science, 2022, 49(7): 226-235.
[14] ZHU Wen-tao, LAN Xian-chao, LUO Huan-lin, YUE Bing, WANG Yang. Remote Sensing Aircraft Target Detection Based on Improved Faster R-CNN [J]. Computer Science, 2022, 49(6A): 378-383.
[15] WANG Jian-ming, CHEN Xiang-yu, YANG Zi-zhong, SHI Chen-yang, ZHANG Yu-hang, QIAN Zheng-kun. Influence of Different Data Augmentation Methods on Model Recognition Accuracy [J]. Computer Science, 2022, 49(6A): 418-423.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!