Computer Science ›› 2023, Vol. 50 ›› Issue (5): 155-160.doi: 10.11896/jsjkx.220400035

• Computer Graphics & Multimedia • Previous Articles     Next Articles

Hyperspectral Image Classification Based on Swin Transformer and 3D Residual Multilayer Fusion Network

WANG Xianwang, ZHOU Hao, ZHANG Minghui, ZHU Youwei   

  1. School of Information,Yunnan University,Kunming 650500,China
  • Received:2022-04-06 Revised:2022-07-04 Online:2023-05-15 Published:2023-05-06
  • About author:WANG Xianwang,born in 1995,postgraduate.His main research interests include hyper-spectral image classification and so on.
    ZHOU Hao,born in 1972,Ph.D,asso-ciate professor.His main research in-terests include digital image processing and computer vision.
  • Supported by:
    District Key Project of Yunnan Province(202202AD080004) and National Natural Science Foundation of China(12263008).

Abstract: Convolutional neural networks(CNNs) are widely seen in in hyperspectral image classification due to their remarkably good local context modeling performance.However,under its inherent limitations of network structure,it fails to exploit and represent sequence attributes from spectral characteristics.To address this problem,an integrated novel network,based on Swin Transformer and 3D residual multi-layer fusion network model,is proposed for hyperspectral image classification.In order to excavate the deep features of hyperspectral images as much as possible,spatial spectrum is extracted by improved 3D residual multi-layer fusion network in ReSTrans network,and the context information in consecutive spectra is captured by self-attention mecha-nism Swin Transformer network model.The final result of classification is obtained by multi-layer perception based on spatial spectrum joint feature.In order to verify the effectiveness of the ReSTrans network model,the improved model is experimentally verified on three hyperspectral data sets of IP,UP and KSC,and the classification accuracy reaches 98.65 %,99.64% and 99.78% respectively.Compared with SST method,the classification performance of the network model improves by 3.55%,0.68% and 1.87% respectively.Experimental results show that the model had good generalization ability and could extract deeper and discriminative features.

Key words: Hyperspectral image classification, 3D residual multilayer fusion network, Self-attention mechanism, Swin Transfor-mer, Spatial spectrum joint feature

CLC Number: 

  • TP751.1
[1]REN S G,WAN S,GU X J,et al.Hyper-spectral image classifi-cation based on multi-scale spatial spectrum identification features[J].Computer Science,2018,45(12):243-250.
[2]ZHU N,LI M.Multilevel selective kernel convolution for retina image classification[J].Journal of Chongqing University of Posts and Telecommunications(Natural Science Edition),2022,34(5):886-893.
[3]LV W,WANG X.Overview of Hyperspectral Image Classification[J].Journal of Sensors,2020,2020(2):1-13.
[4]HAUT J M,PAOLETTI M E,PLAZA J,et al.Visual attention-driven hyperspectral image classification[J].IEEE Transactions on Geos-cience and Remote Sensing,2019,57(10):8065-8080.
[5]WEI X P,YU X C,TAN X,et al.CNN and 3D Gabor filter for hyperspectral image classifica-tion[J].Journal of Computer Aided Design and Graphics,2020,32(1):90-98.
[6]HE M,LI B,CHEN H,et al.Multi-scale 3D deep convolutional neural network for hyperspectral image classification[C]//2017 IEEE International Conference on Image Processing(ICIP).IEEE,2017:3904-3908.
[7]HANG R,LIU Q,HONG D,et al.Cascaded recurrent neural networks for hyperspectral image classification[J].IEEE Transac-tions on Geoscience and Remote Sensing,2019,57(8):5384-5394.
[8]MÜLLER G,RIOS M,SENNRICH A,et al.Why Self-Attention? A Targeted Evaluation of Neural Machine Translation Architectures[J].arXiv:1808.08946,2018.
[9]CARION N,MASSA F,SYNNAEVE G,et al.End-to-end object detection with trans-formers[C]//European Conference on Computer Vision.Berlin:Springer,2020:213-219.
[10]RAMACHANDRAN P,PARMAR N,VASWANI A,et al.Stand-Alone Self-Attention in Vision Models[J].arXiv:1906.05909,2019.
[11]LIU Z,LIN Y,CAO Y,et al.Swin transformer:Hierarchical vision transformer using shifted windows[C]//Proceedings of the IEEE/CVF International Conference on Computer Vision.2021:10012-10022.
[12]HU W,HUANG Y Y,LI H C,et al.Deep Convolutional Neural Networks for Hyper-spectral Image Classification[J].Journal of Sensors,2015,2015:1-12.
[13]LIU,B,YU X C,ZHANG P Q,et al,A semi-supervised convolutional neural network for hyper-spectral image classification[J].Remote Sensing Letters,2017,8(9):839-848.
[14]HAMID A B,BENOIT A,LAMBERT P,et al.3D deep learningapproach for remote sensing image classification[J].IEEE Transactions on Geoscience and Remote Sensing,2018,56(8):4420-4434.
[15]HARA K,KATAOKA H,SATOH Y.Learning spatio-temporal features with 3d residual networks for action recognition[C]//Proceedings of the IEEE International Conference on Computer Vision Workshops.2017.
[16]HE X,CHEN Y,LIN Z.Spatial-spectral transformer for hyperspectral image classification[J].Remote Sensing,2021,13(3):498.
[1] LI Fan, JIA Dongli, YAO Yumin, TU Jun. Graph Neural Network Few Shot Image Classification Network Based on Residual and Self-attention Mechanism [J]. Computer Science, 2023, 50(6A): 220500104-5.
[2] DOU Zhi, HU Chenguang, LIANG Jingyi, ZHENG Liming, LIU Guoqi. Lightweight Target Detection Algorithm Based on Improved Yolov4-tiny [J]. Computer Science, 2023, 50(6A): 220700006-7.
[3] YANG Bin, LIANG Jing, ZHOU Jiawei, ZHAO Mengci. Study on Interpretable Click-Through Rate Prediction Based on Attention Mechanism [J]. Computer Science, 2023, 50(5): 12-20.
[4] ZHANG Jingyuan, WANG Hongxia, HE Peisong. Multitask Transformer-based Network for Image Splicing Manipulation Detection [J]. Computer Science, 2023, 50(1): 114-122.
[5] JIN Fang-yan, WANG Xiu-li. Implicit Causality Extraction of Financial Events Integrating RACNN and BiLSTM [J]. Computer Science, 2022, 49(7): 179-186.
[6] ZHANG Jia-hao, LIU Feng, QI Jia-yin. Lightweight Micro-expression Recognition Architecture Based on Bottleneck Transformer [J]. Computer Science, 2022, 49(6A): 370-377.
[7] ZHAO Dan-dan, HUANG De-gen, MENG Jia-na, DONG Yu, ZHANG Pan. Chinese Entity Relations Classification Based on BERT-GRU-ATT [J]. Computer Science, 2022, 49(6): 319-325.
[8] HU Yan-li, TONG Tan-qian, ZHANG Xiao-yu, PENG Juan. Self-attention-based BGRU and CNN for Sentiment Analysis [J]. Computer Science, 2022, 49(1): 252-258.
[9] WANG Xi, ZHANG Kai, LI Jun-hui, KONG Fang, ZHANG Yi-tian. Generation of Image Caption of Joint Self-attention and Recurrent Neural Network [J]. Computer Science, 2021, 48(4): 157-163.
[10] ZHOU Xiao-shi, ZHANG Zi-wei, WEN Juan. Natural Language Steganography Based on Neural Machine Translation [J]. Computer Science, 2021, 48(11A): 557-564.
[11] ZHANG Peng-fei, LI Guan-yu, JIA Cai-yan. Truncated Gaussian Distance-based Self-attention Mechanism for Natural Language Inference [J]. Computer Science, 2020, 47(4): 178-183.
[12] ZHANG Yi-jie, LI Pei-feng, ZHU Qiao-ming. Event Temporal Relation Classification Method Based on Self-attention Mechanism [J]. Computer Science, 2019, 46(8): 244-248.
[13] FAN Zi-wei, ZHANG Min, LI Zheng-hua. BiLSTM-based Implicit Discourse Relation Classification Combining Self-attention
Mechanism and Syntactic Information
[J]. Computer Science, 2019, 46(5): 214-220.
[14] LI Chang-li, ZHANG Lin, FAN Tang-huai. Hyperspectral Image Classification Based on Adaptive Active Learning and Joint Bilateral Filtering [J]. Computer Science, 2018, 45(12): 223-228.
[15] FENG Hai-liang,PAN Jing-wen and HUANG Hong. Hyperspectral Image Classification Based on Semi-supervised Neighborhood Preserving Embedding [J]. Computer Science, 2014, 41(Z6): 230-233.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!