TY - JOUR
T1 - Hyperspectral Image Classification Based on Multibranch Attention Transformer Networks
AU - Bai, Jing
AU - Wen, Zheng
AU - Xiao, Zhu
AU - Ye, Fawang
AU - Zhu, Yongdong
AU - Alazab, Mamoun
AU - Jiao, Licheng
PY - 2022/8
Y1 - 2022/8
N2 - Deep learning (DL) has become a mainstream method of hyperspectral image (HSI) classification. Many DL-based methods exploit spatial-spectral features to achieve better classification results. However, due to the complex backgrounds in HSIs, existing methods usually show unsatisfactory performance for the class pixels located on the land-cover category boundary area. In large part, this is because the network is susceptible to interference by the irrelevant information around the target pixel in the training stage, resulting in inaccurate feature extraction. In this article, a new multibranch transformer architecture (spectral spatial transformer (SST)-M) that assembles spatial attention and extracts spectral features is proposed to address this problem. The transformer model has a global receptive field and thus can integrate global spatial position information in the HSI cube. Meanwhile, we design a spatial sequence attention model to enhance the useful spatial location features and weaken invalid information. Considering that HSIs contain considerable spectral information, a spectral feature extraction model is designed to extract discriminative spectral features, replacing the widely used principal component analysis (PCA) method and obtaining better classification results than it. Finally, inspired by semantic segmentation, a mask prediction model is designed to classify all of the pixels in the HSI cube; this guides the neural network to learn precise pixel characteristics and spatial distributions. To verify the effectiveness of our algorithm (SST-M), quantitative experiments were conducted in three well-known datasets, namely, Indian Pines (IP), University of Pavia (PU), and Kennedy Space Center (KSC). The experimental results demonstrate that the proposed model achieves better performance than the other state-of-the-art methods.
AB - Deep learning (DL) has become a mainstream method of hyperspectral image (HSI) classification. Many DL-based methods exploit spatial-spectral features to achieve better classification results. However, due to the complex backgrounds in HSIs, existing methods usually show unsatisfactory performance for the class pixels located on the land-cover category boundary area. In large part, this is because the network is susceptible to interference by the irrelevant information around the target pixel in the training stage, resulting in inaccurate feature extraction. In this article, a new multibranch transformer architecture (spectral spatial transformer (SST)-M) that assembles spatial attention and extracts spectral features is proposed to address this problem. The transformer model has a global receptive field and thus can integrate global spatial position information in the HSI cube. Meanwhile, we design a spatial sequence attention model to enhance the useful spatial location features and weaken invalid information. Considering that HSIs contain considerable spectral information, a spectral feature extraction model is designed to extract discriminative spectral features, replacing the widely used principal component analysis (PCA) method and obtaining better classification results than it. Finally, inspired by semantic segmentation, a mask prediction model is designed to classify all of the pixels in the HSI cube; this guides the neural network to learn precise pixel characteristics and spatial distributions. To verify the effectiveness of our algorithm (SST-M), quantitative experiments were conducted in three well-known datasets, namely, Indian Pines (IP), University of Pavia (PU), and Kennedy Space Center (KSC). The experimental results demonstrate that the proposed model achieves better performance than the other state-of-the-art methods.
KW - Convolution
KW - Convolutional neural networks
KW - Deep learning
KW - Feature extraction
KW - hyperspectral image classification
KW - Hyperspectral imaging
KW - multibranch prediction
KW - self-attention mechanism
KW - spatial attention
KW - Three-dimensional displays
KW - Training
KW - transformer model
KW - Transformers
UR - http://www.scopus.com/inward/record.url?scp=85135745459&partnerID=8YFLogxK
U2 - 10.1109/TGRS.2022.3196661
DO - 10.1109/TGRS.2022.3196661
M3 - Article
AN - SCOPUS:85135745459
SN - 0196-2892
VL - 60
SP - 1
EP - 17
JO - IEEE Transactions on Geoscience and Remote Sensing
JF - IEEE Transactions on Geoscience and Remote Sensing
M1 - 5535317
ER -