Abstract
Transformers have proven effective for hyperspectral image classification (HSIC) but often incorporate average pooling that results in information loss. This letter presents WaveFormer, a novel transformer-based approach that leverages wavelet transforms for invertible downsampling. This preserves data integrity while enabling attention learning. Specifically, WaveFormer unifies downsampling with wavelet transforms to decompress feature maps without loss. This provides an efficient tradeoff between performance and computation. Furthermore, the wavelet decomposition enhances the interaction between structural and shape information in image patches and channel maps. To evaluate WaveFormer, we conducted extensive experiments on two benchmark hyperspectral datasets. Our results demonstrate that WaveFormer achieves state-of-the-art classification accuracy, obtaining overall accuracies of 95.66% and 96.54% on the Pavia University and the University of Houston datasets, respectively. By integrating wavelet transforms, WaveFormer presents a new transformer architecture for hyperspectral imagery that achieves superior classification without information loss from average pooling.
| Original language | English |
|---|---|
| Article number | 5502405 |
| Pages (from-to) | 1-5 |
| Number of pages | 5 |
| Journal | IEEE Geoscience and Remote Sensing Letters |
| Volume | 21 |
| DOIs | |
| State | Published - 2024 |
| Externally published | Yes |
Bibliographical note
Publisher Copyright:© 2004-2012 IEEE.
Keywords
- Hyperspectral image classification (HSIC)
- spatial-spectral feature
- spatial-spectral transformers (SSTs)
- wavelet transformer (WaveFormer)
ASJC Scopus subject areas
- Geotechnical Engineering and Engineering Geology
- Electrical and Electronic Engineering