Self-supervised spatial-spectral transformer with Extreme Learning Machine for Hyperspectral Image Classification

  • Muhammad Ahmad*
  • , Manuel Mazzara
  • , Salvatore Distefano
  • , Adil Mehmood Khan
  • , Xin Wu
  • *Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

3 Scopus citations

Abstract

While deep learning has significantly advanced hyperspectral image (HSI) classification, capturing complex spatial-spectral features efficiently and achieving robust performance with limited dynamic masking within masked image sparse attention mechanism enhances computational efficiency by selectively attending to key spectral-spatial patches, effectively reducing complexity, without compromising classification performance. An Extreme Learning Machine (ELM) is integrated as the final classification layer, leveraging SST-extracted features for efficient and lightweight classification. The resulting hybrid SST-ELM model achieves 99.93% overall accuracy on the Salinas dataset, improving upon the baseline by 1.12% with a 5 reduction in training time. Similar improvements are observed on Pavia University (99.06%), Longkou (99.83%), Hanchuan (96.04%), and Honghu (97.10%) datasets.

Original languageEnglish
Pages (from-to)5384-5407
Number of pages24
JournalInternational Journal of Remote Sensing
Volume46
Issue number14
DOIs
StatePublished - 2025

Bibliographical note

Publisher Copyright:
© 2025 Informa UK Limited, trading as Taylor & Francis Group.

Keywords

  • Extreme Learning Machine
  • Hyperspectral Image Classification
  • Self-supervised learning
  • dynamic masking
  • sparse attention
  • spatial-spectral transformer (SST)

ASJC Scopus subject areas

  • General Earth and Planetary Sciences

Fingerprint

Dive into the research topics of 'Self-supervised spatial-spectral transformer with Extreme Learning Machine for Hyperspectral Image Classification'. Together they form a unique fingerprint.

Cite this