Research Activities早稲田大学 研究活動

Multiscanning-Based RNN-Transformer for Hyperspectral Image Classification(Published in IEEE Transactions on Geoscience and Remote Sensing, May 2023)

Journal Title
/掲載ジャーナル名
IEEE Transactions on Geoscience and Remote Sensing
Publication Year and Month
/掲載年月
May, 2023
Paper Title
/論文タイトル
Multiscanning-Based RNN-Transformer for Hyperspectral Image Classification
DOI
/論文DOI
10.1109/TGRS.2023.3277014
 Author of Waseda University
/本学の著者
KAMATA, Seiichiro(Professor, Faculty of Science and Engineering, Graduate School of Information, Production, and Systems):Corresponding Author
Related Websites
/関連Web
Abstract
/抄録
The goal of hyperspectral image (HSI) classification is to assign land-cover labels to each HSI pixel in a patchwise manner. Recently, sequential models, such as recurrent neural networks (RNNs), have been developed as HSI classifiers, which need to scan the HSI patch into a pixel sequence with the scanning order first. However, RNNs have a biased ordering that cannot effectively allocate attention to each pixel in the sequence, and previous methods that use multiple scanning orders to average the features of RNNs are limited by the validity of these orders. To solve this issue, it is naturally inspired by Transformer and its self-attention to discriminatively distribute proper attention for each pixel of the pixel sequence and each scanning order. Hence, in this study, we further develop the sequential HSI classifiers by a specially designed RNN–Transformer (RT) model to feature the multiple sequential characters of the HSI pixels in the HSI patch. Specifically, we introduce a multiscanning-controlled positional embedding strategy for the RT model to complement multiple feature fusion. Furthermore, the RT encoder is proposed for integrating ordering bias and attention reallocation for feature generation at the sequence level. In addition, the spectral–spatial-based soft masked self-attention (SMSA) is proposed for suitable feature enhancement. Finally, an additional fusion Transformer (FT) is deployed for scanning order-level attention allocation. As a result, the whole network can achieve competitive classification performance on four accessible datasets than other state-of-the-art methods. Our study further extends the research on sequential HSI classifiers.
Page Top
WASEDA University

早稲田大学オフィシャルサイト(https://www.waseda.jp/inst/research/)は、以下のWebブラウザでご覧いただくことを推奨いたします。

推奨環境以外でのご利用や、推奨環境であっても設定によっては、ご利用できない場合や正しく表示されない場合がございます。より快適にご利用いただくため、お使いのブラウザを最新版に更新してご覧ください。

このままご覧いただく方は、「このまま進む」ボタンをクリックし、次ページに進んでください。

このまま進む

対応ブラウザについて

閉じる