Jiahua Dong, Hui Yin, Hongliu Li, Wenbo Li, Yulun Zhang, Salman Khan, Fahad Khan, "Dual Hyperspectral Mamba for Efficient Spectral Compressive Imaging", arXiv, 2024
Scene 1 | Scene 5 | Scene 6 | Scene 7 |
---|---|---|---|
- 2024-5-29: This repo is released.
Abstract: Deep unfolding methods have made impressive progress in restoring 3D hyperspectral images (HSIs) from 2D measurements through convolution neural networks or Transformers in spectral compressive imaging. However, they cannot efficiently capture long-range dependencies using global receptive fields, which significantly limits their performance in HSI reconstruction. Moreover, these methods may suffer from local context neglect if we directly utilize Mamba to unfold a 2D feature map as a 1D sequence for modeling global long-range dependencies. To address these challenges, we propose a novel Dual Hyperspectral Mamba (DHM) to explore both global long-range dependencies and local contexts for efficient HSI reconstruction. After learning informative parameters to estimate degradation patterns of the CASSI system, we use them to scale the linear projection and offer noise level for the denoiser (i.e., our proposed DHM). Specifically, our DHM consists of multiple dual hyperspectral S4 blocks (DHSBs) to restore original HSIs. Particularly, each DHSB contains a global hyperspectral S4 block (GHSB) to model long-range dependencies across the entire high-resolution HSIs using global receptive fields, and a local hyperspectral S4 block (LHSB) to address local context neglect by establishing structured state-space sequence (S4) models within local windows. Experiments verify the benefits of our DHM for HSI reconstruction.
Method | Params (M) | FLOPS (G) | PSNR | SSIM | Model Zoo | Result |
---|---|---|---|---|---|---|
DAUHST-9stg | 6.15 | 79.50 | 38.36 | 0.967 | Repo | Repo |
PADUT-12stg | 5.38 | 90.46 | 38.89 | 0.972 | Google Driver | Google Driver |
RDLUF-MixS2-9stg | 1.89 | 115.34 | 39.57 | 0.974 | Repo | Repo |
DERNN-LNLT-3stg | 0.65 | 27.41 | 38.65 | 0.973 | - | - |
DERNN-LNLT-5stg | 0.65 | 45.60 | 39.38 | 0.973 | Repo | Repo |
DERNN-LNLT-7stg | 0.65 | 63.80 | 39.61 | 0.974 | Repo | Repo |
DERNN-LNLT-9stg | 0.65 | 81.99 | 39.93 | 0.976 | Repo | Repo |
DERNN-LNLT-9stg* | 1.09 | 134.18 | 40.33 | 0.977 | Repo | Repo |
DHM-light-3stg | 0.66 | 26.42 | 38.99 | 0.975 | - | - |
DHM-light-5stg | 0.66 | 43.96 | 39.81 | 0.979 | - | - |
DHM-light-7stg | 0.66 | 61.50 | 40.20 | 0.980 | - | - |
DHM-light-9stg | 0.66 | 79.04 | 40.33 | 0.981 | - | - |
DHM-3stg | 0.92 | 36.34 | 39.13 | 0.975 | - | - |
DHM-5stg | 0.92 | 60.50 | 40.16 | 0.980 | - | - |
DHM-7stg | 0.92 | 84.65 | 40.34 | 0.981 | - | - |
DHM-9stg | 0.92 | 108.80 | 40.50 | 0.981 | - | - |
- Create Environment
- Data Preparation
- Simulation Experiement
- Real Experiement
- Results
- Citation
- Acknowledgements
conda create -n your_env_name python=3.9
conda activate your_env_name
conda install cudatoolkit==11.7 -c nvidia
conda install pytorch==1.13.1 torchvision==0.14.1 torchaudio==0.13.1 pytorch-cuda=11.7 -c pytorch -c nvidia
conda install -c "nvidia/label/cuda-11.7.0" cuda-nvcc
conda install packaging
pip install causal-conv1d==1.0.0
pip install mamba_ssm==1.0.1
pip install -r requirements.txt
Download cave_1024_28 (Baidu Disk, code: fo0q
| One Drive), CAVE_512_28 (Baidu Disk, code: ixoe
| One Drive), KAIST_CVPR2021 (Baidu Disk, code: 5mmn
| One Drive), TSA_simu_data (Baidu Disk, code: efu8
| One Drive), TSA_real_data (Baidu Disk, code: eaqe
| One Drive), and then put them into the corresponding folders of datasets/
and recollect them as the following form:
|--real
|-- test_code
|-- train_code
|--simulation
|-- test_code
|-- train_code
|--datasets
|--cave_1024_28
|--scene1.mat
|--scene2.mat
:
|--scene205.mat
|--CAVE_512_28
|--scene1.mat
|--scene2.mat
:
|--scene30.mat
|--KAIST_CVPR2021
|--1.mat
|--2.mat
:
|--30.mat
|--TSA_simu_data
|--mask.mat
|--Truth
|--scene01.mat
|--scene02.mat
:
|--scene10.mat
|--TSA_real_data
|--mask.mat
|--Measurements
|--scene1.mat
|--scene2.mat
:
|--scene5.mat
We use the CAVE dataset (cave_1024_28) as the simulation training set. Both the CAVE (CAVE_1024_28) and KAIST (KAIST_CVPR2021) datasets are used as the real training set.
cd DHM/
# DHM-light 3stage
bash ./scripts/train_DHM_light_3stg_simu.sh
# DHM-light 5stage
bash ./scripts/train_DHM_light_5stg_simu.sh
# DHM-light 7stage
bash ./scripts/train_DHM_light_7stg_simu.sh
# DHM-light 9stage
bash ./scripts/train_DHM_light_9stg_simu.sh
# DHM 3stage
bash ./scripts/train_DHM_3stg_simu.sh
# DHM 5stage
bash ./scripts/train_DHM_5stg_simu.sh
# DHM 7stage
bash ./scripts/train_DHM_7stg_simu.sh
# DHM 9stage
bash ./scripts/train_DHM_9stg_simu.sh
The training log, trained model, and reconstrcuted HSI will be available in DHM/exp/
.
Place the pretrained model to DHM/checkpoints/
Run the following command to test the model on the simulation dataset.
cd DHM/
# DHM-light 3stage
bash ./scripts/test_DHM_light_3stg_simu.sh
# DHM-light 5stage
bash ./scripts/test_DHM_light_5stg_simu.sh
# DHM-light 7stage
bash ./scripts/test_DHM_light_7stg_simu.sh
# DHM-light 9stage
bash ./scripts/test_DHM_light_9stg_simu.sh
# DHM 3stage
bash ./scripts/test_DHM_3stg_simu.sh
# DHM 5stage
bash ./scripts/test_DHM_5stg_simu.sh
# DHM 7stage
bash ./scripts/test_DHM_7stg_simu.sh
# DHM 9stage
bash ./scripts/test_DHM_9stg_simu.sh
The reconstrcuted HSIs will be output into DHM/results/
cd DHM/
# DHM-light 5stage
bash ./scripts/train_DHM_light_5stg_real.sh
The training log and trained model will be available in DHM/exp/
cd DHM/
# DHM-light 5stage
bash ./scripts/test_DHM_light_5stg_real.sh
The reconstrcuted HSI will be output into DHM/results/
We achieved state-of-the-art performance on Simulation dataset. Detailed results can be found in the paper.
Evaluation on Simulation Datasets (click to expand)
- quantitative comparisons in Table 1 of the main paper
- visual comparison in Figure 4 of the main paper
If you find the code helpful in your resarch or work, please cite the following paper(s).
@article{Dong2024DHM,
title={Dual Hyperspectral Mamba for Efficient Spectral Compressive Imaging},
author={Jiahua Dong and Hui Yin and Hongliu Li and Wenbo Li and Yulun Zhang and Salman Khan and Fahad Shahbaz Khan},
year={2024},
journal={arXiv preprint arXiv:2406.00449}
}
Our code is based on following codes, thanks for their generous open source: