in NTIRE 2024
Training Transformer Models by Wavelet Losses Improves Quantitative and Visual Performance in Single Image Super-Resolution
Cansu Korkmaz and A. Murat Tekalp
Also to compare our method, you can download benchmark results from Google Drive
Create python virtual environment and install dependencies
- python -m venv wlt
- source source ./wlt/bin/activate
- pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu118
- pip install -r requirements.txt
- clone this repository
git clone https://github.com/mandalinadagi/Wavelettention
cd Wavelettention
- Download pretrained model from Google Drive and place it under
wavelettention/pretrained_models/
. - Prepare the datasets which can be downloaded from Google Drive.
- Modify the configuration file
options/test_Wavelettention_SRx4.yml
. (path to datasets and the pretrained model) - Run the command
python wavelettention/test.py -opt options/test_Wavelettention_SRx4.yml
. - You can find the images in
results/test_wavelettention/visualization/
folder.
- Download the LSDIR dataset.
- Prepare the Imagenet pretrained x4 HAT-L model.
- Modify the configuration file
options/train_Wavelettention_SRx4.yml
. (path to datasets, pretrained model, set scaling of each wavelet loss term) - Run the command
CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7 python wavelettention/train.py -opt options/train_Wavelettention_SRx4.yml
.
If you find our work helpful in your resarch, please consider citing the following paper.
@inproceedings{korkmaz2024wavelettention,
title={Training Transformer Models by Wavelet Losses Improves Quantitative and Visual Performance in Single Image Super-Resolution},
author={Korkmaz, Cansu and Tekalp, A. Murat},
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops},
month={June},
year={2024}
}
If you have any questions please email at ckorkmaz14@ku.edu.tr
Our code is built on BasicSR and HAT. Thanks to their great work.