Heping Song; Jingyao Gong; Hongjie Jia; Xiangjun Shen; Jianping Gou; Hongying Meng; Le Wang
IEEE Transactions on Circuits and Systems for Video Technology (Early Access)
Deep unfolding networks (DUNs) have attracted significant attention in the field of image compressed sensing (CS) due to its superior performance and good explainability by integrating optimization algorithms with deep networks. However, existing DUNs suffer from low sampling efficiency, and the improvement in reconstruction quality heavily relies on a large number of parameters, although this is not necessary. To address these issues, we propose a Representation Sampling and Hybrid Transformer Network (RHT-Net). We innovatively design a Representation-CS (RCS) model to extract high-level signal representations, sampling highly dense and semantically rich, extremely compact features without relying on observing the original pixels, which also reduces the cross-domain loss during fidelity term correction. In the deep recovery stage, we design a Tri-Scale Sparse Denoising (TSSD) module to extend sparse proximal projections, leveraging multi-scale auxiliary variables to enhance multi-feature flow and memory effects. Subsequently, we develop a hybrid Transformer that includes Global Cross Attention (GCA) and Window Local Attention (WLA), using the measurements to cross-estimate the reconstruction error, thereby generating finer spatial information and local recovery. Experiments demonstrate that RHT-Net outperforms existing SOTA methods by up to 1.17dB in PSNR. Furthermore, RHT-Net-light achieves a 0.43dB gain while reducing model parameters by up to 22 times, underscoring the superior efficiency of our approach.
- Python >= 3.9
- Pytorch >= 2.1.2
- Numpy >= 1.26
- Train data: crop from BSDS dataset.
- Test data: from Urban100, Set11, BSDS200, DIV2K.
- Should decompress and place these datasets in the "./dataset/" directory.
-
We provide pretrained model weights for convenient evaluation, named the weights of RHT-Net as -light.
-
It contains six sampling rates (RHT-Net & Net$^+$)
- Ensure the test dataset path in data_processor.py is correct.
- Ensure that the
trained_modeldirectory contains the pre-trained model weight files. - Run the following scripts to test RHT-Net model:
python3 eval_rht.py
- Ensure the train dataset path in
data_processor.pyis correct. - Run the following scripts to train RHT-Net model:
python3 train.py --epochs 150 --batch-size 16
If you find the code helpful in your research or work, please cite the following paper:
@article{RHTNet25,
title={Representation Sampling and Hybrid Transformer Network for Image Compressed Sensing},
author={Song, Heping and Gong, Jingyao and Jia, Hongjie and Shen, Xiangjun and Gou, Jianping and Meng, Hongying and Wang, Le},
journal={IEEE Transactions on Circuits and Systems for Video Technology},
year={2025},
publisher={IEEE}
}
We thank the authors for sharing their codes: