Thanks to visit codestin.com
Credit goes to github.com

Skip to content
/ RHTNet Public

TCSVT2026 -- Representation Sampling and Hybrid Transformer Network for Image Compressed Sensing

Notifications You must be signed in to change notification settings

songhp/RHTNet

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Representation Sampling and Hybrid Transformer Network for Image Compressed Sensing

Heping Song; Jingyao Gong; Hongjie Jia; Xiangjun Shen; Jianping Gou; Hongying Meng; Le Wang

IEEE Transactions on Circuits and Systems for Video Technology (Early Access)

DOI

Abstract

Deep unfolding networks (DUNs) have attracted significant attention in the field of image compressed sensing (CS) due to its superior performance and good explainability by integrating optimization algorithms with deep networks. However, existing DUNs suffer from low sampling efficiency, and the improvement in reconstruction quality heavily relies on a large number of parameters, although this is not necessary. To address these issues, we propose a Representation Sampling and Hybrid Transformer Network (RHT-Net). We innovatively design a Representation-CS (RCS) model to extract high-level signal representations, sampling highly dense and semantically rich, extremely compact features without relying on observing the original pixels, which also reduces the cross-domain loss during fidelity term correction. In the deep recovery stage, we design a Tri-Scale Sparse Denoising (TSSD) module to extend sparse proximal projections, leveraging multi-scale auxiliary variables to enhance multi-feature flow and memory effects. Subsequently, we develop a hybrid Transformer that includes Global Cross Attention (GCA) and Window Local Attention (WLA), using the measurements to cross-estimate the reconstruction error, thereby generating finer spatial information and local recovery. Experiments demonstrate that RHT-Net outperforms existing SOTA methods by up to 1.17dB in PSNR. Furthermore, RHT-Net-light achieves a 0.43dB gain while reducing model parameters by up to 22 times, underscoring the superior efficiency of our approach.

Network Architecture

structure

Requirements (recommend)

  • Python >= 3.9
  • Pytorch >= 2.1.2
  • Numpy >= 1.26

Datasets

  • Train data: crop from BSDS dataset.
  • Test data: from Urban100, Set11, BSDS200, DIV2K.
  • Should decompress and place these datasets in the "./dataset/" directory.

Quick Start

Pre training model

  • We provide pretrained model weights for convenient evaluation, named the weights of RHT-Net as -light.

  • It contains six sampling rates (RHT-Net & Net$^+$)

Test

  1. Ensure the test dataset path in data_processor.py is correct.
  2. Ensure that the trained_model directory contains the pre-trained model weight files.
  3. Run the following scripts to test RHT-Net model:
    python3 eval_rht.py
    

Train

  1. Ensure the train dataset path in data_processor.py is correct.
  2. Run the following scripts to train RHT-Net model:
    python3 train.py --epochs 150 --batch-size 16
    

Results

Quantitative Results

tables visual

Citation

If you find the code helpful in your research or work, please cite the following paper:

@article{RHTNet25,
  title={Representation Sampling and Hybrid Transformer Network for Image Compressed Sensing},
  author={Song, Heping and Gong, Jingyao and Jia, Hongjie and Shen, Xiangjun and Gou, Jianping and Meng, Hongying and Wang, Le},
  journal={IEEE Transactions on Circuits and Systems for Video Technology},
  year={2025},
  publisher={IEEE}
}

Acknowledgements

We thank the authors for sharing their codes:

FSOINet, OCTUF, CASNet, USBNet, HUNet, NesTD-Net

About

TCSVT2026 -- Representation Sampling and Hybrid Transformer Network for Image Compressed Sensing

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages