-
Python 3.6.12
-
Pytorch 1.2.0+
-
torchvision 0.4.0+
- We released our code for joint training with depth and appearance, which is also our best performance model.
git clone https://github.com/OLobbCode/NoiseLF.git
cd NoiseLF-code/Download the following datasets and unzip them.
- DUT-LF dataset,fetch code is ‘vecy’.
- HFUT dataset.
- LFSD dataset.
- The .txt file link for testing and training is here, code is 'joaa'.
- Set the
c.DATA.TRAIN.ROOTandc.DATA.TRAIN.LISTpath inconfig.pycorrectly. - We demo using VGG-19 as network backbone and train with a initial lr of 1e-5 for 30 epoches.
- After training the result model will be stored under
snapshot/exp_noiselffolder.
Note:only support c.SOLVER.BATCH_SIZE=1
For single dataset testing: you should set c.PHASE='test' in config.py, and set c.DATA.TEST.ROOT , c.DATA.TEST.LIST as yours.
python demo.py For evaluate :
python evaluate.pyAll results saliency maps will be stored under 'Test/Out/exp_noiself_30/' folders in .png formats.
Thanks to MOLF.
