This is the demo of code, model and methods used in my CVPR 2019 paper (link). There are some differences between the model used in the paper and this repository, the model used in the paper is implemented in Tensorflow and the model in this repository is implemented in Pytorch(0.4.1), if you are interested in this work, please run the demo.ipynb in jupyter notebbok to see the model and methods.
The model is trained on Salicon database.
Some saliency prediction examples on OSIE data
All the data annotation is done by myself using labelme
synthetic_data
OSIE-SR
The SegmentationClass folder contains the semantic level masks for each salient regions in the image,and the SegmentationObject folder contains the instance level masks for each salient regions in the image.
1:person head, 10:person part, 16:animal head, 22:animal part, 27:object, 37:text, 46:symbol, 51:vehicle, 57: food, 63:plant, 68:drink, 73:other
Here we report the representation in different backbones (resnet-18) after fine-tuning
Comparison of the activation map in vgg-16 and resnet-18 after fine-tuning, top row is the image and activation maps from vgg-16 after fine-tuning, bottom row is the groud truth saliency map and activation maps from resnet-18 after fine-tuning
you need to first set up your own image path and the binary fixation map path in the code.
python sal_train_pt.py
download the imagenet pretrained weight here
python sal_train_tf.py
@inproceedings{he2019understanding,
title={Understanding and Visualizing Deep Visual Saliency Models},
author={He, Sen and Tavakoli, Hamed R and Borji, Ali and Mi, Yang and Pugeault, Nicolas},
booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition},
pages={10206--10215},
year={2019}
}