Code for the paper "Occlusion Relationship Reasoning with A Feature Separation and Interaction Network". Submit to IJCV.
Authors: Yu Zhou, Rui Lu, Feng Xue,Yuzhe Gao, Xiaojie Guo and Wenqing Cheng
Occlusion relationship reasoning aims to locate where an object occludes others and estimate the depth order of these objects in the 3D space from a 2D image. The former sub-task demands both the accurate location and the semantic indication of the objects, while the latter one needs the depth order among the objects. Although several insightful studies have been proposed, a key characteristic, i.e., the specialty and complementarity between the occlusion boundary detection and the occlusion orientation estimation, is rarely discussed. To verify this claim, in this paper, we propose a network, namely Feature Separation and InteractionNetwork (FSINet), to integrate these properties into a unified end-to-end network, which comprises of a shared encoder-decoder structure and two separated paths for these two sub-tasks. Concretely, the occlusion boundary path contains an Image-level Cue Extractor to capture rich location information of the boundary, a Detail-perceived Semantic Feature Extractor, and a Contextual Correlation Extractor to acquire refined semantic messages of objects. In addition, a Dual-flow Cross Detector is customized to alleviate false-positive and false-negative boundaries. As for the occlusion orientation estimation path, a Scene Context Learner is designed to capture the depth order cue around the boundary. As well, two strip convolutions are built to judge the depth order between objects. The shared decoder supplies the feature interaction, which plays a key role in exploiting the complementarity of the two paths. Extensive experimental results on the PIOD and BSDS ownership datasets are conducted to reveal the superior performance of FSINet over state-of-the-art alternatives. Plus, abundant ablation studies are offered to demonstrate the effectiveness of our design.
The Data Preparation and Evaluation are following Guoxia Wang with his DOOBNet. Thanks for his valuable work.
You may download the dataset original images from PASCAL VOC 2010 and annotations from here. Then you should copy or move JPEGImages
folder in PASCAL VOC 2010 and Data
folder and val_doc_2010.txt in PIOD to data/PIOD/
. You will have the following directory structure:
PIOD
|_ Data
| |_ <id-1>.mat
| |_ ...
| |_ <id-n>.mat
|_ JPEGImages
| |_ <id-1>.jpg
| |_ ...
| |_ <id-n>.jpg
|_ val_doc_2010.txt
Now, you can use data convert tool to augment and generate HDF5 format data for DFNet.
mkdir data/PIOD/Augmentation
python doobscripts/doobnet_mat2hdf5_edge_ori.py \
--dataset PIOD \
--label-dir data/PIOD/Data \
--img-dir data/PIOD/JPEGImages \
--piod-val-list-file data/PIOD/val_doc_2010.txt \
--output-dir data/PIOD/Augmentation
For BSDS ownership dataset, you may download the dataset original images from BSDS300 and annotations from here. Then you should copy or move BSDS300
folder in BSDS300-images and trainfg
and testfg
folder in BSDS_theta to data/BSDSownership/
. And you will have the following directory structure:
BSDSownership
|_ trainfg
| |_ <id-1>.mat
| |_ ...
| |_ <id-n>.mat
|_ testfg
| |_ <id-1>.mat
| |_ ...
| |_ <id-n>.mat
|_ BSDS300
| |_ images
| |_ train
| |_ <id-1>.jpg
| |_ ...
| |_ <id-n>.jpg
| |_ ...
| |_ ...
Note that BSDS ownership's test set are split from 200 train images (100 for train, 100 for test). More information you can check ids in trainfg
and testfg
folder and ids in BSDS300/images/train
folder, or refer to here
Run the following code for BSDS ownership dataset.
mkdir data/BSDSownership/Augmentation
python doobscripts/doobnet_mat2hdf5_edge_ori.py \
--dataset BSDSownership \
--label-dir data/BSDSownership/trainfg \
--img-dir data/BSDSownership/BSDS300/images/train \
--bsdsownership-testfg data/BSDSownership/testfg \
--output-dir data/BSDSownership/Augmentation
Firstly, you need to download the Res50 weight file from Res50 and save resnet50.caffemodel
to the folder $DFNET_ROOT/models/resnet/
.
For training FSINet on PIOD training dataset, you can run:
cd $DFNET_ROOT/examples/DFNet
python write_prototxt.py
bash train.sh
When training completed, you need to modify the save model path model = 'snapshot/dfnet_iter_30000.caffemodel'
in eval.py
and then run python eval.py
to get the results on PIOD testing dataset. For comparation, you can also download our trained model from here. (code: jjnt). The testing results are available at here. (code: ynqs).
For training DFNet on BSDS ownership, you can refer the manner as same as PIOD dataset above. The training model is available at here. (code: 5bmf). The testing results are available at here. (code: 2uni).
Here we provide the PIOD and the BSDS ownership dataset's evaluation and visualization code in doobscripts
folder.
Note that you need to config the necessary paths or variables. More information please refers to doobscripts/README.md
.
To run the evaluation:
run doobscripts/evaluation/EvaluateOcc.m
For visualization, to run the script:
run doobscripts/visulation/PlotAll.m
Tab.1 Comparisons on the PIOD dataset with the state-of-the-art methods.
Method | ODS-E | OIS-E | AP-E | ODS-O | OIS-O | AP-O |
---|---|---|---|---|---|---|
SRF-OCC | .345 | .369 | .207 | .268 | .286 | .152 |
DOC-HED | .509 | .532 | .468 | .460 | .479 | .405 |
DOC-DMLFOV | .669 | .684 | .677 | .601 | .611 | .585 |
DOOBNet | .736 | .746 | .723 | .702 | .712 | .683 |
OFNet | .751 | .762 | .773 | .718 | .728 | .729 |
FSINet | .762 | .774 | .779 | .733 | .743 | .738 |
Tab.2 Comparisons on the BSDS ownership dataset with the state-of-the-art methods.
Method | ODS-E | OIS-E | AP-E | ODS-O | OIS-O | AP-O |
---|---|---|---|---|---|---|
SRF-OCC | .511 | .544 | .442 | .419 | .448 | .337 |
DOC-HED | .658 | .685 | .602 | .522 | .545 | .428 |
DOC-DMLFOV | .579 | .609 | .519 | .463 | .491 | .369 |
DOOBNet | .647 | .668 | .539 | .555 | .570 | .440 |
OFNet | .662 | .689 | .585 | .583 | .607 | .501 |
FSINet | .657 | .692 | .598 | .591 | .620 | .515 |