Fig .1: Proposed Pixel-GCN framework for Video Object Segmentation.
Table 1: Quantitative comparison of the proposed Pixel-GCN with existing state-ofthe-art methods on DAVIS-2016 and DAVIS-2017 validation datasets for VOS. “ ↑ ” - higher is better. “ ∗ ” - w/o proposed aggregation functions (Amotion & Aocc) in rows 9 & 10 (Best results in bold).
Table 2: Quantitative comparison of various models on CamVid dataset having features as input for semantic segmentation in videos. “FS”: Feature → Segmentation, whereas “FF”: Feature → Feature. “ ∗ ” - without proposed aggregation functions (A_motion & A_occ) in rows 2 & 3
(Best results in bold). MO refers to the moving objects categories and SEG represents segmentation.
RGB
PReMVOS (ACCV, 2018) [2]
Pixel-GCN (Ours)
(ICONIP, 2019)
Ground Truth
RGB
PReMVOS (ACCV, 2018) [2]
Pixel-GCN (Ours)
(ICONIP, 2019)
Ground Truth
RGB DeepLab V3+ Baseline (ECCV, 2018) [3] Pixel-GCN (Ours) (ICONIP, 2019)
RGB
Ground Truth
VidSeg-GAN (ICVGIP, 2018) [4]
TempSeg-GAN (VISAPP, 2019) [5]
Pixel-GCN (ICONIP, 2019)