Incorporating Depth Information into Few-Shot Semantic Segmentation
Affiliation auteurs | !!!! Error affiliation !!!! |
Titre | Incorporating Depth Information into Few-Shot Semantic Segmentation |
Type de publication | Conference Paper |
Year of Publication | 2021 |
Auteurs | Zhang Y, Sidibe D, Morel O, Meriaudeau F |
Conference Name | 2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR) |
Publisher | Int Assoc Pattern Recognit; IEEE Comp Soc; Italian Assoc Comp Vis Pattern Recognit & Machine Learning |
Conference Location | 10662 LOS VAQUEROS CIRCLE, PO BOX 3014, LOS ALAMITOS, CA 90720-1264 USA |
ISBN Number | 978-1-7281-8808-9 |
Résumé | Few-shot segmentation presents a significant challenge for semantic scene understanding under limited supervision. Namely, this task targets at generalizing the segmentation ability of the model to new categories given a few samples. In order to obtain complete scene information, we extend the RGB-centric methods to take advantage of complementary depth information. In this paper, we propose a two-stream deep neural network based on metric learning. Our method, known as RDNet, learns class-specific prototype representations within RGB and depth embedding spaces, respectively. The learned prototypes provide effective semantic guidance on the corresponding RGB and depth query image, leading to more accurate performance. Moreover, we build a novel outdoor scene dataset, known as Cityscapes-3(i), using labeled RGB images and depth images from the Cityscapes dataset. We also perform ablation studies to explore the effective use of depth information in few-shot segmentation tasks. Experiments on Cityscapes-3(i) show that our method achieves excellent results with visual and complementary geometric cues from only a few labeled examples. |
DOI | 10.1109/ICPR48806.2021.9412921 |