Abstract
In order to learn object segmentation models in videos, conventional methods require a large amount of pixel-wise ground truth annotations. However, collecting such supervised data is time-consuming and labor-intensive. In this paper, we exploit existing annotations in source images and transfer such visual information to segment videos with unseen object categories. Without using any annotations in the target video, we propose a method to jointly mine useful segments and learn feature representations that better adapt to the target frames. The entire process is decomposed into two tasks: (1) solving a submodular function for selecting object-like segments, and (2) learning a CNN model with a transferable module for adapting seen categories in the source domain to the unseen target video. We present an iterative update scheme between two tasks to self-learn the final solution for object segmentation. Experimental results on numerous benchmark datasets show that the proposed method performs favorably against the state-of-the-art algorithms.
Original language | English |
---|---|
Title of host publication | Computer Vision – ACCV 2018 - 14th Asian Conference on Computer Vision, Revised Selected Papers |
Editors | Hongdong Li, C.V. Jawahar, Greg Mori, Konrad Schindler |
Publisher | Springer Verlag |
Pages | 615-631 |
Number of pages | 17 |
ISBN (Print) | 9783030208691 |
DOIs | |
Publication status | Published - 2019 |
Event | 14th Asian Conference on Computer Vision, ACCV 2018 - Perth, Australia Duration: 2018 Dec 2 → 2018 Dec 6 |
Publication series
Name | Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) |
---|---|
Volume | 11364 LNCS |
ISSN (Print) | 0302-9743 |
ISSN (Electronic) | 1611-3349 |
Conference
Conference | 14th Asian Conference on Computer Vision, ACCV 2018 |
---|---|
Country/Territory | Australia |
City | Perth |
Period | 18/12/2 → 18/12/6 |
Bibliographical note
Funding Information:Acknowledgments. This work is supported in part by Ministry of Science and Technology under grants MOST 105-2221-E-001-030-MY2 and MOST 107-2628-E-001-005-MY3.
Funding Information:
This work is supported in part by Ministry of Science and Technology under grants MOST 105-2221-E-001-030-MY2 and MOST 107-2628-E-001-005-MY3.
Publisher Copyright:
© 2019, Springer Nature Switzerland AG.
All Science Journal Classification (ASJC) codes
- Theoretical Computer Science
- Computer Science(all)