Please use this identifier to cite or link to this item: http://hdl.handle.net/2440/116990
Citations
Scopus Web of Science® Altmetric
?
?
Type: Conference paper
Title: Bayesian semantic instance segmentation in open set world
Author: Pham, T.
Vijay Kumar, B.
Do, T.
Carneiro, G.
Reid, I.
Citation: Computer Vision - ECCV 2018 15th European Conference: Proceedings, Part I, 2018 / Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (ed./s), vol.11214 LNCS, pp.3-18
Publisher: Springer
Issue Date: 2018
Series/Report no.: Lecture Notes in Computer Science; 11205
ISBN: 9783030012458
ISSN: 0302-9743
1611-3349
Conference Name: European Conference on Computer Vision (ECCV) (08 Sep 2018 - 14 Sep 2018 : Munich)
Abstract: This paper addresses the semantic instance segmentation task in the open-set conditions, where input images can contain known and unknown object classes. The training process of existing semantic instance segmentation methods requires annotation masks for all object instances, which is expensive to acquire or even infeasible in some realistic scenarios, where the number of categories may increase boundlessly. In this paper, we present a novel open-set semantic instance segmentation approach capable of segmenting all known and unknown object classes in images, based on the output of an object detector trained on known object classes. We formulate the problem using a Bayesian framework, where the posterior distribution is approximated with a simulated annealing optimization equipped with an efficient image partition sampler. We show empirically that our method is competitive with state-of-the-art supervised methods on known classes, but also performs well on unknown classes when compared with unsupervised methods.
Rights: © Springer Nature Switzerland AG 2018
RMID: 0030102452
DOI: 10.1007/978-3-030-01249-6_1
Appears in Collections:Australian Institute for Machine Learning publications
Computer Science publications

Files in This Item:
There are no files associated with this item.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.