Please use this identifier to cite or link to this item: https://hdl.handle.net/2440/83991
Type: Conference paper
Title: Semantic parsing for priming object detection in RGB-D Scenes
Author: Cadena Lerma, C.
Kosecka, J.
Citation: 3rd Workshop on Semantic Perception, Mapping and Exploration (SPME), 2013 / pp.1-6
Publisher: SPME
Publisher Place: Online
Issue Date: 2013
Conference Name: Semantic Perception, Mapping and Exploration (2013 : Karlsruhe, Germany)
Statement of
Responsibility: 
César Cadena and Jana Kǒsecka
Abstract: The advancements in robot autonomy and capabilities for carrying out more complex tasks in unstructured indoors environments can be greatly enhanced by endowing existing environment models with semantic information. In this paper we describe an approach for semantic parsing of indoors environments into semantic categories of Ground, Structure, Furniture and Props. Instead of striving to categorize all object classes and instances encountered in the environment, this choice of semantic labels separates clearly objects and nonobject categories. We use RGB-D images of indoors environments and formulate the problem of semantic segmentation in the Conditional Random Fields Framework. The appearance and depth information enables us induce the graph structure of the random field, which can be effectively approximated by a tree, and to design robust geometric features, which are informative for separation and characterization of different categories. These two choices notably improve the efficiency and performance of the semantic parsing tasks. We carry out the experiments on a NYU V2 dataset and achieve superior or comparable performance and the fraction of computational cost.
Rights: Copyright status unknown
Description (link): http://www.spme.ws/2013
Published version: http://www.spme.ws/spme13_cadena.pdf?attredirects=0
Appears in Collections:Aurora harvest
Computer Science publications

Files in This Item:
There are no files associated with this item.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.