Semantic Amodal Segmentation

09/04/2015
by   Yan Zhu, et al.
0

Common visual recognition tasks such as classification, object detection, and semantic segmentation are rapidly reaching maturity, and given the recent rate of progress, it is not unreasonable to conjecture that techniques for many of these problems will approach human levels of performance in the next few years. In this paper we look to the future: what is the next frontier in visual recognition? We offer one possible answer to this question. We propose a detailed image annotation that captures information beyond the visible pixels and requires complex reasoning about full scene structure. Specifically, we create an amodal segmentation of each image: the full extent of each region is marked, not just the visible pixels. Annotators outline and name all salient regions in the image and specify a partial depth order. The result is a rich scene structure, including visible and occluded portions of each region, figure-ground edge information, semantic labels, and object overlap. We create two datasets for semantic amodal segmentation. First, we label 500 images in the BSDS dataset with multiple annotators per image, allowing us to study the statistics of human annotations. We show that the proposed full scene annotation is surprisingly consistent between annotators, including for regions and edges. Second, we annotate 5000 images from COCO. This larger dataset allows us to explore a number of algorithmic ideas for amodal segmentation and depth ordering. We introduce novel metrics for these tasks, and along with our strong baselines, define concrete new challenges for the community.

READ FULL TEXT

page 1

page 2

page 3

page 5

page 7

page 8

page 10

page 11

research
12/10/2020

Amodal Segmentation Based on Visible Region Segmentation and Shape Prior

Almost all existing amodal segmentation methods make the inferences of o...
research
12/27/2018

Finite State Machines for Semantic Scene Parsing and Segmentation

We introduce in this work a novel stochastic inference process, for scen...
research
06/10/2021

3D Semantic Mapping from Arthroscopy using Out-of-distribution Pose and Depth and In-distribution Segmentation Training

Minimally invasive surgery (MIS) has many documented advantages, but the...
research
12/12/2016

COCO-Stuff: Thing and Stuff Classes in Context

Semantic classes can be either things (objects with a well-defined shape...
research
11/29/2011

Picture Collage with Genetic Algorithm and Stereo vision

In this paper, a salient region extraction method for creating picture c...
research
06/20/2018

Fluid Annotation: a human-machine collaboration interface for full image annotation

We introduce Fluid Annotation, an intuitive human-machine collaboration ...
research
09/27/2019

Invisible Marker: Automatic Annotation for Object Manipulation

We propose invisible marker for accurate automatic annotation to manipul...

Please sign up or login with your details

Forgot password? Click here to reset