Skip to content

Latest commit

 

History

History
48 lines (35 loc) · 2.13 KB

README.md

File metadata and controls

48 lines (35 loc) · 2.13 KB

ShapeFormer: Shape Prior Visible-to-Amodal Transformer-based Amodal Instance Segmentation

Table of Contents

Introduction

alt text

The figure above illustrates our ShapeFormer architecture. The main implementation of this network can be found here.

Usage

1. Installation

This work uses aistron for implementation. Please follow the instruction at here

2. Data preparation

Please download the below datasets. More preparation instruction can be found here.

2.1. KINS dataset

Download the Images from KITTI dataset.

The Amodal Annotations could be found at KINS dataset

2.2. D2SA dataset

The D2S Amodal dataset could be found at mvtec-d2sa.

2.3. COCOA-cls dataset

The COCOA dataset annotation from here (reference from github.com/YihongSun/Bayesian-Amodal) The images of COCOA dataset is the train2014 and val2014 of COCO dataset.

3. Training, Testing and Demo

Configuration files for training AISFormer on each datasets are available here. To train, test and run demo, see the example scripts at scripts/:

Citation

@article{tran2024shapeformer,
  title={ShapeFormer: Shape Prior Visible-to-Amodal Transformer-based Amodal Instance Segmentation},
  author={Tran, Minh and Bounsavy, Winston and Vo, Khoa and Nguyen, Anh and Nguyen, Tri and Le, Ngan},
  journal={arXiv preprint arXiv:2403.11376},
  year={2024}
}