- Notifications
You must be signed in to change notification settings - Fork117
UPSNet: A Unified Panoptic Segmentation Network
License
uber-research/UPSNet
Folders and files
| Name | Name | Last commit message | Last commit date | |
|---|---|---|---|---|
Repository files navigation
UPSNet is initially described in aCVPR 2019 oral paper.
This repository is tested under Python 3.6, PyTorch 0.4.1. And model training is done with 16 GPUs by usinghorovod. It should also work under Python 2.7 / PyTorch 1.0 and with 4 GPUs.
© Uber, 2018-2019. Licensed under the Uber Non-Commercial License.
If you find UPSNet is useful in your research, please consider citing:
@inproceedings{xiong19upsnet, Author = {Yuwen Xiong, Renjie Liao, Hengshuang Zhao, Rui Hu, Min Bai, Ersin Yumer, Raquel Urtasun}, Title = {UPSNet: A Unified Panoptic Segmentation Network}, Conference = {CVPR}, Year = {2019}}COCO 2017 (trained on train-2017 set)
| test split | PQ | SQ | RQ | PQTh | PQSt | |
|---|---|---|---|---|---|---|
| UPSNet-50 | val | 42.5 | 78.0 | 52.4 | 48.5 | 33.4 |
| UPSNet-101-DCN | test-dev | 46.6 | 80.5 | 56.9 | 53.2 | 36.7 |
Cityscapes
| PQ | SQ | RQ | PQTh | PQSt | |
|---|---|---|---|---|---|
| UPSNet-50 | 59.3 | 79.7 | 73.0 | 54.6 | 62.7 |
| UPSNet-101-COCO (ms test) | 61.8 | 81.3 | 74.8 | 57.6 | 64.8 |
We recommend using Anaconda3 as it already includes many common packages.
We recommend using 4~16 GPUs with at least 11 GB memory to train our model.
Clone this repo to$UPSNet_ROOT
Runinit.sh to build essential C++/CUDA modules and download pretrained model.
For Cityscapes:
Assuming you already downloaded Cityscapes dataset at$CITYSCAPES_ROOT and TrainIds label images are generated, please create a soft link byln -s $CITYSCAPES_ROOT data/cityscapes underUPSNet_ROOT, and runinit_cityscapes.sh to prepare Cityscapes dataset for UPSNet.
For COCO:
Assuming you already downloaded COCO dataset at$COCO_ROOT and haveannotations andimages folders under it, please create a soft link byln -s $COCO_ROOT data/coco underUPSNet_ROOT, and runinit_coco.sh to prepare COCO dataset for UPSNet.
Training:
python upsnet/upsnet_end2end_train.py --cfg upsnet/experiments/$EXP.yaml
Test:
python upsnet/upsnet_end2end_test.py --cfg upsnet/experiments/$EXP.yaml
We provide serveral config files (16/4 GPUs for Cityscapes/COCO dataset) under upsnet/experiments folder.
The model weights that can reproduce numbers in our paper are available now. Please follow these steps to use them:
Rundownload_weights.sh to get trained model weights for Cityscapes and COCO.
For Cityscapes:
python upsnet/upsnet_end2end_test.py --cfg upsnet/experiments/upsnet_resnet50_cityscapes_16gpu.yaml --weight_path ./model/upsnet_resnet_50_cityscapes_12000.pth
python upsnet/upsnet_end2end_test.py --cfg upsnet/experiments/upsnet_resnet101_cityscapes_w_coco_16gpu.yaml --weight_path ./model/upsnet_resnet_101_cityscapes_w_coco_3000.pth
For COCO:
python upsnet/upsnet_end2end_test.py --cfg upsnet/experiments/upsnet_resnet50_coco_16gpu.yaml --weight_path model/upsnet_resnet_50_coco_90000.pth
python upsnet/upsnet_end2end_test.py --cfg upsnet/experiments/upsnet_resnet101_dcn_coco_3x_16gpu.yaml --weight_path model/upsnet_resnet_101_dcn_coco_270000.pth
About
UPSNet: A Unified Panoptic Segmentation Network
Topics
Resources
License
Uh oh!
There was an error while loading.Please reload this page.
Stars
Watchers
Forks
Releases
Packages0
Uh oh!
There was an error while loading.Please reload this page.