Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up

[ICCV2023] Segment Every Reference Object in Spatial and Temporal Spaces

License

NotificationsYou must be signed in to change notification settings

FoundationVision/UniRef

Repository files navigation

Official implementation ofUniRef++, an extended version of ICCV2023UniRef.

UniRef

Highlights

  • UniRef/UniRef++ is a unified model for four object segmentation tasks, namely referring image segmentation (RIS), few-shot segmentation (FSS), referring video object segmentation (RVOS) and video object segmentation (VOS).
  • At the core of UniRef++ is the UniFusion module for injecting various reference information into network. And we implement it using flash attention with high efficiency.
  • UniFusion could play as the plug-in component for foundation models likeSAM.

Schedule

  • Add Training Guide
  • Add Evaluation Guide
  • Add Data Preparation
  • Release Model Checkpoints
  • Release Code

Results

video_demo.mp4

Referring Image Segmentation

RIS

Referring Video Object Segmentation

RVOS

Video Object Segmentation

VOS

Zero-shot Video Segmentation & Few-shot Image Segmentation

zero-few-shot

Model Zoo

Objects365 Pretraining

ModelCheckpoint
R50model
Swin-Lmodel

Imge-joint Training

ModelRefCOCOFSS-1000Checkpoint
R5076.385.2model
Swin-L79.987.7model

Video-joint Training

The results are reported on the validation set.

ModelRefCOCOFSS-1000Ref-Youtube-VOSRef-DAVIS17Youtube-VOS18DAVIS17LVOSCheckpoint
UniRef++-R5075.679.161.563.581.981.560.1model
UniRef++-Swin-L79.185.466.967.283.283.967.2model

Installation

SeeINSTALL.md

Getting Started

Please seeDATA.md for data preparation.

Please seeEVAL.md for evaluation.

Please seeTRAIN.md for training.

Citation

If you find this project useful in your research, please consider cite:

@article{wu2023uniref++,title={UniRef++: Segment Every Reference Object in Spatial and Temporal Spaces},author={Wu, Jiannan and Jiang, Yi and Yan, Bin and Lu, Huchuan and Yuan, Zehuan and Luo, Ping},journal={arXiv preprint arXiv:2312.15715},year={2023}}
@inproceedings{wu2023uniref,title={Segment Every Reference Object in Spatial and Temporal Spaces},author={Wu, Jiannan and Jiang, Yi and Yan, Bin and Lu, Huchuan and Yuan, Zehuan and Luo, Ping},booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision},pages={2538--2550},year={2023}}

Acknowledgement

The project is based onUNINEXT codebase. We also refer to the repositoriesDetectron2,Deformable DETR,STCN,SAM. Thanks for their awsome works!

About

[ICCV2023] Segment Every Reference Object in Spatial and Temporal Spaces

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages


[8]ページ先頭

©2009-2025 Movatter.jp