# DVIS_Plus **Repository Path**: lytlm1994/DVIS_Plus ## Basic Information - **Project Name**: DVIS_Plus - **Description**: No description available - **Primary Language**: Python - **License**: MIT - **Default Branch**: main - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 0 - **Created**: 2024-01-24 - **Last Updated**: 2024-01-24 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README
# [DVIS++: Improved Decoupled Framework for Universal Video Segmentation](https://arxiv.org/abs/2312.13305) [Tao Zhang](https://scholar.google.com/citations?user=3xu4a5oAAAAJ&hl=zh-CN), XingYe Tian, Yikang Zhou, [ShunPing Ji](https://scholar.google.com/citations?user=FjoRmF4AAAAJ&hl=zh-CN), Xuebo Wang, Xin Tao, Yuan Zhang, Pengfei Wan, Zhongyuan Wang and [Yu Wu](https://scholar.google.com/citations?hl=zh-CN&user=23SZHUwAAAAJ) [![PWC](https://img.shields.io/endpoint.svg?url=https://paperswithcode.com/badge/dvis-improved-decoupled-framework-for/video-instance-segmentation-on-ovis-1)](https://paperswithcode.com/sota/video-instance-segmentation-on-ovis-1?p=dvis-improved-decoupled-framework-for) [![PWC](https://img.shields.io/endpoint.svg?url=https://paperswithcode.com/badge/dvis-improved-decoupled-framework-for/video-instance-segmentation-on-youtube-vis-1)](https://paperswithcode.com/sota/video-instance-segmentation-on-youtube-vis-1?p=dvis-improved-decoupled-framework-for) [![PWC](https://img.shields.io/endpoint.svg?url=https://paperswithcode.com/badge/dvis-improved-decoupled-framework-for/video-instance-segmentation-on-youtube-vis-2)](https://paperswithcode.com/sota/video-instance-segmentation-on-youtube-vis-2?p=dvis-improved-decoupled-framework-for) [![PWC](https://img.shields.io/endpoint.svg?url=https://paperswithcode.com/badge/dvis-improved-decoupled-framework-for/video-instance-segmentation-on-youtube-vis-3)](https://paperswithcode.com/sota/video-instance-segmentation-on-youtube-vis-3?p=dvis-improved-decoupled-framework-for) [![PWC](https://img.shields.io/endpoint.svg?url=https://paperswithcode.com/badge/dvis-improved-decoupled-framework-for/video-semantic-segmentation-on-vspw)](https://paperswithcode.com/sota/video-semantic-segmentation-on-vspw?p=dvis-improved-decoupled-framework-for) [![PWC](https://img.shields.io/endpoint.svg?url=https://paperswithcode.com/badge/dvis-improved-decoupled-framework-for/video-panoptic-segmentation-on-vipseg)](https://paperswithcode.com/sota/video-panoptic-segmentation-on-vipseg?p=dvis-improved-decoupled-framework-for)
## News - DVIS and DVIS++ achieved **1st place** in the VPS Track of the PVUW challenge at CVPR 2023. `2023.5.25` - DVIS and DVIS++ achieved **1st place** in the VIS Track of the 5th LSVOS challenge at ICCV 2023. `2023.8.15` ## Features - DVIS++ is a universal video segmentation framework that supports VIS, VPS and VSS. - DVIS++ can run in both online and offline modes. - DVIS++ achieved SOTA performance on YTVIS 2019&2021&2022, OVIS, VIPSeg and VSPW datasets. - OV-DVIS++ is the first open-vocabulary video universal segmentation framework with powerful zero-shot segmentation capability. ## Demos ### VIS ### VSS ### VPS ### Open-vocabulary demos ## Installation See [Installation Instructions](INSTALL.md). ## Getting Started See [Preparing Datasets for DVIS++](datasets/README.md). See [Getting Started with DVIS++](GETTING_STARTED.md). ## Model Zoo Trained models are available for download in the [DVIS++ Model Zoo](MODEL_ZOO.md). ## Citing DVIS and DVIS++ ```BibTeX @article{zhang2023dvis, title={DVIS: Decoupled Video Instance Segmentation Framework}, author={Zhang, Tao and Tian, Xingye and Wu, Yu and Ji, Shunping and Wang, Xuebo and Zhang, Yuan and Wan, Pengfei}, journal={arXiv preprint arXiv:2306.03413}, year={2023} } @article{zhang2023dvisplus, title={DVIS++: Improved Decoupled Framework for Universal Video Segmentation}, author={Tao Zhang and Xingye Tian and Yikang Zhou and Shunping Ji and Xuebo Wang and Xin Tao and Yuan Zhang and Pengfei Wan and Zhongyuan Wang and Yu Wu}, journal={arXiv preprint arXiv:2312.13305}, year={2023}, } ``` ## Acknowledgement This repo is largely based on [Mask2Former](https://github.com/facebookresearch/Mask2Former), [MinVIS](https://github.com/NVlabs/MinVIS), [VITA](https://github.com/sukjunhwang/VITA), [CTVIS](https://github.com/KainingYing/CTVIS), [FC-CLIP](https://github.com/bytedance/fc-clip) and [DVIS](https://github.com/zhang-tao-whu/DVIS). Thanks for their excellent works.