demo.mp4
Panoptic Video Scene Graph Generation
Jingkang Yang,
Wenxuan Peng,
Xiangtai Li,
Zujin Guo,
Liangyu Chen,
Bo Li,
Zheng Ma,
Kaiyang Zhou,
Wayne Zhang,
Chen Change Loy,
Ziwei Liu,
S-Lab, Nanyang Technological University & SenseTime Research
The Panoptic Video Scene Graph Generation (PVSG) Task aims to interpret a complex scene video with a dynamic scene graph representation, with each node in the scene graph grounded by its pixel-accurate segmentation mask tube in the video.
Given a video, PVSG models need to generate a dynamic (temporal) scene graph that is grounded by panoptic mask tubes. |
We carefully collect 400 videos, each featuring dynamic scenes and rich in logical reasoning content. On average, these videos are 76.5 seconds long (5 FPS). The collection comprises 289 videos from VidOR, 55 videos from EpicKitchen, and 56 videos from Ego4D.
Please access the dataset via this link, and put the downloaded zip files to the place below.
├── assets
├── checkpoints
├── configs
├── data
├── data_zip
│ ├── Ego4D
│ │ ├── ego4d_masks.zip
│ │ └── ego4d_videos.zip
│ ├── EpicKitchen
│ │ ├── epic_kitchen_masks.zip
│ │ └── epic_kitchen_videos.zip
│ ├── VidOR
│ │ ├── vidor_masks.zip
│ │ └── vidor_videos.zip
│ └── pvsg.json
├── datasets
├── models
├── scripts
├── tools
├── utils
├── .gitignore
├── environment.yml
└── README.md
Please run unzip_and_extract.py
to unzip the files and extract frames from the videos. If you use zip
, make sure to use unzip -j xxx.zip
to remove junk paths. You should have your data
directory looks like this:
data
├── ego4d
│ ├── frames
│ ├── masks
│ └── videos
├── epic_kitchen
│ ├── frames
│ ├── masks
│ └── videos
├── vidor
│ ├── frames
│ ├── masks
│ └── videos
└── pvsg.json
We suggest our users to play with ./tools/Visualize_Dataset.ipynb
to quickly get familiar with PSG dataset.
To setup the environment, we use conda
to manage our dependencies.
Our developers use CUDA 10.1
to do experiments.
You can specify the appropriate cudatoolkit
version to install on your machine in the environment.yml
file, and then run the following to create the conda
environment:
conda env create -f environment.yml
conda activate openpvsg
You shall manually install the following dependencies.
# Install mmcv
pip install mmcv-full==1.4.0 -f https://download.openmmlab.com/mmcv/dist/cu101/torch1.7.0/index.html
conda install -c conda-forge pycocotools
pip install mmdet==2.25.0
# already within environment.yml
pip install timm
python -m pip install scipy
pip install git+https://github.com/cocodataset/panopticapi.git
# for unitrack
pip install imageio==2.6.1
pip install lap==0.4.0
pip install cython_bbox==0.1.3
# for vps
pip install seaborn
pip install ftfy
pip install regex
# If you're using wandb for logging
pip install wandb
wandb login
Download the pretrained models for tracking if you are interested in IPS+Tracking solution.
IPS+Tracking & Relation Modeling
# Train IPS
sh scripts/train/train_ips.sh
# Tracking and save query features
sh scripts/utils/prepare_qf_ips.sh
# Prepare for relation modeling
sh scripts/utils/prepare_rel_set.sh
# Train relation models
sh scripts/train/train_relation.sh
# Test
sh scripts/test/test_relation_full.sh
VPS & Relation Modeling
# Train VPS
sh scripts/train/train_vps.sh
# Save query features
sh scripts/utils/prepare_qf_vps.sh
# Prepare for relation modeling
sh scripts/utils/prepare_rel_set.sh
# Train relation models
sh scripts/train/train_relation.sh
# Test
sh scripts/test/test_relation_full.sh
Method | M2F ckpt | vanilla | filter | conv | transformer |
---|---|---|---|---|---|
mask2former_ips | link | link | link | link | link |
mask2former_vps | link | link | link | link | link |
If you find our repository useful for your research, please consider citing our paper:
@inproceedings{yang2023pvsg,
author = {Yang, Jingkang and Peng, Wenxuan and Li, Xiangtai and Guo, Zujin and Chen, Liangyu and Li, Bo and Ma, Zheng and Zhou, Kaiyang and Zhang, Wayne and Loy, Chen Change and Liu, Ziwei},
title = {Panoptic Video Scene Graph Generation},
booktitle = {CVPR},
year = {2023},
}