Semi-supervised video object segmentation targets at segmenting a particular object instance throughout the entire video sequence given only the object mask of the first frame. Different from previous video object segmentation datasets such as DAVIS, we will provide much larger scale of training and test data to foster various kinds of algorithms. In addition, our test dataset will have unseen categories which do not exist in the training dataset, in order to evaluate the generalization ability of algorithms.
Our dataset for video object segmentation was first released in 2018 in conjunction with a workshop challenge. In 2019, we further augment the dataset with more videos and object annotations (a subset of this dataset is carried out for the task of video instance segmentation).
The 2018 version was used for “the 1st large-scale video object segmentation challenge” at ECCV 2018. It contains three subsets.
- Training: 3471 videos with dense (6 fps) object annotations, 65 categories and 5945 unique object instances.
- Validation: 474 videos, 65 training (seen) categories, 26 unseen categories and 894 unique object instances.
- Test: 508 videos, 65 training (seen) categories, 29 unseen categories and 915 unique object instances.
The 2019 version will be used for “the 2nd large-scale video object segmentation challenge” at ICCV 2019”. It augments the 2018 version with more videos and annotations.
- Training: 3471 videos, 65 categories and 6459 unique object instances.
- Validation: 507 videos, 65 training categories, 26 unseen categories and 1063 unique object instances.
- Test: 541 videos, 65 training categories, 29 unseen categories and 1092 unique object instances.
Following previous video object segmentation challenge DAVIS, we use Region Jaccard (J) and Boundary F measure (F) as our evaluation metrics. Differently, we evaluate the metrics on different sets of categories, i.e. seen and unseen, in order to evaluate the generalization ability of proposed methods. In detail,
- We compute J and F on both seen and unseen categories respectively, averaged over all corresponding objects.
- The final score is the average of the four scores: J for seen categories, F for seen categories, J for unseen categories, and F for unseen categories.
Note that in our videos, some objects first appear at middle frames instead of the first frame, be sure not to miss them!
We have set up evaluation servers on CodaLab for the convenience of evaluating new algorithms. For more details of how to submit your results, please check the following links.
- Evaluation server for the 2019 version (will be released soon).
- Evaluation server for the 2018 version.
- 2019 version (will be released soon).
- 2018 version.