Visualizing Promptable and Open-Vocabulary Segmentation Across Multiple Datasets

Written by segmentation | Published 2024/11/13
Tech Story Tags: vision-language-model | open-vocabulary-segmentation | uni-ovseg | image-mask-pairs | image-text-pairs | clip-embedding | image-mask-text-triplets | uni-ovseg-framework

TLDRThis section presents visualizations of promptable and open-vocabulary segmentation results across several datasets, showcasing the framework's performance in segmenting and predicting masks.via the TL;DR App

Authors:

(1) Zhaoqing Wang, The University of Sydney and AI2Robotics;

(2) Xiaobo Xia, The University of Sydney;

(3) Ziye Chen, The University of Melbourne;

(4) Xiao He, AI2Robotics;

(5) Yandong Guo, AI2Robotics;

(6) Mingming Gong, The University of Melbourne and Mohamed bin Zayed University of Artificial Intelligence;

(7) Tongliang Liu, The University of Sydney.

Table of Links

Abstract and 1. Introduction

2. Related works

3. Method and 3.1. Problem definition

3.2. Baseline and 3.3. Uni-OVSeg framework

4. Experiments

4.1. Implementation details

4.2. Main results

4.3. Ablation study

5. Conclusion

6. Broader impacts and References

A. Framework details

B. Promptable segmentation

C. Visualisation

C. Visualisation

We illustrate a wide range of visualisations of promptable segmentation and open-vocabulary segmentation across multiple datasets.

This paper is available on arxiv under CC BY 4.0 DEED license.


Written by segmentation | I break groups into smaller groups by common traits
Published by HackerNoon on 2024/11/13