Efficient RGB-D semantic segmentation has received considerable attention in mobile robots, which plays a vital role in analyzing and recognizing environmental information.
SegPrompt: Boosting Open-world Segmentation via Category-level Prompt Learning
In this work, we propose a novel training mechanism termed SegPrompt that uses category information to improve the model's class-agnostic segmentation ability for both known and unknown categories.
Description of SegPrompt: Boosting Open-world Segmentation via Category-level Prompt Learning
GitHub Link
The GitHub link is https://github.com/aim-uofa/segpromptIntroduce
The repository "aim-uofa/SegPrompt" contains the official implementation of the ICCV 2023 paper titled "SegPrompt Boosting Open-World Segmentation via Category-level Prompt Learning." The authors propose SegPrompt for improving open-world segmentation through category-level prompt learning. They introduce a new benchmark called LVIS-OW, which involves reorganizing COCO and LVIS datasets into Known-Seen-Unseen categories for better evaluating open-world models. The repository provides dataset preparation instructions, benchmark details, and evaluation scripts. Acknowledgments are given to related repositories like Mask2Former and Detectron2, and the paper encourages proper citation if the project is used. In this work, we propose a novel training mechanism termed SegPrompt that uses category information to improve the model's class-agnostic segmentation ability for both known and unknown categories.Content
1Zhejiang University, 2The University of Adelaide, Please follow the instructions in Mask2Former Here we provide our proposed new benchmark LVIS-OW. First prepare COCO and LVIS dataset, place them under $DETECTRON2_DATASETS following Detectron2 The dataset structure is as follows: Or you can directly use the command to generate from the json file of COCO and LVIS. We thank the following repos for their great works: If you found this project useful for your paper, please kindly cite our paper.Alternatives & Similar Tools
Recent leading zero-shot video object segmentation (ZVOS) works devote to integrating appearance and motion information by elaborately designing feature fusion modules and identically applying them in multiple feature stages.
Google Gemini, a multimodal AI by DeepMind, processes text, audio, images, and more. Gemini outperforms in AI benchmarks, is optimized for varied devices, and has been tested for safety and bias, adhering to responsible AI practices.
Cerelyze - Enabling engineers to rapidly reproduce scientific research
Video ReTalking, advanced real-world talking head video according to input audio, producing a high-quality
Then transplant it to the real world to solve complex problems
Compare SegPrompt: Boosting Open-world Segmentation via Category-level Prompt Learning
Quick compare routes for nearby alternatives.
SegPrompt: Boosting Open-world Segmentation via Category-level Prompt Learning vs Spatial-information Guided Adaptive Context-aware Network for Efficient RGB-D Semantic Segmentation
Compare SegPrompt: Boosting Open-world Segmentation via Category-level Prompt Learning with Spatial-information Guided Adaptive Context-aware Network for Efficient RGB-D Semantic Segmentation and jump into the preserved compare route.
Open compare route →SegPrompt: Boosting Open-world Segmentation via Category-level Prompt Learning vs Isomer: Isomerous Transformer for Zero-shot Video Object Segmentation
Compare SegPrompt: Boosting Open-world Segmentation via Category-level Prompt Learning with Isomer: Isomerous Transformer for Zero-shot Video Object Segmentation and jump into the preserved compare route.
Open compare route →SegPrompt: Boosting Open-world Segmentation via Category-level Prompt Learning vs Free Google Gemini: the best largest and most capable AI model
Compare SegPrompt: Boosting Open-world Segmentation via Category-level Prompt Learning with Free Google Gemini: the best largest and most capable AI model and jump into the preserved compare route.
Open compare route →