Show simple item record

dc.contributor.authorHu, J
dc.contributor.authorLin, J
dc.contributor.authorGong, S
dc.contributor.authorCai, W
dc.date.accessioned2024-05-17T14:50:26Z
dc.date.available2024-05-17T14:50:26Z
dc.date.issued2024-03-25
dc.identifier.issn2159-5399
dc.identifier.urihttps://qmro.qmul.ac.uk/xmlui/handle/123456789/96948
dc.description.abstractCamouflaged object detection (COD) approaches heavily rely on pixel-level annotated datasets. Weakly-supervised COD (WSCOD) approaches use sparse annotations like scribbles or points to reduce annotation efforts, but this can lead to decreased accuracy. The Segment Anything Model (SAM) shows remarkable segmentation ability with sparse prompts like points. However, manual prompt is not always feasible, as it may not be accessible in real-world application. Additionally, it only provides localization information instead of semantic one, which can intrinsically cause ambiguity in interpreting targets. In this work, we aim to eliminate the need for manual prompt. The key idea is to employ Cross-modal Chains of Thought Prompting (CCTP) to reason visual prompts using the semantic information given by a generic text prompt. To that end, we introduce a test-time instance-wise adaptation mechanism called Generalizable SAM (GenSAM) to automatically generate and optimize visual prompts from the generic task prompt for WSCOD. In particular, CCTP maps a single generic text prompt onto image-specific consensus foreground and background heatmaps using vision-language models, acquiring reliable visual prompts. Moreover, to test-time adapt the visual prompts, we further propose Progressive Mask Generation (PMG) to iteratively reweight the input image, guiding the model to focus on the targeted region in a coarse-to-fine manner. Crucially, all network parameters are fixed, avoiding the need for additional training. Experiments on 3 benchmarks demonstrate that GenSAM outperforms point supervision approaches and achieves comparable results to scribble supervision ones, solely relying on general task descriptions. Our codes is in https://github.com/jyLin8100/GenSAM.en_US
dc.format.extent12511 - 12518
dc.publisherAssociation for the Advancement of Artificial Intelligenceen_US
dc.rightsThis is a pre-copyedited, author-produced version accepted for publication in Proceedings of the AAAI Conference on Artificial Intelligence following peer review. The version of record is available at https://ojs.aaai.org/index.php/AAAI/article/view/29144
dc.titleRelax Image-Specific Prompt Requirement in SAM: A Single Generic Prompt for Segmenting Camouflaged Objectsen_US
dc.typeConference Proceedingen_US
dc.rights.holder© 2024, Association for the Advancement of Artificial Intelligence
dc.identifier.doi10.1609/aaai.v38i11.29144
pubs.issue11en_US
pubs.notesNot knownen_US
pubs.publication-statusPublisheden_US
pubs.volume38en_US
rioxxterms.funderDefault funderen_US
rioxxterms.identifier.projectDefault projecten_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record