RobustSAM: Segment Anything Robustly on Degraded Images

Wei Ting Chen, Yu Jiet Vong, Sy Yen Kuo, Sizhuo Ma*, Jian Wang*

*此作品的通信作者

研究成果: 期刊稿件會議文章同行評審

2 引文 斯高帕斯(Scopus)

摘要

Segment Anything Model (SAM) has emerged as a transformative approach in image segmentation, acclaimed for its robust zero-shot segmentation capabilities and flexible prompting system. Nonetheless, its performance is challenged by images with degraded quality. Addressing this limitation, we propose the Robust Segment Anything Model (RobustSAM), which enhances SAM's performance on low-quality images while preserving its promptability and zero-shot generalization. Our method leverages the pre-trained SAM model with only marginal parameter increments and computational requirements. The additional parameters of RobustSAM can be optimized within 30 hours on eight GPUs, demonstrating its feasibility and practicality for typical research laboratories. We also introduce the Robust-Seg dataset, a collection of 688K image-mask pairs with different degradations designed to train and evaluate our model optimally. Extensive experiments across various segmentation tasks and datasets confirm RobustSAM's superior performance, especially under zero-shot conditions, underscoring its potential for extensive real-world application. Additionally, our method has been shown to effectively improve the performance of SAM-based downstream tasks such as single image dehazing and deblurring.

原文英語
頁(從 - 到)4081-4091
頁數11
期刊Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
DOIs
出版狀態已出版 - 2024
事件2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2024 - Seattle, 美國
持續時間: 16 06 202422 06 2024

文獻附註

Publisher Copyright:
© 2024 IEEE.

指紋

深入研究「RobustSAM: Segment Anything Robustly on Degraded Images」主題。共同形成了獨特的指紋。

引用此