Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Generalized SAM: Efficient Fine-Tuning of SAM for Variable Input Image Sizes

About

There has been a lot of recent research on improving the efficiency of fine-tuning foundation models. In this paper, we propose a novel efficient fine-tuning method that allows the input image size of Segment Anything Model (SAM) to be variable. SAM is a powerful foundational model for image segmentation trained on huge datasets, but it requires fine-tuning to recognize arbitrary classes. The input image size of SAM is fixed at 1024 x 1024, resulting in substantial computational demands during training. Furthermore, the fixed input image size may result in the loss of image information, e.g. due to fixed aspect ratios. To address this problem, we propose Generalized SAM (GSAM). Different from the previous methods, GSAM is the first to apply random cropping during training with SAM, thereby significantly reducing the computational cost of training. Experiments on datasets of various types and various pixel counts have shown that GSAM can train more efficiently than SAM and other fine-tuning methods for SAM, achieving comparable or higher accuracy.

Sota Kato, Hinako Mitsuoka, Kazuhiro Hotta• 2024

Related benchmarks

TaskDatasetResultRank
Semantic segmentationCityscapes
mIoU76.25
218
Semantic segmentationCamVid
mIoU73.99
70
Semantic segmentationISBI 2012
mIoU80.53
13
Semantic segmentationKvasir-Seg
mIoU88.76
13
Semantic segmentationM-Building
mIoU80.69
9
Semantic segmentationSynapse
mIoU72.78
9
Semantic segmentationTrans10K
mIoU89.19
9
Showing 7 of 7 rows

Other info

Follow for update