Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Building a generalized affordance grounding model to identify actionable
regions on objects is vital for real-world applications. Existing methods to
train the model can be divided into weakly and fully supervised ways. However,
the former method requires a complex training framework design and can not
infer new actions without an auxiliary prior. While the latter often struggle
with limited annotated data and components trained from scratch despite being
simpler. This study focuses on fully supervised affordance grounding and
overcomes its limitations by proposing AffordanceSAM, which extends SAM's
generalization capacity in segmentation to affordance grounding. Specifically,
we design an affordance-adaption module and curate a coarse-to-fine annotated
dataset called C2F-Aff to thoroughly transfer SAM's robust performance to
affordance in a three-stage training manner. Experimental results confirm that
AffordanceSAM achieves state-of-the-art (SOTA) performance on the AGD20K
benchmark and exhibits strong generalized capacity.