A4-Agent: An Agentic Framework for Zero-Shot Affordance Reasoning

1HKUST(GZ), 2HKUST, 3SJTU, 4Knowin
*Equal contribution     †Corresponding author
A4-Agent Teaser

Overview of A4-Agent. An affordance-centric vision-language agent that predicts actionable regions based on complex task instructions. By integrating image generation, object detection, segmentation, and a vision-language model, our framework imagines plausible interactions to accurately localize the action-specific part. A4-Agent is completely zero-shot and achieves state-of-the-art performance on all benchmarks.

Abstract

Affordance prediction, which identifies interaction regions on objects based on language instructions, is critical for embodied AI. Prevailing end-to-end models couple high-level reasoning and low-level grounding into a single monolithic pipeline and rely on training over annotated datasets, which leads to poor generalization on novel objects and unseen environments.

In this paper, we move beyond this paradigm by proposing A4-Agent, a training-free agentic framework that decouples affordance prediction into a three-stage pipeline. Our framework coordinates specialized foundation models at test time: (1) a Dreamer that employs generative models to visualize how an interaction would look; (2) a Thinker that utilizes large vision-language models to decide what object part to interact with; and (3) a Spotter that orchestrates vision foundation models to precisely locate where the interaction area is.

By leveraging the complementary strengths of pre-trained models without any task-specific fine-tuning, our zero-shot framework significantly outperforms state-of-the-art supervised methods across multiple benchmarks and demonstrates robust generalization to real-world settings.

Motivation

A4-Agent Motivation

Affordance prediction fundamentally requires two complementary capabilities: high-level reasoning (interpreting instructions and identifying relevant parts) and low-level grounding (precisely localizing these parts).

Prevailing end-to-end models tightly couple these processes into a single monolithic pipeline. This design introduces several issues, including a trade-off between reasoning and grounding, limited generalization to novel objects, and reduced flexibility. We question whether entangling these capabilities is the right path. Instead, we propose A4-Agent, which decouples reasoning and grounding, allowing for the coordination of specialized foundation models without task-specific training.

Method

A4-Agent Pipeline

We decompose the task into a three-stage pipeline, with each stage managed by a specialized expert leveraging powerful foundation models:
1) Dreamer: Drawing inspiration from human cognitive processes, the Dreamer initiates an imagination phase. It employs generative models to synthesize visual scenarios depicting how an interaction would look.
2) Thinker: The Thinker utilizes leading Vision-Language Models (VLMs) to interpret task instructions. Integrating visual observations with the imagined scenarios, it generates structured textual descriptions that specify what to interact with.
3) Spotter: The Spotter orchestrates robust vision foundation models to execute precise spatial localization, pinpointing exactly where the interaction area is within the visual input.

Experiments

Quantitative Results

We evaluate A4-Agent on three benchmarks: ReasonAff, RAGNet, and UMD. Crucially, our framework is completely zero-shot—it has never been trained or fine-tuned on any of these datasets.

Results on ReasonAff Dataset

Model gIoU↑ cIoU↑ P50-95 P50
VLPart4.213.880.851.31
OVSeg16.5210.594.129.89
SAN10.2113.453.177.18
LISA-7B38.1740.5819.6933.62
SAM4MLLM45.5133.6422.7943.48
AffordanceLLM48.4938.6120.1942.11
InternVL3-8B31.7924.6821.9335.41
Qwen2.5VL-7B25.1820.5415.8226.00
AffordanceVLM30.5025.5418.3130.29
Seg-Zero59.2648.0345.8761.33
Vision Reasoner63.0452.7047.2367.33
Affordance-R167.4162.7255.2274.50
A4-Agent (Ours)70.5264.6255.2275.24

Results on RAGNet Dataset

Model Zero-shot 3DOI HANDAL-easy HANDAL-hard
gIoU↑cIoU↑ gIoU↑cIoU↑ gIoU↑cIoU↑
G-DINO4.13.93.63.03.43.1
LISA12.38.115.511.912.38.1
GLaMM4.42.94.73.55.03.5
Vision-Reasoner39.630.329.619.827.716.7
Affordance-R139.033.443.138.740.737.9
AffordanceVLM38.139.458.358.158.257.8
A4-Agent (Ours)63.958.361.161.761.059.6

Results on UMD Dataset

Model gIoU↑ cIoU↑ P50 P50-95
LISA-7B41.9041.2339.6519.33
SAM4MLLM12.408.414.120.05
AffordanceLLM43.1138.9741.5622.36
Qwen2.5VL-7B33.2129.8325.1710.45
InternVL3-7B30.4628.7318.679.94
AffordanceVLM25.4117.969.3725.10
Seg-Zero44.2639.3039.9316.53
Vision Reasoner44.0039.7139.0416.10
Affordance-R149.8542.2453.3534.08
A4-Agent (Ours)65.3859.8177.3143.78

Qualitative Results on ReasonAff

Qualitative Results on ReasonAff

A4-Agent demonstrates superior reasoning ability on the ReasonAff dataset, which requires deep understanding of implicit contextual instructions. Our zero-shot framework effectively interprets complex queries and accurately localizes the actionable regions, outperforming supervised baselines.


Qualitative Results on RAGNet

Qualitative Results on RAGNet

Qualitative comparison on the RAGNet dataset. Our zero-shot method effectively reasons over task instructions to identify correct regions and precisely localize them with masks, closely matching ground truth. This outperforms baseline methods including AffordanceVLM trained on this dataset.


Open-World Generalization

Open-World Generalization

We further evaluate A4-Agent on open-world images to assess its robustness. The results show that A4-Agent generalizes well to:

  • Novel Objects: Identifying actionable regions on objects not seen in standard datasets.
  • Complex Scenes: Selecting the correct tool part in cluttered environments.
  • Deep Reasoning: Logically deducing appropriate tools for a given task (e.g., using a rock as a substitute for a hammer).
This confirms the potential of our training-free, agentic approach for real-world applications.

BibTeX

@article{zhang2025a4agent,
  title={A4-Agent: An Agentic Framework for Zero-Shot Affordance Reasoning}, 
  author={Zhang, Zixin and Chen, Kanghao and Wang, Hanqing and Zhang, Hongfei and Chen, Harold Haodong and Liao, Chenfei and Guo, Litao and Chen, Ying-Cong},
  journal={arXiv preprint arXiv:2512.14442},
  year={2025}
}