GRIM: Task-Oriented Grasping with Conditioning on Generative Examples
DOI:
https://doi.org/10.1609/aaai.v40i22.38873Abstract
Task-Oriented Grasping (TOG) presents a significant challenge, requiring a nuanced understanding of task semantics, object affordances, and the functional constraints dictating how an object should be grasped for a specific task. To address these challenges, we introduce GRIM (Grasp Re-alignment via Iterative Matching), a novel training-free framework for task-oriented grasping. Initially, a coarse alignment strategy is developed using a combination of geometric cues and principal component analysis (PCA)-reduced DINO features for similarity scoring. Subsequently, the full grasp pose associated with the retrieved memory instance is transferred to the aligned scene object and further refined against a set of task-agnostic, geometrically stable grasps generated for the scene object, prioritizing task compatibility. In contrast to existing learning-based methods, GRIM demonstrates strong generalization capabilities, achieving robust performance with only a small number of conditioning examples.Downloads
Published
2026-03-14
How to Cite
Shailesh, S., Raj, A., Kumar, N., Shukla, P., Melnik, A., Beetz, M., & Nandi, G. C. (2026). GRIM: Task-Oriented Grasping with Conditioning on Generative Examples. Proceedings of the AAAI Conference on Artificial Intelligence, 40(22), 18118–18125. https://doi.org/10.1609/aaai.v40i22.38873
Issue
Section
AAAI Technical Track on Intelligent Robotics