Abstract:
A challenging task in embodied artificial intelligence is enabling the robot to carry out a navigational task following natural language instruction. In the task, the nav...Show MoreMetadata
Abstract:
A challenging task in embodied artificial intelligence is enabling the robot to carry out a navigational task following natural language instruction. In the task, the navigator needs to understand objects, directions, as well as room types, which serve as landmarks for navigation. Although it is easy to encode objects and directions with an external encoder like an object detector, current navigators struggle to encode room type information properly due to the low accuracy offered by existing classifiers. This inadequacy poses confusion that navigators find difficult to overcome. Even humans may sometimes fail to determine the exact type of a room since multiple room types may exist in one panorama. To mitigate this problem, we propose to encode room type information in a prompt manner. Specifically, we first establish multi-modal, learnable prompt pools containing knowledge of room types. By querying the prompt pools, the navigator can obtain room-type prompts of the current view, and incorporate them into the navigator using a prompt-based learning method. Experimental results on the REVERIE, R2R and SOON datasets demonstrate the effectiveness of our approach.
Published in: IEEE Transactions on Circuits and Systems for Video Technology ( Volume: 34, Issue: 10, October 2024)
Funding Agency:
Citations are not available for this document.
Cites in Papers - |
Cites in Papers - IEEE (1)
Select All
1.
Heqian Qiu, Lanxiao Wang, Taijin Zhao, Fanman Meng, Qingbo Wu, Hongliang Li, "MCCE-REC: MLLM-Driven Cross-Modal Contrastive Entropy Model for Zero-Shot Referring Expression Comprehension", IEEE Transactions on Circuits and Systems for Video Technology, vol.35, no.1, pp.754-768, 2025.