Abstract
In applications involving multiple conversational agents, each of these agents has its own view of a visual scene, and thus all the agents must establish common visual landmarks in order to coordinate their space understanding and to coherently share generated spatial descriptions of this scene. Whereas natural language processing approaches contribute to define the common ground through dialogues between these agents, we propose in this paper a computer-vision system to determine the object of reference for both agents efficiently and automatically. Our approach consists in processing each agent's view by computing the related, visual interest points, and then by matching them in order to extract the salient and meaningful landmark. Our approach has been successfully tested on real-world data, and its performance and design allow its use for embedded robotic system communication.
Original language | English |
---|---|
Title of host publication | Proceedings of the 8th International Conference on Agents and Artificial Intelligence |
Subtitle of host publication | ICAART |
Publisher | SciTePress |
Pages | 566-569 |
Number of pages | 4 |
Volume | 2 |
ISBN (Print) | 9789897581724 |
DOIs | |
Publication status | Published - 26 Feb 2016 |
Externally published | Yes |
Event | International Conference on Agents and Artificial Intelligence - Rome, Italy Duration: 24 Feb 2016 → 26 Feb 2016 Conference number: 8 http://www.icaart.org/?y=2016 |
Conference
Conference | International Conference on Agents and Artificial Intelligence |
---|---|
Abbreviated title | ICAART 2016 |
Country/Territory | Italy |
City | Rome |
Period | 24/02/16 → 26/02/16 |
Internet address |
Keywords
- Qualitative Spatial Reasoning
- Object Detention
- Local Feature Descriptors
- Feature Extraction
- Visual Scene Understanding
- Automated Image Annotation
- Robotics
- Autonomic Agents