As a guest user you are not logged in or recognized by your IP address. You have
access to the Front Matter, Abstracts, Author Index, Subject Index and the full
text of Open Access publications.
Despite recent progress in integrating Large Language Models (LLMs) to enhance robotic capabilities, key challenges remain—particularly in interpreting perceptions, interacting with humans, and executing tasks in real-world environments. These challenges are largely due to LLM hallucinations and lack of formal semantic knowledge that hinder contextual understanding. To address this issue, we propose a semantic framework that combines an LLM and a commonsense knowledge graph to align the robot’s perceptions with a suitable task execution plan. The framework integrates pre-trained deep learning models for object detection, face identification, speech recognition, an LLM, and a local knowledge graph based on Resource Description Framework (RDF). The knowledge graph including scene and object description, as well spatial knowledge is used to select and execute the suitable task plan for the given context. Experimental results demonstrate the effectiveness of framework’s components in improving the robot’s interaction and and actuation in its environment.
This website uses cookies
We use cookies to provide you with the best possible experience. They also allow us to analyze user behavior in order to constantly improve the website for you. Info about the privacy policy of IOS Press.
This website uses cookies
We use cookies to provide you with the best possible experience. They also allow us to analyze user behavior in order to constantly improve the website for you. Info about the privacy policy of IOS Press.