LLM-aided Human-Robot Interfacing for A Prompt-to-Control Application - ArticAI2024
This work introduces an innovative solution for seamless human-robot interaction, enabling operators to control a mobile robot using natural language commands. Here, a custom large language model (LLM), is hosted on a server and equipped with contextual knowledge of the environment, that can process free form text commands provided through a user-friendly interface, and thus, eliminating the need for prior knowledge of machine-specific languages. The LLM translates these directives into high-level plans structured around three core questions: Go where?, Find what?, and Do what?. These plans are then transmitted to the robot, which autonomously navigates to the specified location and identifies target objects using an optimized, lightweight artificial intelligence (AI) model designed for real-time performance. Demonstrations validate the system's ability to generate precise, actionable plans from broad commands and execute tasks efficiently, highlighting its potential to enhance human-robot collaboration.
Relevant Resources
Paper: [to be included]
Github repository: https://github.com/ICONgroupCWC/LLM-aided-Prompt-to-Control