This project explores the integration of Large Language Models (LLMs), specifically ChatGPT, into robotic decision-making frameworks to enhance real-time human-robot collaboration. By leveraging a hybrid decision-making model, this system balances reactive control with deliberative planning, ensuring adaptive, efficient, and safe robotic behavior.
β
LLM-Based Task Planning β Uses ChatGPT to process high-level natural language commands and convert them into actionable robot instructions.
β
MoveIt Integration β Implements a hybrid planner in ROS2 MoveIt for dynamic path planning and obstacle avoidance.
β
Human Detection and Interaction β Utilizes a RGB-D camera and YOLO_ROS for real-time human pose estimation and collision avoidance.\
The robotic system is structured into several key components:
- UR5e Robotic Arm β Primary manipulator executing planned trajectories.
- ChatGPT Task Planner β Processes human commands and generates structured task plans.
- MoveIt Hybrid Planning Framework β Provides real-time reactive path planning and obstacle avoidance.
- PrimeSense 3D Depth Camera β Enables human pose estimation and dynamic collision marker generation.
- YOLO ROS Wrapper β Detects humans and objects to generate collision boundaries for safer interactions.
- MoveIt hybrid planner successfully adapted to dynamic environments, ensuring real-time path corrections.
- Human pose detection provided accurate real-time collision avoidance, though response times needed further optimization.
- The global planner effectively executed pre-defined tasks with high accuracy.
- Real-time hybrid planning was validated in simulations but awaits further testing in physical environments.
π Improve real-time responsiveness with optimized service calls.
π Enhance human detection accuracy using LiDAR or multi-camera setups.
π Scale the system for complex environments, such as factory floors or clinical settings.
π Explore alternative LLMs for improved command interpretation.
π€ Nicholas Bell β Lead Developer & Researcher
π© Contact:Β https://www.linkedin.com/in/nickojbell/
This project is licensed under the MIT License. See the LICENSE file for details.
π For a detailed breakdown, refer to the full Thesis Document.
This project utilizes external repositories to extend its functionality. A big thank you to the contributors of these projects for their efforts in making robotics more accessible and powerful!
-
OpenNI2 Submodule
π Path: src/OpenNI2
π Repository: structureio/OpenNI2 -
OpenNI2 Camera Submodule
π Path: src/openni2_camera
π Repository: ros-drivers/openni2_camera -
YOLO_ROS Perception Submodule
π Repository: mgonzs13/yolo_ros