VVM was my first venture into ROS2 development, combining voice commands, computer vision, and robotic manipulation. This project demonstrates how AI can make robots understand and respond to human commands naturally.
The system works in three main steps:
Say "pick up the red ball" and the robot will find the red ball, calculate its position, and grab it!
Voice Video Manipulator system architecture
Object detection and distance calculation in real-time
The vision system processes camera feeds to identify and locate objects:
Complete voice-controlled pick-and-place operation
This project was my introduction to ROS2 development. It taught me how to:
The experience gained from VVM became the foundation for all my future robotics projects.