This project showcases a real-time vision-based control system designed to catch a falling pen using a robotic arm.
The goal was to demonstrate dynamic motion control and low-latency visual feedback for intercepting a rapidly moving target — a challenging benchmark in robotic control due to the system’s tight timing constraints.
System Architecture
The entire workflow integrates high-speed computer vision with a predictive control loop.
Object Detection & Tracking
A high-frame-rate camera continuously captures the workspace.
OpenCV-based motion tracking isolates the pen in flight by analyzing frame differences and centroid motion.
The pen’s 2D trajectory is projected to a 3D coordinate using camera calibration data.
Trajectory Prediction
A Kalman Filter estimates the pen’s velocity and acceleration in real time.
The predicted intersection point (catch location) is computed within tens of milliseconds of release.
This prediction compensates for communication and actuation latency in the robot’s control loop.
Robotic Control
The robot arm (Franka Emika Panda or similar) uses inverse kinematics (IK) to compute the optimal joint configuration for interception.
A PD control law refines motion during final approach to ensure smooth and stable capture.
Timing synchronization between perception and control is maintained using ROS 2 publisher–subscriber nodes running at 100 Hz.