← Back to Portfolio

Real-Time Vision-Guided Robotic Arm

YOLOE segmentation + WebSocket control pipeline

YOLOE ONNX OpenCV WebSocket Inverse Kinematics Raspberry Pi Edge Inference Custom Fine-Tuning

A real-time object detection and tracking pipeline that streams normalized coordinates to a robotic arm via WebSocket. YOLOE-11 segmentation models are fine-tuned on custom classes using visual or text prompts, then exported to ONNX for low-latency edge inference on Raspberry Pi or standard USB cameras.

The robot receives continuous JSON payloads — pixel coordinates, normalized (0–1) position, and confidence — which feed directly into an IK solver for end-effector targeting. Demonstrated on coin detection (Mexican pesos, dining tokens, USD) and building/logo recognition.

Demo

Detection Examples

Technical Implementation

Pipeline