YOLOE segmentation + WebSocket control pipeline
A real-time object detection and tracking pipeline that streams normalized coordinates to a robotic arm via WebSocket. YOLOE-11 segmentation models are fine-tuned on custom classes using visual or text prompts, then exported to ONNX for low-latency edge inference on Raspberry Pi or standard USB cameras.
The robot receives continuous JSON payloads — pixel coordinates, normalized (0–1) position, and confidence — which feed directly into an IK solver for end-effector targeting. Demonstrated on coin detection (Mexican pesos, dining tokens, USD) and building/logo recognition.
results[0].speed['inference']