🤖 AI Summary
Spherical robots face significant challenges in achieving realistic flocking behavior due to severe visual constraints—limited to perceiving only neighboring agents’ positions, optical sizes, and optical flow on a panoramic retina.
Method: This paper proposes a purely vision-driven, distributed flocking control framework integrating robot-in-the-loop panoramic visual sensing, dense optical flow estimation, virtual anchor constraints, and local interaction modeling.
Contribution/Results: For the first time, the approach stably reproduces canonical flocking phases—including aggregation and milling—in a physical swarm of 10 spherical robots. It bridges the behavioral gap between simulation and real-world deployment: empirical flocking dynamics closely match theoretical predictions and classical models (e.g., Reynolds-style rules). The system demonstrates strong scalability and robustness under resource constraints, offering a novel paradigm for vision-based coordination in resource-limited multi-agent systems.
📝 Abstract
The implementation of collective motion, traditionally, disregard the limited sensing capabilities of an individual, to instead assuming an omniscient perception of the environment. This study implements a visual flocking model in a ``robot-in-the-loop'' approach to reproduce these behaviors with a flock composed of 10 independent spherical robots. The model achieves robotic collective motion by only using panoramic visual information of each robot, such as retinal position, optical size and optic flow of the neighboring robots. We introduce a virtual anchor to confine the collective robotic movements so to avoid wall interactions. For the first time, a simple visual robot-in-the-loop approach succeed in reproducing several collective motion phases, in particular, swarming, and milling. Another milestone achieved with by this model is bridging the gap between simulation and physical experiments by demonstrating nearly identical behaviors in both environments with the same visual model. To conclude, we show that our minimal visual collective motion model is sufficient to recreate most collective behaviors on a robot-in-the-loop system that is scalable, behaves as numerical simulations predict and is easily comparable to traditional models.