🤖 AI Summary
In densely populated poultry farming scenarios, accurate counting of silky-feathered black-boned chickens is severely hindered by heavy occlusion and the inherent limitations of single-view imaging. To address this, we propose a tri-planar multi-view density estimation framework: via geometric camera calibration and projective mapping, features from multiple viewpoints are jointly aligned onto three semantically meaningful horizontal planes—ground, mid-level, and top. A spatial feature alignment module ensures cross-view consistency, while fused tri-planar features generate a high-fidelity, scene-level density map. We further introduce MV-SilkyChicken, the first real-world, multi-view dataset of silky chickens captured in operational farming environments. Extensive experiments demonstrate that our method achieves 95.1% counting accuracy in practical deployments—substantially outperforming both single-view baselines and conventional feature fusion approaches. This work delivers a robust, end-to-end, and deployable solution for intelligent livestock management.
📝 Abstract
Accurate animal counting is essential for smart farming but remains difficult in crowded scenes due to occlusions and limited camera views. To address this, we propose a tri-plane-based multi-view chicken counting model (TP-MVCC), which leverages geometric projection and tri-plane fusion to integrate features from multiple cameras onto a unified ground plane. The framework extracts single-view features, aligns them via spatial transformation, and decodes a scene-level density map for precise chicken counting. In addition, we construct the first multi-view dataset of silkie chickens under real farming conditions. Experiments show that TP-MVCC significantly outperforms single-view and conventional fusion comparisons, achieving 95.1% accuracy and strong robustness in dense, occluded scenarios, demonstrating its practical potential for intelligent agriculture.