WalnutData: A UAV Remote Sensing Dataset of Green Walnuts and Model Evaluation

๐Ÿ“… 2025-02-27
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
To address the scarcity of high-quality remote sensing datasets for green walnuts in agricultural computer vision, this paper introduces WalnutDataโ€”the first dedicated UAV-based remote sensing dataset for green walnuts, comprising 30,240 images and 706,208 annotated instances. We propose a high-granularity annotation paradigm that systematically models the combined effects of illumination (front-lit vs. back-lit) and occlusion (occluded vs. non-occluded), yielding four distinct subsets. Furthermore, we establish the first authoritative benchmark and unified evaluation protocol for this task, conducting comprehensive experiments using state-of-the-art detectors including YOLOv5, YOLOv8, and Faster R-CNN. Both the full dataset and benchmark results are publicly released, significantly improving detection accuracy and robustness under complex field conditions. The resource has garnered substantial attention on GitHub.

Technology Category

Application Category

๐Ÿ“ Abstract
The UAV technology is gradually maturing and can provide extremely powerful support for smart agriculture and precise monitoring. Currently, there is no dataset related to green walnuts in the field of agricultural computer vision. Thus, in order to promote the algorithm design in the field of agricultural computer vision, we used UAV to collect remote-sensing data from 8 walnut sample plots. Considering that green walnuts are subject to various lighting conditions and occlusion, we constructed a large-scale dataset with a higher-granularity of target features - WalnutData. This dataset contains a total of 30,240 images and 706,208 instances, and there are 4 target categories: being illuminated by frontal light and unoccluded (A1), being backlit and unoccluded (A2), being illuminated by frontal light and occluded (B1), and being backlit and occluded (B2). Subsequently, we evaluated many mainstream algorithms on WalnutData and used these evaluation results as the baseline standard. The dataset and all evaluation results can be obtained at https://github.com/1wuming/WalnutData.
Problem

Research questions and friction points this paper is trying to address.

Lack of green walnut dataset in agricultural computer vision.
UAV-collected dataset addresses varying lighting and occlusion.
Evaluates mainstream algorithms on new WalnutData baseline.
Innovation

Methods, ideas, or system contributions that make the work stand out.

UAV collected high-granularity walnut data
Constructed WalnutData for algorithm evaluation
Evaluated mainstream algorithms on green walnuts
Mingjie Wu
Mingjie Wu
Yunnan Normal University
Small Object DetectionSuper-Resolution ReconstructionIntelligent AgricultureUAV Image Process
C
Chenggui Yang
School of Information, Yunnan Normal University; Engineering Research Center of Computer Vision and Intelligent Control Technology, Department of Education of Yunnan Province
H
Huihua Wang
School of Physics and Electronic Information, Yunnan Normal University; Engineering Research Center of Computer Vision and Intelligent Control Technology, Department of Education of Yunnan Province
C
Chen Xue
School of Information, Yunnan Normal University; Engineering Research Center of Computer Vision and Intelligent Control Technology, Department of Education of Yunnan Province
Y
Yibo Wang
School of Information, Yunnan Normal University; Engineering Research Center of Computer Vision and Intelligent Control Technology, Department of Education of Yunnan Province
H
Haoyu Wang
School of Information, Yunnan Normal University; Engineering Research Center of Computer Vision and Intelligent Control Technology, Department of Education of Yunnan Province
Y
Yansong Wang
School of Information, Yunnan Normal University; Engineering Research Center of Computer Vision and Intelligent Control Technology, Department of Education of Yunnan Province
Can Peng
Can Peng
University of Oxford
Computer VisionIncremental Learning
Y
Yuqi Han
School of Information, Yunnan Normal University; Engineering Research Center of Computer Vision and Intelligent Control Technology, Department of Education of Yunnan Province
R
Ruoyu Li
School of Physics and Electronic Information, Yunnan Normal University; Engineering Research Center of Computer Vision and Intelligent Control Technology, Department of Education of Yunnan Province
L
Lijun Yun
School of Information, Yunnan Normal University; Engineering Research Center of Computer Vision and Intelligent Control Technology, Department of Education of Yunnan Province
Z
Zaiqing Chen
School of Information, Yunnan Normal University; Engineering Research Center of Computer Vision and Intelligent Control Technology, Department of Education of Yunnan Province
S
Songfan Shi
School of Information, Yunnan Normal University; Engineering Research Center of Computer Vision and Intelligent Control Technology, Department of Education of Yunnan Province
L
Luhao Fang
School of Information, Yunnan Normal University; Engineering Research Center of Computer Vision and Intelligent Control Technology, Department of Education of Yunnan Province
S
Shuyi Wan
School of Information, Yunnan Normal University; Engineering Research Center of Computer Vision and Intelligent Control Technology, Department of Education of Yunnan Province
T
Tingfeng Li
School of Information, Yunnan Normal University; Engineering Research Center of Computer Vision and Intelligent Control Technology, Department of Education of Yunnan Province
S
Shuangyao Liu
School of Information, Yunnan Normal University; Engineering Research Center of Computer Vision and Intelligent Control Technology, Department of Education of Yunnan Province
H
Haotian Feng
School of Information, Yunnan Normal University; Engineering Research Center of Computer Vision and Intelligent Control Technology, Department of Education of Yunnan Province