Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
162 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A fast accurate fine-grain object detection model based on YOLOv4 deep neural network (2111.00298v1)

Published 30 Oct 2021 in cs.CV and cs.LG

Abstract: Early identification and prevention of various plant diseases in commercial farms and orchards is a key feature of precision agriculture technology. This paper presents a high-performance real-time fine-grain object detection framework that addresses several obstacles in plant disease detection that hinder the performance of traditional methods, such as, dense distribution, irregular morphology, multi-scale object classes, textural similarity, etc. The proposed model is built on an improved version of the You Only Look Once (YOLOv4) algorithm. The modified network architecture maximizes both detection accuracy and speed by including the DenseNet in the back-bone to optimize feature transfer and reuse, two new residual blocks in the backbone and neck enhance feature extraction and reduce computing cost; the Spatial Pyramid Pooling (SPP) enhances receptive field, and a modified Path Aggregation Network (PANet) preserves fine-grain localized information and improve feature fusion. Additionally, the use of the Hard-Swish function as the primary activation improved the model's accuracy due to better nonlinear feature extraction. The proposed model is tested in detecting four different diseases in tomato plants under various challenging environments. The model outperforms the existing state-of-the-art detection models in detection accuracy and speed. At a detection rate of 70.19 FPS, the proposed model obtained a precision value of $90.33 \%$, F1-score of $93.64 \%$, and a mean average precision ($mAP$) value of $96.29 \%$. Current work provides an effective and efficient method for detecting different plant diseases in complex scenarios that can be extended to different fruit and crop detection, generic disease detection, and various automated agricultural detection processes.

Citations (193)

Summary

  • The paper introduces an enhanced YOLOv4 algorithm that achieves a mean average precision of 96.29% at 70.19 FPS for plant disease detection.
  • It leverages a DenseNet backbone along with novel residual blocks and a modified PANet to efficiently extract and fuse fine-grained features.
  • Experimental results demonstrate high precision (90.33%) and an F1-score of 93.64%, underscoring its potential for real-time precision agriculture applications.

An Evaluation of a High-Performance Object Detection Framework for Plant Disease Detection Using an Enhanced YOLOv4 Algorithm

The paper discusses a novel advancement in precision agriculture by introducing a fine-grained object detection framework specifically tailored for plant disease detection. This proposed model is based on an improved version of the YOLOv4 algorithm and addresses common challenges such as dense distribution of disease symptoms, irregular morphology, and multiscale object classes, which often constrain traditional disease detection methods.

The enhanced YOLOv4 model introduces several architectural innovations to balance detection speed and accuracy effectively. A significant improvement involves employing DenseNet in the backbone network to optimize feature transfer and reuse. This change is complemented by the addition of two novel residual blocks in the backbone and neck components, designed to augment feature extraction capabilities while also reducing computing costs. Another vital enhancement is the integration of a Spatial Pyramid Pooling (SPP) layer that amplifies the model's receptive field. Besides, the Path Aggregation Network (PANet) incorporated in the neck component has been modified to better preserve fine-grained localized information, thereby improving feature fusion.

The model's efficacy was tested on a dataset of tomato plant diseases under various challenging environmental conditions. Experimental results demonstrate the model's superior performance in both detection accuracy and processing speed when compared to existing state-of-the-art detection frameworks. At a processing rate of 70.19 frames per second (FPS), the model achieved a precision of 90.33%, an F1-score of 93.64%, and a mean average precision (mAP) of 96.29%. These results underscore the model's robustness and suitability for real-time applications.

One more technical nuance is the adoption of the Hard-Swish activation function, which reportedly improves the model's capability in nonlinear feature extraction. Such functions play a crucial role in deep learning models by influencing the speed and accuracy of convergence during training, thereby enhancing the model's overall performance.

The implications of this research are noteworthy for practical agricultural applications, as it provides a potent tool for early disease detection in plants, which can significantly mitigate potential losses in crop yields. On a theoretical level, this research contributes to improving object detection models by refining the architecture of neural networks to handle complex tasks involving fine-grained details. This approach can potentially be extrapolated to other domains requiring high precision in object detection, such as autonomous driving or medical diagnostics.

Looking ahead, further refinements could be considered in enhancing the model’s feature extraction depth for even more challenging datasets. Additionally, extending this architecture to support a broader range of agricultural tasks and other industries remains a promising avenue for future developments in artificial intelligence. The paper stands as a testament to the advancing frontier of AI applications in agriculture, demonstrating how targeted improvements in model architecture can yield significant practical dividends.