Recognition of tea buds based on an improved YOLOv7 model

Mengxue Song, Ce Liu, Liqing Chen, Lichao Liu, Jingming Ning, Chuanyang Yu

Abstract


The traditional recognition algorithm is prone to miss detection targets in the complex tea garden environment, and it is difficult to satisfy the requirement for tea bud recognition accuracy and efficiency. In this study, the YOLOv7 model was developed to improve tea bud recognition accuracy for some extreme tea garden scenarios. In the improved model, a lightweight MobileNetV3 network is adopted to replace the original backbone network, which reduces the size of the model and improves detection efficiency. The convolutional block attention module is introduced to enhance the attention to the features of small and occluded tea buds, suppressing the interference of the complex tea garden environment on tea bud recognition and strengthening the feature extraction capability of the recognition model. Moreover, to further improve recognition accuracy for dense and occlusive scenarios, the soft non-maximum suppression strategy is integrated into the recognition model. Experimental results show that the improved YOLOv7 model has the precision, recall, and mean average precision (mAP) values of 88.3%, 87.4%, and 88.5%, respectively. Compared with the Faster R-CNN, SSD, and original YOLOv7 algorithms, the mAP of the improved YOLOv7 model is increased by 7.4, 7.9, and 3.9 percentage points, respectively, and its recognition speed is also promoted by 94.9%, 46.2%, and 16.9%. The proposed model can rapidly and accurately identify the tea buds in multiple complex tea garden scenarios - such as dense distribution, being close to the background color, and mutual occlusion - with high generalization and robustness, which can provide theoretical and technical support for the recognition of tea-picking robots.
Keywords: tea bud recognition, YOLOv7, lightweight MobileNetV3 network, CBAM
DOI: 10.25165/j.ijabe.20241706.9095

Citation: Song M X, Liu C, Chen L Q, Liu L C, Ning J M, Yu C Y. Recognition of tea buds based on an improved YOLOv7 model. Int J Agric & Biol Eng, 2024; 17(6): 238–244.

Keywords


tea bud recognition, YOLOv7, lightweight MobileNetV3 network, CBAM

Full Text:

PDF

References


Chen T C, Li H X, Chen J Z, Zeng Z H, Han C Y, Wu W B. Detection network for multi-size and multi-target tea bud leaves in the field of view via improved YOLOv7. Computers and Electronics in Agriculture, 2024; 218: 108700.

Lu J Q, Luo H X, Yu C R, Xiao L, Huang J W, Wu H W, et al. Tea bud DG: A lightweight tea bud detection model based on dynamic detection head and adaptive loss function. Computers and Electronics in Agriculture, 2024; 227: 109522.

Xie D, Chen L, Liu L, Chen L, Wang H. Actuators and sensors for application in agricultural robots: A review. Machines, 2022; 10(10): 913.

Zhang L, Zou L, Wu C Y, Jia J M, Chen J N. Method of famous tea sprout identification and segmentation based on improved watershed algorithm. Computers and Electronics in Agriculture, 2021; 184: 106108.

Zhang L, Zhang H D, Chen Y D, Dai S H, Li X M, Imou K, Liu Z H, et al. Real-time monitoring of optimum timing for harvesting fresh tea leaves based on machine vision. Int J Agric & Biol Eng, 2019; 12(1): 6–9.

Shao P D, Wu M H, Wang X W, Zhou J, Liu S, Wang Y S. Research on the tea bud recognition based on improved k-means algorithm. MATEC Web of Conferences, 2018; 232: 03050.

Chen C L, Lu J Z, Zhou M C, Yi J, Liao M, Gao Z M. A YOLOv3-based computer vision system for identification of tea buds and the picking point. Computers and Electronics in Agriculture, 2022; 198: 107116.

Hu H M, Kaizu Y, Zhang H D, Xu Y W, Imou K, Li M, et al. Recognition and localization of strawberries from 3D binocular cameras for a strawberry picking robot using coupled YOLO/Mask R-CNN. Int J Agric & Biol Eng, 2022; 15(6): 175–179,

Liu B, Wei S S, Zhang F, Guo N W, Fan H Y, Yao W. Tomato leaf disease recognition based on multi-task distillation learning. Frontiers in Plant Science, 2024; 14: 1330527.

Li F M, Zheng Y H, Liu S, Sun F B, Bai H R. A multi-objective apple leaf disease detection algorithm based on improved TPH-YOLOV5. Applied Fruit Science, 2024; 66(2): 339–415.

Li Y T, He L Y, Jia J M, Chen J N, Lyu J, Wu C A Y. High-efficiency tea shoot detection method via a compressed deep learning model. Int J Agric & Biol Eng, 2022; 15(3): 159–166.

Wang Y W, Xiao M H, Wang S, Jiang Q, Wang X C, Zhang Y N. Detection of famous tea buds based on improved YOLOv7 network. Agriculture-Basel, 2023; 13(6): 1190.

Li J, Li J H, Zhao X, Su X H, Wu W B. Lightweight detection networks for tea bud on complex agricultural environment via improved YOLO v4. Computers and Electronics in Agriculture, 2023; 211: 107955.

Gui J S, Wu D W, Xu H R, Chen J N, Tong J H. Tea bud detection based on multi-scale convolutional block attention module. Journal of Food Process Engineering, 2024; 47(2): e14556.

Aqel D, Al-Zubi S, Mughaid A, Jararweh Y. Extreme learning machine for plant diseases classification: a sustainable approach for smart agriculture. Cluster Computing-The Journal of Networds Software Tools and Applications, 2021; 25(3): 2303–2303.

Zhang X W, Xuan C Z, Hou Z F. Recognition model for coated red clover seeds using YOLOv5s optimized with an attention module. Int J Agric & Biol Eng, 2023; 16(6): 207–214.

Gu Z A, He D Q, Huang J D, Chen J Q, Wu X H, Huang B C, et al. Simultaneous detection of fruits and fruiting stems in mango using improved YOLOv8 model deployed by edge device. Computers and Electronics in Agriculture, 2024; 227: 109512.

Zhao L L, Zhu M L. MS-YOLOv7: YOLOv7 based on multi-scale for object detection on UAV aerial photography. Drones, 2023; 7(3): 188.

Du W S, Jia Z H, Sui S S, Liu P. Table grape inflorescence detection and clamping point localisation based on channel pruned YOLOV7-TP. Biosystems Engineering, 2023; 235: 100–115.

Ma H R, Li Y W, Zhang X Y, Li Y Y, Li Z Q, Zhang R Q, et al. Target Detection for coloring and ripening potted dwarf apple fruits based on improved YOLOv7-RSES. Applied Sciences-Basel, 2024; 14(11): 4523.

Zhang Y, Sun Y P, Wang Z, Jiang Y. YOLOv7-RAR for urban vehicle detection. Sensors, 2023; 23(4): 1801.

Howard A, Sandler M, Chen B, Wang W, Chen L C, Tan M, et al. Searching for MobileNetV3. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019; pp.1314–1324.

Howard A G, Zhu M L, Chen B, Kalenichenko D, Wang W J, Weyand T, et al. MobileNets: efficient convolutional neural networks for mobile vision applications. 2017; arXiv: 1704.04861. https://arxiv.org/pdf/1704.04861

Sandler M, Howard A, Zhu M, Zhmoginov A, Chen L C. MobileNetV2: inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018; pp.4510-4520. https://arxiv.org/abs/1704.04861

Woo S, Park J, Lee J Y, Kweon I S. CBAM: convolutional block attention module. Computer Vision - ECCV. Cham: Springer International Publishing, 2018; 3-19.

Khan R U, Khattak H, Wong W S, AlSalman H, Mosleh M A A, Rahman S M M. Intelligent malaysian sign language translation system using convolutional-based attention module with residual network. Computational Intelligence and Neuroscience, 2022; 2021: 9023010.

Bodla N, Singh B, Chellapp.R, Davis L S. Soft-NMS—improving object detection with one line of code. Proceedings of the IEEE International Conference on Computer Vision, 2017; pp.5561–5569.




Copyright (c) 2024 International Journal of Agricultural and Biological Engineering

Creative Commons License
This work is licensed under a Creative Commons Attribution 4.0 International License.

2023-2026 Copyright IJABE Editing and Publishing Office