Abstract
The weeds compete with the crops for the nutrients. However, it is necessary to control weeds in order to prevent them from affecting the productivity of the land. In the context of the robotization of this process, the Deep Learning can help to localize and identify the weeds, thanks to the object detector. The accuracy of an object detector depends on its backbones which are networks their function is feature extraction. In this research, we have shown the performance of the most popular CNNs, which are used as backbones, for the identification and distinction of crop weeds, as well as proposed a model that combines the architecture of the two models MobileNet and ResNet. The proposed model had the properties of MobileNet in terms speed and precision, our modification has resulted in a significant improvement. The obtained results, showed us that the MobileNet models are the most suitable for as backbone to the future object detector that we want to design it to detect and ultra-localize the weeds.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Mohammed, H., Tannouche, A., Ounejjar, Y.: Weed detection in pea cultivation with the faster RCNN ResNet 50 convolutional neural network. Revue d’Intelligence Artificielle 36(1), 13–18 (2022). https://doi.org/10.18280/ria.360102
Zou, Z., Shi, Z., Guo, Y., Ye, J: Object Detection in 20 Years: A Survey (2019)
Ashqar, B.A.M., Abu-Nasser, B.S., Abu-Naser, S.S.: Plant Seedlings Classification Using Deep Learning (2019)
Li, B., Lima, D.: Facial expression recognition via ResNet-50. Int. J. Cogn. Comput. Eng. 2, 57–64 (2021). https://doi.org/10.1016/j.ijcce.2021.02.002
Venkateswarlu, I.B., Kakarla, J., Prakash, S.: Face mask detection using MobileNet and global pooling block. In: 4th IEEE Conference on Information and Communication Technology, CICT 2020. Institute of Electrical and Electronics Engineers Inc. (2020)
Beckmann, P., Kegler, M., Cernak, M.: Word-level Embeddings for Cross-Task Transfer Learning in Speech Processing (2019). https://doi.org/10.23919/EUSIPCO54536.2021.9616254
Michele, A., Colin, V., Santika, D.D.: Mobilenet convolutional neural networks and support vector machines for palmprint recognition. In: Procedia Computer Science. Elsevier B.V., pp. 110–117 (2019)
Wang, W., Li, Y., Zou, T., et al.: A novel image classification approach via dense-mobilenet models. Mob. Inf. Syst. 2020,(2020). https://doi.org/10.1155/2020/7602384
Zheng, Y.Y., Kong, J.L., Jin, X.B., et al.: Cropdeep: the crop vision dataset for deep-learning-based classification and detection in precision agriculture. Sensors (Switzerland) 19,(2019). https://doi.org/10.3390/s19051058
Chiu, Y.-C., Tsai, C.-Y., Ruan, D., et al.: Mobilenet-SSDv2: An Improved Object Detection Model for Embedded Systems (2020)
Howard, A.G., Zhu, M., Chen, B., et al.: MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications (2017)
Sandler, M., Howard, A., Zhu, M., et al.: MobileNetV2: Inverted Residuals and Linear Bottlenecks (2018)
He, K., Zhang, X., Ren, S., Sun, J.: Deep Residual Learning for Image Recognition (2015)
Grandini, M., Bagli, E., Visani, G.: Metrics for Multi-Class Classification: an Overview (2020)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Habib, M., Sekhra, S., Tannouche, A., Ounejjar, Y. (2023). The Identification of Weeds and Crops Using the Popular Convolutional Neural Networks. In: Motahhir, S., Bossoufi, B. (eds) Digital Technologies and Applications. ICDTA 2023. Lecture Notes in Networks and Systems, vol 668. Springer, Cham. https://doi.org/10.1007/978-3-031-29857-8_49
Download citation
DOI: https://doi.org/10.1007/978-3-031-29857-8_49
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-29856-1
Online ISBN: 978-3-031-29857-8
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)