Abstract
The productivity and efficiency of a traditional agriculture crop production cycle can be improved by computer vision and machine learning. Innovative Information and Communication Technology (ICT) solutions in sensing, processing, and learning have attracted significant attention in precision agriculture. The power of visualizing the real world allows spatial or spatiotemporal information gathering and its representation. On the other hand, learning allows processing the information for planning, reasoning, and inference. The combination of computer vision and machine learning facilitates the automation of various tasks in the crop cycle. Compared to conventional agricultural practices, tasks are performed automatically and with greater fidelity. Computer vision has improved due to the following: availability of solid-state photo-detectors, increased sensor size, better spatial and spectral resolution, increased frame rate, and shutter speed. The development of remote sensing platforms like Unmanned Aerial Vehicles, terrestrial robots, and satellites allows capturing images at different scales. All the above factors increase the quality of the visual information. At the same time, information processing has become more accessible due to the growth of parallel computing (Graphics Processing Unit), efficient open-source machine learning libraries, large-scale annotated datasets, and reproducible research. Enrichment in information and increased processing power have seen innovations in precision agriculture like crop health and growth monitoring, prevention and control of pest and crop disease, automatic harvesting of crops, automated crop quality testing, and automated farm management. Computer vision and machine learning go hand in hand with precision agriculture, and therefore, they are seamlessly integrated into this chapter. It showcases publicly available large datasets and state-of-the-art algorithms developed while using them. After thorough analysis, it is evident that computer vision combined with intelligent techniques allows processing in more complex environments. It, in turn, will produce generalized and robust automation systems for precision agriculture.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Achanta R, Shaji A, Smith K, Lucchi A, Fua P, Süsstrunk S (2012) SLIC superpixels compared to state-of-the-art superpixel methods. IEEE PAMI 34(11):2274–2282
Aich S, Stavness I (2017) Leaf counting with deep convolutional and deconvolutional networks. In: Proceedings of the IEEE international conference on computer vision workshops, pp 2080–2089
Altaheri H, Alsulaiman M, Muhammad G (2019a) Date fruit classification for robotic harvesting in a natural environment using deep learning. IEEE Access (7):117115–117133
Altaheri H, Alsulaiman M, Muhammad G, Amin SU, Bencherif M, Mekhtiche M (2019) Date fruit dataset for intelligent harvesting. Data brief (26):104514
Arabidopsis Thaliana Root segmentation (2021) https://sites.google.com/sinc.unl.edu.ar/root-segmentation-challenge/home#h.x8m73yix931s. Accessed 10 June 2021
Badrinarayanan V, Kendall A, Cipolla R (2017) Segnet A deep convolutional encoder-decoder architecture for image segmentation. IEEE PAMI 39(12):2481–2495
Barbedo JGA (2013) Digital image processing techniques for detecting, quantifying and classifying plant diseases. SpringerPlus 2(1):1–12
Barbedo JGA (2016) A review on the main challenges in automatic plant disease identification based on visible range images. Biosyst Eng (144):52–60
Bargoti S, Underwood J (2017) Deep fruit detection in orchards. In: Proceedings IEEE international conference on robotics and automation, pp 3626–3633
Bochkovskiy A, Wang CY, Liao HYM (2020) Yolov4: Optimal speed and accuracy of object detection. arXiv preprint arXiv:2004.10934
Chebrolu N, Lottes P, Schaefer A, Winterhalter W, Burgard W, Stachniss C (2017) Agricultural robot dataset for plant classification, localization and mapping on sugar beet fields. Int J Rob Res 36(10):1045–1052
David E, Madec S, Sadeghi-Tehran P, Aasen H, Zhen B, Liu S Guo W (2020) Global Wheat Head Detection (GWHD) dataset: a large and diverse dataset of high-resolution RGB-labelled images to develop and benchmark wheat head detection methods. Plant Phenomics
David E, Ogidi F, Guo W, Baret F, Stavness, I (2021) Global wheat challenge 2020: Analysis of the competition design and winning models. arXiv preprint arXiv:2105.06182
de Oliveira ME, Corrêa CG (2020) Virtual reality and augmented reality applications in agriculture: a literature review. In: Proceedings 22nd symposium on virtual and augmented reality, pp 1–9
Di Cicco M, Potena C, Grisetti G, Pretto A (2017) Automatic model based dataset generation for fast and accurate crop and weeds detection. In: Proceedings IEEE/RSJ international conference on intelligent robots and systems pp 5188–5195
Dobermann A, Blackmore BS, Cook S, Adamchuk VI (2004) Precision farming: challenges and future directions. In New directions for a diverse planet. in: proceeding 4th international crop science congress, pp 1–19
Dong W, Yuanquan C, Daoliang L, Wanbin Z, Weiming T, Taisheng D Shaozhong K (2019) Foresight of disruptive technologies in agricultural engineering. Stra Stu of Chin Acad Eng 20(6):57–63
dos Santos Ferreira A, Freitas DM, da Silva GG, Pistori H, Folhes MT (2017a) Weed detection in soybean crops using ConvNets. Comput Electron Agric (143)314–324
dos Santos Ferreira A, Pistori H, Matte FD, Gonçalves da S, Gercina (2017b) Data for: Weed Detection in Soybean Crops Using ConvNets, Mendeley Data, v2, https://doi.org/10.17632/3fmjm7ncc6.2. Accessed 24 June 2021
FAO (2009) How to Feed the World in 2050. http://www.fao.org/fileadmin/templates/wsfs/docs/expert_paper/How_to_Feed_the_World_in_2050.pdf. Accessed 8 June 2021
Fountas S, Aggelopoulou K, Gemtos TA (2016) Precision Agriculture. In: Supply chain management for sustainable food networks. Wiley, Chichester, UK, pp 41–65
Gaggion N, Ariel F, Daric V, Lambert É, Legendre S, Roulé T, Ferrante E (2020) ChronoRoot: high-throughput phenotyping by deep segmentation networks reveals novel temporal parameters of plant root system architecture. https://doi.org/10.1101/2020.10.27.350553
Gebbers R, Adamchuk VI (2010) Precision agriculture and food security. Science 327(5967):828–831
Hammer GL, van Oosterom E, McLean G, Chapman S C, Broad I, Harland P, Muchow R C (2010) Adapting APSIM to model the physiology and genetics of complex adaptive traits in field crops. J Exp Bot 61(8):2185–2202
Haralick RM, Shapiro LG (1991) Glossary of computer vision terms. Pattern Recognit 24(1):69–93
He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE international conference on pattern recognition, pp 770–778
Huang T (1996) Computer vision: evolution and promise. https://cds.cern.ch/record/400313/files/p21.pdf. Accessed 16 June 2021
ISPA (2019) Definition of precision. https://www.springer.com/journal/11119/updates/17240272. Accessed 8 June 2021
Karkee M, Bhusal S, Zhang Q (2019) Apple dataset benchmark from orchard environment in modern fruiting wall. https://research.libraries.wsu.edu:8443/xmlui/handle/2376/17721. Accessed 16 June 2021
Koirala A, Walsh KB, Wang Z, McCarthy C (2019) Deep learning for real-time fruit detection and orchard fruit load estimation: Benchmarking of ‘MangoYOLO.’ Precis Agric 20(6):1107–1135
Kuznetsova A, Rom H, Alldrin N, Uijlings J, Krasin I, Pont-Tuset J, Ferrari V (2020) The open images dataset v4. Int J Comput Vis 1–26
Leaf counting challenge (2021) https://data-challenges.fz-juelich.de/web/challenges/challenge-page/85/overview. Accessed 10 June 2021
Lee DH (2013) Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks. In Workshop on challenges in representation learning, international conference on machine learning 3(2)
Leminen MS, Mathiassen SK, Dyrmann M, Laursen MS, Paz LC, Jørgensen, RN (2020). Open plant phenotype database of common weeds in denmark. Remote Sens 12(8):1246
Liu W, Anguelov D, Erhan D, Szegedy C, Reed S, Fu CY, Berg AC (2016) SSD: Single shot multibox detector. In European conference on computer vision, pp 21–37
Lobet G(2017) Image analysis in plant sciences: publish then perish. Trends Plant Sci 22(7):559–566
Lu Y, Young S (2020) A survey of public datasets for computer vision tasks in precision agriculture. Comput Electron Agric (178):105760
Mahajan S, Das A, Sardana, HK (2015) Image acquisition techniques for assessment of legume quality. Trends Food Sci Technol 42(2):116–133
Masjedi A, Zhao J, Thompson AM, Yang KW, Flatt JE, Crawford MM, Chapman S (2018) Sorghum biomass prediction using UAV-based remote sensing data and crop model simulation. In: International geoscience and remote sensing symposium 2018, pp 7719–7722
Minervini M, Fischbach A, Scharr H, Tsaftaris SA(2016) Finely-grained annotated datasets for image-based plant phenotyping. Pattern Recognit Lett (81):80–89
National Research Council (1997) Precision agriculture in the 21st century: geospatial and information technologies in crop management. Washington, DC, USA: National Academy Press
Olsen A, Konovalov DA, Philippa B, Ridd P, Wood JC, Johns J, White RD (2019) DeepWeeds: A multi-class weed species image dataset for deep learning. Sci Rep 9(1):1–12
Parikh A, Raval MS, Parmar C, Chaudhary S (2016) Disease detection and severity estimation in cotton plant from unconstrained images. In: IEEE international conference on data science and advanced analytics, pp 594–601
Pierce FJ, Nowak, P (1999) Aspects of precision agriculture. Adv Agron (67):1–85
Pierce FJ, Robert PC Mangold G (1994) Site-specific management: The pros, the cons, and the realities. In: Proceedings of the integrated crop management conference, Iowa State University, p 17–21
Redmon J, Farhadi A (2017) YOLO9000: better, faster, stronger. In: Proceedings of the IEEE International Conference on Pattern Recognition, pp 7263–7271
Redmon J, Farhadi A (2018) Yolov3: An incremental improvement. arXiv preprint arXiv:1804.02767
Redmon J, Divvala S, Girshick R, Farhadi A (2016) You only look once: Unified, real-time object detection. In: Proceedings of the IEEE international conference on pattern recognition, p 779–788
Ren S, He K, Girshick R, Sun J (2016) Faster R-CNN: towards real-time object detection with region proposal networks. IEEE PAMI 39(6):1137–1149
Samuel AL (1959) Some studies in machine learning using the game of checkers. IBM J Res Dev.3(3):210–229
Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556
Sorghum Biomass challenge (2021) https://www.kaggle.com/c/sorghum-biomassprediction/overview/description Accessed 9 June 2021
Stafford JV (1996) Essential technology for precision agriculture. In: Robert PC, Rust RH, Larson WE (eds), precision agriculture. proceedings third international conference on precision agriculture, pp 595–604
Stafford JV (2000) Implementing precision agriculture in the 21st century. J Agric Eng Res 76(3):267–275
Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z (2016) Rethinking the inception architecture for computer vision. In: IEEE international conference on pattern recognition, pp 2818–2826
Tan M, Pang R, Le QV (2020) Efficientdet: Scalable and efficient object detection. In: Proceedings of the international conference on pattern recognition, pp 10781–10790
Teimouri N, Dyrmann M, Nielsen PR, Mathiassen, SK Somerville GJ, Jørgensen RN (2018) Weed growth stage estimator using deep convolutional neural networks. Sensors 18(5):1580
United Nations (2019) World Population Prospects 2019. https://population.un.org/wpp/Download/Standard/CSV/. Accessed 8 June 2021
Walter A, Liebisch F, Hund A (2015) Plant phenotyping: from bean weighing to image analysis. Plant methods 11(1): 1–11
Wikipedia—Computer Vision. (2021) https://en.wikipedia.org/wiki/Computer_vision. Accessed 8 June 2021
Yang X, Shu L, Chen J, Ferrag MA, Wu J, Nurellari E, Huang K (2020) A survey on smart agriculture: Development modes, technologies, and security and privacy challenges. J Auto Sinica 8(2): 273–302
Yun S, Han D, Oh S J, Chun S, Choe J, Yoo Y (2019) Cutmix: Regularization strategy to train strong classifiers with localizable features. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 6023–6032
Zareiforoush H, Minaei S, Alizadeh MR, Banakar A (2015) Potential applications of computer vision in quality inspection of rice: a review. Food Eng Rev 7(3):321–345
Zeiler MD, Fergus R (2014) Visualizing and understanding convolutional networks. In European conference on computer vision, pp 818–833
Zhang H, Cisse M, Dauphin YN, Lopez-Paz D (2017). Mixup: Beyond empirical risk minimization. arXiv preprint arXiv:1710.09412
Zhang N, Wang M, Wang N (2002) Precision agriculture—a worldwide overview. Comput Electron Agric 36(2–3):113-132
Acknowledgements
The authors wholeheartedly thank all reviewers for super quick responses and insightful comments. We believe that it has significantly improved the quality of the chapter. The authors would like to thank Ms. Shailaja Sampat for the initiation of the work and interactions. It is imperative to thank all researchers contributing to creating datasets, open-source software, and their contribution to reproducible research.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this chapter
Cite this chapter
Raval, M.S., Chaudhary, S., Adinarayana, J. (2022). Computer Vision and Machine Learning in Agriculture. In: Reddy, G.P.O., Raval, M.S., Adinarayana, J., Chaudhary, S. (eds) Data Science in Agriculture and Natural Resource Management. Studies in Big Data, vol 96. Springer, Singapore. https://doi.org/10.1007/978-981-16-5847-1_5
Download citation
DOI: https://doi.org/10.1007/978-981-16-5847-1_5
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-16-5846-4
Online ISBN: 978-981-16-5847-1
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)