فصلنامه علمی- پژوهشی اطلاعات جغرافیایی « سپهر»

فصلنامه علمی- پژوهشی اطلاعات جغرافیایی « سپهر»

استخراج مرز زمین‌های کشاورزی از تصاویر ماهواره‌ای با یادگیری عمیق و شبکه‌های عصبی پیچشی

نوع مقاله : مقاله پژوهشی

نویسندگان
1 استادیار گروه عمران، دانشکده مهندسی، دانشگاه بوعلی سینا، همدان ، ایران
2 دانشجوی کارشناسی ارشد، گروه عمران، دانشکده مهندسی، دانشگاه بوعلی سینا، همدان ، ایران
3 دانشجوی دکتری، گروه کامپیوتر، دانشکده مهندسی، دانشگاه بوعلی سینا، همدان ، ایران
چکیده
هدف از استخراج اطلاعات مرزی، ایجاد پایگاه‌ اطلاعات زمین‌های کشاورزی است. با داشتن این معلومات می‌توان میزان مصرف آب و برداشت محصول را تخمین زد و هدایت ماشین‌آلات کشاورزی را خودکار نمود. برداشت میدانی مرز زمین‌ها زمان‌بر و پرهزینه است و برای این کار می‌توان روش‌های سنجش از دوری را مورد استفاده قرار داد. اما ترسیم دستی مرز زمین‌های کشاورزی از روی تصاویر سنجش از دوری، همچنان دشوار است. بنابراین استفاده از روش‌های خودکار راه‌حل مناسبی به نظر می رسد. این روش‌ها را می‌توان به دو دسته روش‌های مرسوم پردازش تصویر و روش‌های مبتنی بر یادگیری ماشین طبقه‌بندی کرد. روش‌های مرسوم پردازش تصویر مانند تشخیص لبه، مشکلاتی چون موقعیت‌یابی نادرست و دقت تشخیص ضعیف دارند. بنابراین محققان الگوریتم‌های تشخیص لبه نوین را بر اساس یادگیری‌عمیق، پیشنهاد نموده‌اند. شبکه‌های ‌عصبی‌ پیچشی ازجمله آن‌ها هستند که در این پژوهش به‌ کار رفته‌اند. برای پیاده‌سازی روش پیشنهادی، از زبان برنامه‌نویسی پایتون نسخه 3.11 در چارچوب کتابخانه keras استفاده شد. یکی از مشکلات استفاده از شبکه‌های عصبی پیچشی، کمبود مجموعه‌ داده آموزشی مناسب است. در پژوهش حاضر برای حل این مشکل، از فنون انتقال ‌یادگیری و تنظیم‌ دقیق استفاده شد. مجموعه‌ داده دسترسی آزاد فرانسه در کنار سه مجموعه داده ‌همدان، بهار و خرسان، برای آموزش شبکه‌های ‌عصبی ‌پیچشی به کار رفت. هشت سناریوی آزمایشی طراحی شد. پنج مورد از آن‌ها با تنظیم ‌دقیق و سه مورد دیگر بدون تنظیم ‌دقیق انجام گرفت. همچنین پنج حالت معماری مختلف از شبکه U-Net با شبکه‌های ‌پایه مختلف پیاده‌سازی شد. برای ارزیابی عملکرد، معیارهای Dice Score، IoU، Accuracy، Recall و F1-Score محاسبه شدند. در پایان مشخص شد انتقال ‌یادگیری و تنظیم ‌دقیق، روشی برای جبران کمبود داده‌های آموزشی و افزایش دقت هستند. در این پژوهش دقت عملکرد سناریوی سوم آزمایش،0.73T  در معیارIoU بود. وقتی همین سناریو با استفاده از تنظیم دقیق انجام شد، دقت آن 0.14 بهبود یافت و به 0.87 رسید. همچنین مکانیسم Attention در ترکیب با معماری‌های شبکه عصبی، دقت استخراج مرز را بهبود داد.
کلیدواژه‌ها

موضوعات


عنوان مقاله English

Extraction of agricultural land borders from satellite images using deep learning and convolutional neural networks

نویسندگان English

Morteza Heidarimozaffar 1
Sajjad Yavari 2
Zahra Dalvand 3
1 Assistant professor, Faculty of engineering, Bu-Ali Sina University, Hamedan, Iran
2 M.Sc. Student, Faculty of engineering, Bu-Ali Sina University, Hamedan, Iran
3 Ph.D Student, Faculty of engineering, Bu-Ali Sina University, Hamedan, Iran
چکیده English

Extended Abstract
Introduction
The purpose of extracting borders is to create a database of agricultural lands. With this information, it is possible to estimate the amount of water consumption and crop harvest and automate agricultural machinery's guidance. The field survey of land borders is time-consuming and expensive. Therefore, remote sensing methods can be used for this. Manually delineating agricultural land boundaries from remote sensing images is still difficult. Thus, using automatic algorithms is a good solution. These algorithms are divided into Traditional algorithms and algorithms based on deep learning. Traditional algorithms have problems such as incorrect positioning and poor detection accuracy. Therefore, researchers have proposed a variety of edge detection algorithms based on deep learning. Convolutional neural networks are among the algorithms used in this research.
 
Materials & Methods
To train convolutional neural networks, the open-access dataset of France and three datasets of Hamadan, Bahar, and Khersan were used. To conduct this research, Global mapper & Adobe Photoshop software was used to prepare images, draw label polygons, and unify the format of labels belonging to different datasets. Python programming language, Anaconda environment, and Cross and Tensor Flow libraries have been used to implement deep learning algorithms. Eight scenarios were tested with different datasets. Five of them were performed without the use of transfer learning and fine-tuning, and the other three were performed using fine-tuning. By combining U-Net architecture and pre-trained neural networks, five hybrid architecture modes were constructed. The attention mechanism was also used to improve the performance of the models. To evaluate the performance, Dice Score, IoU, Accuracy, Recall, and F1-Score metrics were calculated.
 
Results & Discussion
In the first scenario, the training and testing of the model was done on the French dataset. The sameness of these datasets in this scenario helps to compare different backbones. In the second scenario, the training and testing were done on the Hamadan dataset. In this scenario, due to the same training and test datasets, the performance of different backbones was compared. In the third scenario, training was done on the French dataset and testing on the Hamadan dataset. The purpose of designing this scenario is to evaluate the performance of the models on the images of different regions, with different spatial resolutions and sensors. In the fourth scenario, training was done on the French dataset and testing on the Bahar dataset. This scenario examines the possibility of using images with a spatial resolution of 10 meters for training and testing models. In the fifth scenario, training was done on the French dataset and testing was done on the Khersan dataset. In this scenario, the generalizability of the model was investigated. In the sixth scenario, training on the French dataset, fine-tuning on the Hamedan dataset, and testing on the Hamedan dataset were done. In this scenario, the effect of transfer learning was investigated. Also, the result of using images of different regions on the performance of the model was evaluated. In the seventh scenario, training on the French dataset, fine-tuning on the Hamadan dataset, and testing on the Khersan dataset were performed. One of the goals of this scenario is to evaluate the generalizability of the model. In the eighth scenario, the model was trained on the French dataset and fine-tuned on the Hamedan dataset. Finally, the test was performed on the Bahar dataset. The purpose of this scenario is to evaluate the effect of the spatial resolution of the test dataset images on the performance of the model.
 
Conclusion
In the first scenario, the accuracy of detection in the first, fourth, and fifth architectural states is similar and equal to 0.71. In the IoU metric. The recognition quality of the model is not favorable for small plots. One of the reasons is the low spatial resolution of images. In the second scenario, compared to the first scenario, the first, fourth, and fifth architectural states have suffered a drop in accuracy. Reducing the number of training datasets can be one of the reasons. In the third scenario, the number of Training images increased compared to the second scenario. As a result, the accuracy of the model's performance in the fourth architectural mode increased by 0.03 in the IOU metric compared to the second scenario. However, due to the difference in the resolution of the training and test datasets, the performance of the model was not satisfactory. The accuracy of the model in the fourth scenario decreased by 0.57 compared to the third scenario. The main difference between the third and fourth scenarios is in the spatial resolution of the test images. This shows that to achieve the desired results, the test images should have sufficient spatial resolution. In the fifth scenario, the accuracy of the first architectural mode decreased by 0.10 and the fourth architectural mode decreased by 0.18 in the IOU metric. This decrease in accuracy is due to the difference in topography and the size and shape of agricultural lands in the regions of France and Khersan. It can be concluded that the generalizability of the model is not suitable. The sixth scenario is the same as the third scenario, only fine-tuning on the Hamadan dataset has been added. The detection accuracy of the sixth scenario in the fourth architectural mode reached 0.87 and compared to the third scenario in the same architectural mode, it increased by 0.14. This shows that fine-tuning is effective in increasing model accuracy. The 7th scenario compared to the 5th scenario in the 4th architectural mode had a 0.04 increase in accuracy in the IOU metric due to the use of fine tuning. But still, the detection accuracy in this scenario was not desired. One of the reasons for this problem is the difference between agricultural lands in France and Khersan regions. It can be concluded that the existing models for use in different regions do not have proper generalization capabilities.

کلیدواژه‌ها English

The border of agricultural land
Remote sensing
Deep learning
Automatic boundary extraction algorithms
Convolutional neural network
Training dataset
1- Al-Wassai, F., & Kalyankar, N. (2013). Major Limitations of Satellite images.
2- Alzubaidi, L., Zhang, J., Humaidi, A. J., Al-Dujaili, A., Duan, Y., Al-Shamma, O., . . . Farhan, L. (2021). Review of deep learning: concepts, CNN architectures, challenges, applications, future directions. Journal of Big Data, 8(1), 53. doi:10.1186/s40537-021-00444-8
3- Arnold, L., Rebecchi, S., Chevallier, S., & Paugam-Moisy, H. (2011). An Introduction to Deep Learning (Vol. 1).
4- Chollet, F. (2017). Deep learning with python. New York, NY: Manning Publications.
www.github.com
5- Haton, J.-P. (2006). A brief introduction to artificial intelligence (Vol. 39).
6- Hoeser, T., Bachofer, F., & Kuenzer, C. (2020). Object Detection and Image Segmentation with Deep Learning on Earth Observation Data: A ReviewPart II: Applications. Remote Sensing, 12(18). doi:10.3390/rs12183053
7- Hoeser, T., & Kuenzer, C. (2020). Object Detection and Image Segmentation with Deep Learning on Earth Observation Data: A Review-Part I: Evolution and Recent Trends. Remote Sensing, 12(10). doi:10.3390/rs12101667
8- Krizhevsky, A., Sutskever, I., & Hinton, G. (2012). ImageNet Classification with Deep Convolutional Neural Networks. Neural Information Processing Systems, 25. doi:10.1145/3065386
9- Lagunas, M., & Garces, E. (2018). Transfer Learning for Illustration Classification.
10- Li, X., Xu, X., Yang, R., & Pu, F. (2020, 26 Sept.-2 Oct. 2020). DBC: Deep Boundaries Combination for Farmland Boundary Detection Based on UAV Imagery. Paper presented at the IGARSS 2020 - 2020 IEEE International Geoscience and Remote Sensing Symposium.
11- Long, J., Shelhamer, E., & Darrell, T. (2015). Fully convolutional networks for semantic segmentation.
12- Masoud, K. M., Persello, C., & Tolpekin, V. A. (2020). Delineation of Agricultural Field Boundaries from Sentinel-2 Images Using a Novel Super-Resolution Contour Detector Based on Fully Convolutional Networks. Remote Sensing, 12(1). doi:10.3390/rs12010059
13- Niu, Z., Zhong, G., & Yu, H. (2021). A review on the attention mechanism of deep learning. Neurocomputing, 452, 48-62. doi:https://doi.org/10.1016/j.neucom.2021.03.091
14- Persello, C., Tolpekin, V. A., Bergado, J. R., & de By, R. A. (2019). Delineation of agricultural fields in smallholder farms from satellite images using fully convolutional networks and combinatorial grouping. Remote Sensing of Environment, 231, 111253. doi:https://doi.org/10.1016/j.rse.2019.111253
15- Phiri, D., Simwanda, M., Salekin, S., Nyirenda, V. R., Murayama, Y., & Ranagalage, M. (2020). Sentinel-2 Data for Land Cover/Use Mapping: A Review. Remote Sensing, 12(14). doi:10.3390/rs12142291
16- Richards, J. (2022). Remote Sensing Digital Image Analysis.
17- Ronneberger, O., Fischer, P., & Brox, T. (2015, 2015//). U-Net: Convolutional Networks for Biomedical Image Segmentation. Paper presented at the Medical Image Computing and Computer-Assisted Intervention MICCAI 2015, Cham.
18- Sefercik, U. G., Alkan, M., Atalay, C., Jacobsen, K., Büyüksalih, G., & Karakış, S. (2020). Optimizing the Achievable Information Content Extraction from WorldView-4 Stereo Imagery. PFG Journal of Photogrammetry, Remote Sensing and Geoinformation Science, 88(6), 449-461. doi:10.1007/s41064-020-00127-8
19- Sun, R., Lei, T., Chen, Q., Wang, Z., Du, X., Zhao, W., & Nandi, A. (2022). Survey of Image Edge Detection. Frontiers in Signal Processing, 2, 826967. doi:10.3389/frsip.2022.826967
20- Tan, M., & Le, Q. (2019). EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks.
21- Wang, X., Shu, L., Han, R., Yang, F., Gordon, T., Wang, X., & Xu, H. (2023). A Survey of Farmland Boundary Extraction Technology Based on Remote Sensing Images. Electronics, 12(5). doi:10.3390/electronics12051156
22- Watkins, B., & van Niekerk, A. (2019). A comparison of object-based image analysis approaches for field boundary delineation using multi-temporal Sentinel-2 imagery. Computers and Electronics in Agriculture, 158, 294-302. doi:https://doi.org/10.1016/j.compag.2019.02.009
23- Zhang, Z., Zhang, X., Peng, C., Cheng, D., & Sun, J. (2018). ExFuse: Enhancing Feature Fusion for Semantic Segmentation.
24- Zhuang, F., Qi, Z., Duan, K., Xi, D., Zhu, Y., Zhu, H., . . . He, Q. (2021). A Comprehensive Survey on Transfer Learning. Proceedings of the IEEE, 109(1), 43-76. doi:10.1109/JPROC.2020.3004555