DOI QR코드

DOI QR Code

Localization of ripe tomato bunch using deep neural networks and class activation mapping

  • Seung-Woo Kang (Department of Biosystems Machinery Engineering, Chungnam National University) ;
  • Soo-Hyun Cho (Department of Biosystems Machinery Engineering, Chungnam National University) ;
  • Dae-Hyun Lee (Department of Biosystems Machinery Engineering, Chungnam National University) ;
  • Kyung-Chul Kim (Department of Agricultural Engineering, National Institute of Agricultural Sciences)
  • 투고 : 2023.02.09
  • 심사 : 2023.07.10
  • 발행 : 2023.09.01

초록

In this study, we propose a ripe tomato bunch localization method based on convolutional neural networks, to be applied in robotic harvesting systems. Tomato images were obtained from a smart greenhouse at the Rural Development Administration (RDA). The sample images for training were extracted based on tomato maturity and resized to 128 × 128 pixels for use in the classification model. The model was constructed based on four-layer convolutional neural networks, and the classes were determined based on stage of maturity, using a Softmax classifier. The localization of the ripe tomato bunch region was indicated on a class activation map. The class activation map could show the approximate location of the tomato bunch but tends to present a local part or a large part of the ripe tomato bunch region, which could lead to poor performance. Therefore, we suggest a recursive method to improve the performance of the model. The classification results indicated that the accuracy, precision, recall, and F1-score were 0.98, 0.87, 0.98, and 0.92, respectively. The localization performance was 0.52, estimated by the Intersection over Union (IoU), and through input recursion, the IoU was improved by 13%. Based on the results, the proposed localization of the ripe tomato bunch area can be incorporated in robotic harvesting systems to establish the optimal harvesting paths.

키워드

과제정보

본 결과물은 농림축산식품부 및 과학기술정보통신부, 농촌진흥청의 재원으로 농림식품기술기획평가원과 재단법인 스마트팜연구개발사업단의 스마트팜다부처 패키지혁신기술개발사업의 지원을 받아 연구되었음(421031-04).

참고문헌

  1. Huang Y, Si W, Chen K, Sun Y. 2020. Assessment of tomato maturity in different layers by spatially resolved spectroscopy. Sensors 20:7229.
  2. Kim TH, Lee DH, Kang SW, Cho SH, Kim KC. 2022a. Estimation of tomato maturity as a continuous index using deep neural networks. Korean Journal of Agricultural Science 49:785-793. [in Korean] https://doi.org/10.7744/KJOAS.20220072
  3. Kim TH, Lee DH, Kim KC, Choi TY, Yu JM. 2022b. Tomato maturity estimation using deep neural network. Applied Sciences 13:412.
  4. Kim JM, Pyo HR, Jang IH, Kang JH, Ju BK, Ko KE. 2022c. Tomato harvesting robotic system based on Deep-ToMaToS: Deep learning network using transformation loss for 6D pose estimation of maturity classified tomatoes with side-stem. Computers and Electronics in Agriculture 201:107300.
  5. Kim WS, Lee DH, Kim YJ. 2020. Machine vision-based automatic disease symptom detection of onion downy mildew. Computers and Electronics in Agriculture 168:105099.
  6. Lee KB, Kim YB, Cho BH, Kim WK, Kim MJ, Hong YK, Kim KC. 2022. Implementation of 3D location detection embedded system for tomato harvesting robots. The Journal of Korean Institute of Communications and Information Sciences 47:2007-2019. [in Korean] https://doi.org/10.7840/kics.2022.47.11.2007
  7. Seo D, Cho B, Kim KC. 2021. Development of monitoring robot system for tomato fruits in hydroponic greenhouses. Agronomy 11:2211.
  8. Won JH, Yang CJ, Kwon KD, Hong YK, Kim GH. 2022. Study on the technology of target recognition using RGBD camera for automation harvest. In Proceeding of Korea Society for Agricultural Machinery 219. [in Korean]
  9. Zhang F, Gao J, Zhou H, Zhang J, Zou K, Yuan T. 2022. Three-dimensional pose detection method based on keypoints detection network for tomato bunch. Computers and Electronics in Agriculture 195:106824.
  10. Zhou B, Khosla A, Lapedriza A, Oliva A, Torralba A. 2016. Learning deep features for discriminative localization. In Proceedings of the IEEE conference on computer vision and pattern recognition 2921:2929.