Author
Listed:
- Ziang Niu
(College of Biosystems Engineering and Food Science, Zhejiang University, Hangzhou 310058, China)
- Ting Huang
(College of Biosystems Engineering and Food Science, Zhejiang University, Hangzhou 310058, China)
- Chengjia Xu
(College of Biosystems Engineering and Food Science, Zhejiang University, Hangzhou 310058, China)
- Xinyue Sun
(College of Biosystems Engineering and Food Science, Zhejiang University, Hangzhou 310058, China)
- Mohamed Farag Taha
(School of Agricultural Engineering, Jiangsu University, Zhenjiang 212013, China
Department of Soil and Water Sciences, Faculty of Environmental Agricultural Sciences, Arish University, North Sinai 45516, Egypt)
- Yong He
(College of Biosystems Engineering and Food Science, Zhejiang University, Hangzhou 310058, China)
- Zhengjun Qiu
(College of Biosystems Engineering and Food Science, Zhejiang University, Hangzhou 310058, China)
Abstract
Maize leaf area offers valuable insights into physiological processes, playing a critical role in breeding and guiding agricultural practices. The Azure Kinect DK possesses the real-time capability to capture and analyze the spatial structural features of crops. However, its further application in maize leaf area measurement is constrained by RGB–depth misalignment and limited sensitivity to detailed organ-level features. This study proposed a novel approach to address and optimize the limitations of the Azure Kinect DK through the multimodal coupling of RGB-D data for enhanced organ-level crop phenotyping. To correct RGB–depth misalignment, a unified recalibration method was developed to ensure accurate alignment between RGB and depth data. Furthermore, a semantic information-guided depth inpainting method was proposed, designed to repair void and flying pixels commonly observed in Azure Kinect DK outputs. The semantic information was extracted using a joint YOLOv11-SAM2 model, which utilizes supervised object recognition prompts and advanced visual large models to achieve precise RGB image semantic parsing with minimal manual input. An efficient pixel filter-based depth inpainting algorithm was then designed to inpaint void and flying pixels and restore consistent, high-confidence depth values within semantic regions. A validation of this approach through leaf area measurements in practical maize field applications—challenged by a limited workspace, constrained viewpoints, and environmental variability—demonstrated near-laboratory precision, achieving an MAPE of 6.549%, RMSE of 4.114 cm 2 , MAE of 2.980 cm 2 , and R 2 of 0.976 across 60 maize leaf samples. By focusing processing efforts on the image level rather than directly on 3D point clouds, this approach markedly enhanced both efficiency and accuracy with the sufficient utilization of the Azure Kinect DK, making it a promising solution for high-throughput 3D crop phenotyping.
Suggested Citation
Ziang Niu & Ting Huang & Chengjia Xu & Xinyue Sun & Mohamed Farag Taha & Yong He & Zhengjun Qiu, 2025.
"A Novel Approach to Optimize Key Limitations of Azure Kinect DK for Efficient and Precise Leaf Area Measurement,"
Agriculture, MDPI, vol. 15(2), pages 1-20, January.
Handle:
RePEc:gam:jagris:v:15:y:2025:i:2:p:173-:d:1566936
Download full text from publisher
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jagris:v:15:y:2025:i:2:p:173-:d:1566936. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
We have no bibliographic references for this item. You can help adding them by using this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.