A Comparative Analysis of Deep Learning Methods for Ghaf Tree Detection and Segmentation from UAV-based Images
- Mohamed Ali
- Feb 11
- 2 min read
Authors: Hani Shanableh, Mohamed Barakat A. Gibril, Ahmed Mansour, Aditya Dixit, Rami Al-Ruzouq, Nezar Hammouri, Fouad Lamghari, Safa M. Ahmed, Ratiranjan Jena, Tilal Mohamed, Mohammed Abdulraheem Almarzouqi, Nedal Salem Alafayfeh, and Simon Zerisenay Ghebremeskel.

Selected images from the testing dataset (a–f) with corresponding annotations, presenting results from four models: Mask R-CNN with MViTv2-tiny Mask R-CNN with ViT-base, Mask DINO, and Mask R-CNN with Swin transformer.
Abstract: The Prosopis cineraria, commonly known as the Ghaf tree, is an ecologically significant species that prevents desertification, enhances soil fertility, and supports biodiversity within arid ecosystems. Mapping and monitoring Ghaf trees using unmanned aerial systems (UAVs) and deep learning are essential for advancing conservation efforts through reliable, automated assessments. In this study, we performed a comparative analysis of several transformer-based deep learning models, including Mask DETR with Improved Denoising Anchor Boxes (Mask DINO) and Mask R-CNN models based on the Vision Transformer (ViT), Swin Transformer, and Enhanced Multiscale ViT (MViTv2), for mapping Ghaf trees from UAV images captured in diverse urban and agricultural environments. Results demonstrated strong potential for the assessed instance segmentation architectures in mapping Ghaf trees, achieving mean average precision values of 80% to 84.2% for detection and 82.2% to 85.1% for segmentation, with F1-scores ranging from 83.55% to 88.3% for detection and 85.5% to 88.6% for segmentation. This study underscores the effectiveness of transformer-based deep learning architectures for mapping Ghaf trees from UAV images, with findings and refinements that can be applied and extended to map other native tree species and support broader conservation initiatives.



Comments