2.1.1Problem statement¶
Hyperspectral remote sensing has demonstrated potential in automated vegetation mapping applications Govender et al., 2009Zhong et al., 2022. In addition to condition monitoring, the enhanced spectral depth is particularly valuable in discriminating plant species in complex forest canopies Pereira Martins-Neto et al., 2023. Despite the increasing accessibility of hyperspectral sensors, they are rarely deployed in real-world conservation scenarios due to their cost and complexity.
In recent years, deep learning algorithms have emerged as powerful tools for computer vision tasks and offer several key advantages over traditional classification approaches in the context of vegetation mapping. DL algorithms have proven highly effective in semantic segmentation tasks targeting vegetation species using high resolution hyperspectral data Zhang et al., 2020. Likewise, studies using high resolution RGB-only datasets for similar task and targets have also achieved high accuracy with deep learning approaches Egli & Höpke, 2020Garzon-Lopez & Lasso, 2020.
Despite strong evidence showing that deep learning is effective in species-level vegetation mapping tasks without access to rich spectral information, there are limited studies comparing the two input feature sets Nezami et al., 2020. Studies comparing high-resolution RGB and hyperspectral datasets typically focus on assessing plant condition in agricultural scenarios, where the additional spectral depth significantly improves accuracy Xu et al., 2025.
Research comparing the performance of deep learning segmentation models trained on RGB imagery versus those informed by hyperspectral data is sparse, particularly in complex forest canopies. Furthermore, there is limited understanding of the underlying mechanisms driving performance differences between these approaches. Specifically whether improved accuracy stems from the additional spectral information content itself, or from the complex spectral-spatial relationships that deep learning models can exploit within hyperspectral datasets.
2.1.2Proposed method¶
Airborne data collection will employ fixed-wing aircraft platforms to acquire co-registered optical and hyperspectral imagery across all study sites. Flight operations will target 500 metres above ground level to optimise the balance between spatial resolution and survey efficiency. This altitude specification will yield ground sampling distances of approximately 0.03 metres for RGB imagery and 0.40 metres for hyperspectral data.
Optical imagery acquisition will utilise a PhaseOne PAS 150 frame camera system, with subsequent processing into three-band orthomosaics using photogrammetric workflows. Hyperspectral data collection will employ simultaneous deployment of Specim FX10 and AFX17 sensors (Specim Ltd, Finland). The FX10 is sensitive to near-infrared (NIR) wavelengths (400-1000 nm) across 224 spectral bands, while the AFX17 is sensitive to shortwave infrared (SWIR) wavelengths (900-1700 nm) across 224 bands.
Co-registration between hyperspectral and RGB datasets will leverage the methodology developed by, Haynes et al., Haynes et al., 2025 which utilised the same sensor combination as we propose. The two datasets will be stacked into a multiband raster during preprocessing.
Ground truth data will be established through manual delineation of individual tree crowns for target taxa using geographic information system software. Crown boundary annotation will be performed on the analysis-ready imagery by expert botanists. To ensure annotation accuracy and reduce interpretation bias, a randomly selected subset of desktop-derived annotations will undergo field validation through direct ground-based observation.
The initial reference dataset will be augmented using established data augmentation techniques Mumuni & Mumuni, 2022 to increase sample size and improve model robustness.
The analysis will evaluate the contribution of spectral information to species-level semantic segmentation accuracy through systematic comparison of two-dimensional (2D) and three-dimensional (3D) implementations of established deep learning architectures. This comparison will isolate the value of spatial-spectral relationships in hyperspectral data while controlling for architectural differences.
The analysis will concentrate on 2D and 3D variants of U-Net architectures Ronneberger et al., 2015, which have demonstrated established efficacy in remote sensing applications, particularly for vegetation segmentation in high-resolution imagery [Flood et al. (2019); schieferMappingForestTree2020]. By constraining the comparison to a single architectural family, the contribution of spatial-spectral relationships can be isolated without confounding effects from different network designs.
Model performance will be assessed using pixel-wise accuracy metrics appropriate for semantic segmentation tasks in remote sensing contexts Maxwell et al., 2021. Cross-validation strategies will ensure robust performance estimates and evaluate model transferability across different study sites.
To understand the spectral and spatial features driving species discrimination, model interpretation will employ multiple explainability techniques. Gradient-weighted Class Activation Mapping (Grad-CAM) will identify spatial regions most influential for species classification decisions Onishi & Ise, 2021. Shapley Additive Explanations (SHAP) values will quantify the contribution of individual spectral bands to model predictions Huang et al., 2020. These interpretation methods will provide insights into whether spectral relationships are as critical as spatial relationships for accurate species identification.
2.1.3Key innovation¶
This research addresses three critical knowledge gaps in remote sensing-based vegetation mapping. First, it will determine which deep neural network architecture yields the most accurate segmentation of Tasmanian vegetation species using hyperspectral imagery. Second, it will quantify the spectral requirements for accurate vegetation segmentation, specifically evaluating the potential of RGB-only datasets as alternatives to hyperspectral data for operational mapping applications. Third, it will investigate whether spectral relationships are as important as spatial relationships for species identification, providing fundamental insights into the mechanisms underlying deep learning-based vegetation discrimination.
- Govender, M., Chetty, K., & Bulcock, H. (2009). A Review of Hyperspectral Remote Sensing and Its Application in Vegetation and Water Resource Studies. Water SA, 33(2). 10.4314/wsa.v33i2.49049
- Zhong, H., Lin, W., Liu, H., Ma, N., Liu, K., Cao, R., Wang, T., & Ren, Z. (2022). Identification of Tree Species Based on the Fusion of UAV Hyperspectral Image and LiDAR Data in a Coniferous and Broad-Leaved Mixed Forest in Northeast China. Frontiers in Plant Science, 13, 964769. 10.3389/fpls.2022.964769
- Pereira Martins-Neto, R., Garcia Tommaselli, A. M., Imai, N. N., Honkavaara, E., Miltiadou, M., Saito Moriya, E. A., & David, H. C. (2023). Tree Species Classification in a Complex Brazilian Tropical Forest Using Hyperspectral and LiDAR Data. Forests, 14(5), 945. 10.3390/f14050945
- Zhang, B., Zhao, L., & Zhang, X. (2020). Three-Dimensional Convolutional Neural Network Model for Tree Species Classification Using Airborne Hyperspectral Images. Remote Sensing of Environment, 247, 111938. 10.1016/j.rse.2020.111938
- Egli, S., & Höpke, M. (2020). CNN-based Tree Species Classification Using High Resolution RGB Image Data from Automated UAV Observations. Remote Sensing, 12(23), 3892. 10.3390/rs12233892
- Garzon-Lopez, C. X., & Lasso, E. (2020). Species Classification in a Tropical Alpine Ecosystem Using UAV-borne RGB and Hyperspectral Imagery. Drones, 4(4), 69.
- Nezami, S., Khoramshahi, E., Nevalainen, O., Pölönen, I., & Honkavaara, E. (2020). Tree Species Classification of Drone Hyperspectral and RGB Imagery with Deep Learning Convolutional Neural Networks. Remote Sensing, 12(7), 1070. 10.3390/rs12071070
- Xu, Y., Mao, Y., Li, H., Shen, J., Xu, X., Wang, S., Zaman, S., Ding, Z., & Wang, Y. (2025). A Deep Learning Model Based on RGB and Hyperspectral Images for Efficiently Detecting Tea Green Leafhopper Damage Symptoms. Smart Agricultural Technology, 10, 100817. 10.1016/j.atech.2025.100817
- Haynes, R. S., Lucieer, A., Turner, D., & Cimoli, E. (2025). Co-Registration of Multi-Modal UAS Pushbroom Imaging Spectroscopy and RGB Imagery Using Optical Flow. Drones, 9(2), 132. 10.3390/drones9020132
- Mumuni, A., & Mumuni, F. (2022). Data Augmentation: A Comprehensive Survey of Modern Approaches. Array, 16, 100258. 10.1016/j.array.2022.100258
- Ronneberger, O., Fischer, P., & Brox, T. (2015). U-Net: Convolutional Networks for Biomedical Image Segmentation. arXiv. 10.48550/arXiv.1505.04597
- Flood, N., Watson, F., & Collett, L. (2019). Using a U-net Convolutional Neural Network to Map Woody Vegetation Extent from High Resolution Satellite Imagery across Queensland, Australia. International Journal of Applied Earth Observation and Geoinformation, 82, 101897. 10.1016/j.jag.2019.101897
- Maxwell, A. E., Warner, T. A., & Guillén, L. A. (2021). Accuracy Assessment in Convolutional Neural Network-Based Deep Learning Remote Sensing Studies—Part 2: Recommendations and Best Practices. Remote Sensing, 13(13), 2591. 10.3390/rs13132591
- Onishi, M., & Ise, T. (2021). Explainable Identification and Mapping of Trees Using UAV RGB Image and Deep Learning. Scientific Reports, 11(1), 903. 10.1038/s41598-020-79653-9
- Huang, X., Kroening, D., Ruan, W., Sharp, J., Sun, Y., Thamo, E., Wu, M., & Yi, X. (2020). A Survey of Safety and Trustworthiness of Deep Neural Networks: Verification, Testing, Adversarial Attack and Defence, and Interpretability. Computer Science Review, 37, 100270. 10.1016/j.cosrev.2020.100270