Self-supervised Learning for Hyperspectral Images of Trees
- URL: http://arxiv.org/abs/2509.05630v1
- Date: Sat, 06 Sep 2025 07:25:39 GMT
- Title: Self-supervised Learning for Hyperspectral Images of Trees
- Authors: Moqsadur Rahman, Saurav Kumar, Santosh S. Palmate, M. Shahriar Hossain,
- Abstract summary: This paper focuses on self-supervised learning to create neural network embeddings reflecting vegetation properties of trees from aerial hyperspectral images of crop fields.<n> Experimental results demonstrate that a constructed tree representation, using a vegetation property-related embedding space, performs better in downstream machine learning tasks compared to the direct use of hyperspectral vegetation properties as tree representations.
- Score: 2.2399170518036913
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Aerial remote sensing using multispectral and RGB imagers has provided a critical impetus to precision agriculture. Analysis of the hyperspectral images with limited or no labels is challenging. This paper focuses on self-supervised learning to create neural network embeddings reflecting vegetation properties of trees from aerial hyperspectral images of crop fields. Experimental results demonstrate that a constructed tree representation, using a vegetation property-related embedding space, performs better in downstream machine learning tasks compared to the direct use of hyperspectral vegetation properties as tree representations.
Related papers
- Multi-Image Super Resolution Framework for Detection and Analysis of Plant Roots [2.6737900598082835]
We propose a novel underground imaging system that captures multiple overlapping views of plant roots.<n>We use a deep learning-based Multi-Image Super Resolution (MISR) framework to enhance root visibility and detail.<n>Our approach outperforms state-of-the-art super resolution baselines, achieving a 2.3 percent reduction in BRISQUE.
arXiv Detail & Related papers (2026-01-09T02:30:48Z) - Data Augmentation and Resolution Enhancement using GANs and Diffusion Models for Tree Segmentation [49.13393683126712]
Urban forests play a key role in enhancing environmental quality and supporting biodiversity in cities.<n> accurately detecting trees is challenging due to complex landscapes and the variability in image resolution caused by different satellite sensors or UAV flight altitudes.<n>We propose a novel pipeline that integrates domain adaptation with GANs and Diffusion models to enhance the quality of low-resolution aerial images.
arXiv Detail & Related papers (2025-05-21T03:57:10Z) - DeepForest: Sensing Into Self-Occluding Volumes of Vegetation With Aerial Imaging [5.849820033135295]
Long-standing limitation of remote sensing to penetrate deep into dense canopy layers.<n> LiDAR and radar are currently considered the primary options for measuring 3D vegetation structures.<n>Our approach allows sensing deep into self-occluding vegetation volumes, such as forests.
arXiv Detail & Related papers (2025-02-04T09:45:49Z) - BranchPoseNet: Characterizing tree branching with a deep learning-based pose estimation approach [0.0]
This paper presents an automated pipeline for detecting tree whorls in proximally laser scanning data using a pose-estimation deep learning model.
Accurate whorl detection provides valuable insights into tree growth patterns, wood quality, and offers potential for use as a biometric marker to track trees throughout the forestry value chain.
arXiv Detail & Related papers (2024-09-23T07:10:11Z) - Classification of Single Tree Decay Stages from Combined Airborne LiDAR
Data and CIR Imagery [1.4589991363650008]
This study, for the first time, automatically categorizing individual trees (Norway spruce) into five decay stages.
Three different Machine Learning methods - 3D point cloud-based deep learning (KPConv), Convolutional Neural Network (CNN), and Random Forest (RF)
All models achieved promising results, reaching overall accuracy (OA) of up to 88.8%, 88.4% and 85.9% for KPConv, CNN and RF, respectively.
arXiv Detail & Related papers (2023-01-04T22:20:16Z) - Semantic Image Segmentation with Deep Learning for Vine Leaf Phenotyping [59.0626764544669]
In this study, we use Deep Learning methods to semantically segment grapevine leaves images in order to develop an automated object detection system for leaf phenotyping.
Our work contributes to plant lifecycle monitoring through which dynamic traits such as growth and development can be captured and quantified.
arXiv Detail & Related papers (2022-10-24T14:37:09Z) - End-to-end deep learning for directly estimating grape yield from
ground-based imagery [53.086864957064876]
This study demonstrates the application of proximal imaging combined with deep learning for yield estimation in vineyards.
Three model architectures were tested: object detection, CNN regression, and transformer models.
The study showed the applicability of proximal imaging and deep learning for prediction of grapevine yield on a large scale.
arXiv Detail & Related papers (2022-08-04T01:34:46Z) - Potato Crop Stress Identification in Aerial Images using Deep
Learning-based Object Detection [60.83360138070649]
The paper presents an approach for analyzing aerial images of a potato crop using deep neural networks.
The main objective is to demonstrate automated spatial recognition of a healthy versus stressed crop at a plant level.
Experimental validation demonstrated the ability for distinguishing healthy and stressed plants in field images, achieving an average Dice coefficient of 0.74.
arXiv Detail & Related papers (2021-06-14T21:57:40Z) - Generative Hierarchical Features from Synthesizing Images [65.66756821069124]
We show that learning to synthesize images can bring remarkable hierarchical visual features that are generalizable across a wide range of applications.
The visual feature produced by our encoder, termed as Generative Hierarchical Feature (GH-Feat), has strong transferability to both generative and discriminative tasks.
arXiv Detail & Related papers (2020-07-20T18:04:14Z) - Single Image Deraining via Scale-space Invariant Attention Neural
Network [58.5284246878277]
We tackle the notion of scale that deals with visual changes in appearance of rain steaks with respect to the camera.
We propose to represent the multi-scale correlation in convolutional feature domain, which is more compact and robust than that in pixel domain.
In this way, we summarize the most activated presence of feature maps as the salient features.
arXiv Detail & Related papers (2020-06-09T04:59:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.