Learning to Predict Aboveground Biomass from RGB Images with 3D Synthetic Scenes
- URL: http://arxiv.org/abs/2511.23249v1
- Date: Fri, 28 Nov 2025 15:00:05 GMT
- Title: Learning to Predict Aboveground Biomass from RGB Images with 3D Synthetic Scenes
- Authors: Silvia Zuffi,
- Abstract summary: We propose a novel learning-based method for estimating aboveground biomass from a single ground-based RGB image.<n>We leverage the recently introduced synthetic 3D SPREAD dataset, which provides realistic forest scenes.<n>Our approach achieves a median AGB estimation error of 1.22 kg/m2 on held-out SPREAD data and 1.94 kg/m2 on a real-image dataset.
- Score: 8.063045613475234
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Forests play a critical role in global ecosystems by supporting biodiversity and mitigating climate change via carbon sequestration. Accurate aboveground biomass (AGB) estimation is essential for assessing carbon storage and wildfire fuel loads, yet traditional methods rely on labor-intensive field measurements or remote sensing approaches with significant limitations in dense vegetation. In this work, we propose a novel learning-based method for estimating AGB from a single ground-based RGB image. We frame this as a dense prediction task, introducing AGB density maps, where each pixel represents tree biomass normalized by the plot area and each tree's image area. We leverage the recently introduced synthetic 3D SPREAD dataset, which provides realistic forest scenes with per-image tree attributes (height, trunk and canopy diameter) and instance segmentation masks. Using these assets, we compute AGB via allometric equations and train a model to predict AGB density maps, integrating them to recover the AGB estimate for the captured scene. Our approach achieves a median AGB estimation error of 1.22 kg/m^2 on held-out SPREAD data and 1.94 kg/m^2 on a real-image dataset. To our knowledge, this is the first method to estimate aboveground biomass directly from a single RGB image, opening up the possibility for a scalable, interpretable, and cost-effective solution for forest monitoring, while also enabling broader participation through citizen science initiatives.
Related papers
- Direct Estimation of Tree Volume and Aboveground Biomass Using Deep Regression with Synthetic Lidar Data [4.588276431770691]
estimation of forest biomass is crucial for monitoring carbon sequestration and informing climate change mitigation strategies.<n>Existing methods often rely on allometric models, which estimate individual tree biomass by relating it to measurable biophysical parameters, e.g., trunk diameter and height.<n>This study proposes a direct approach that leverages synthetic point cloud data to train a deep regression network, which is then applied to real point clouds for plot-level wood volume estimation.
arXiv Detail & Related papers (2026-03-04T23:56:41Z) - Estimating Pasture Biomass from Top-View Images: A Dataset for Precision Agriculture [19.0810931631268]
We present a comprehensive dataset of 1,162 annotated top-view images of pastures collected across 19 locations in Australia.<n>Each image captures a 70cm * 30cm quadrat and is paired with on-ground measurements including biomass sorted by component (green, dead, and legumes fraction), vegetation height, and Normalized Difference Vegetation Index (NDVI) from Active Optical Sensors (AOS)<n>The dataset is released and hosted in a Kaggle competition that challenges the international Machine Learning community with the task of pasture biomass estimation.
arXiv Detail & Related papers (2025-10-27T01:35:00Z) - Statistical Confidence Rescoring for Robust 3D Scene Graph Generation from Multi-View Images [56.134885746889026]
semantic scene graph estimation methods utilize ground truth 3D annotations to accurately predict target objects, predicates, and relationships.<n>We overcome the noisy reconstructed pseudo point-based geometry from predicted depth maps and reduce the amount of background noise present in multi-view image features.<n>Our method outperforms current methods purely using multi-view images as the initial input.
arXiv Detail & Related papers (2025-08-05T21:25:50Z) - Bringing SAM to new heights: Leveraging elevation data for tree crown segmentation from drone imagery [68.69685477556682]
Current monitoring methods involve ground measurements, requiring extensive cost, time and labor.<n>Drone remote sensing and computer vision offer great potential for mapping individual trees from aerial imagery at broad-scale.<n>We compare methods leveraging Segment Anything Model (SAM) for the task of automatic tree crown instance segmentation in high resolution drone imagery.<n>We also study the integration of elevation data into models, in the form of Digital Surface Model (DSM) information, which can readily be obtained at no additional cost from RGB drone imagery.
arXiv Detail & Related papers (2025-06-05T12:43:11Z) - Estimating the Diameter at Breast Height of Trees in a Forest With a Single 360 Camera [52.85399274741336]
Forest inventories rely on accurate measurements of the diameter at breast height (DBH) for ecological monitoring, resource management, and carbon accounting.<n>While LiDAR-based techniques can achieve centimeter-level precision, they are cost-prohibitive and operationally complex.<n>We present a low-cost alternative that only needs a consumer-grade 360 video camera.
arXiv Detail & Related papers (2025-05-06T01:09:07Z) - CAP-Net: A Unified Network for 6D Pose and Size Estimation of Categorical Articulated Parts from a Single RGB-D Image [86.75098349480014]
This paper tackles category-level pose estimation of articulated objects in robotic manipulation tasks.<n>We propose a single-stage Network, CAP-Net, for estimating the 6D poses and sizes of Categorical Articulated Parts.<n>We introduce the RGBD-Art dataset, the largest RGB-D articulated dataset to date, featuring RGB images and depth noise simulated from real sensors.
arXiv Detail & Related papers (2025-04-15T14:30:26Z) - Estimation of forest height and biomass from open-access multi-sensor
satellite imagery and GEDI Lidar data: high-resolution maps of metropolitan
France [0.0]
This study uses a machine learning approach that was previously developed to produce local maps of forest parameters.
We used the GEDI Lidar mission as reference height data, and the satellite images from Sentinel-1, Sentinel-2 and ALOS-2 PALSA-2 to estimate forest height.
The height map is then derived into volume and aboveground biomass (AGB) using allometric equations.
arXiv Detail & Related papers (2023-10-23T07:58:49Z) - Mapping historical forest biomass for stock-change assessments at parcel
to landscape scales [0.0]
Map products can help identify where, when, and how forest carbon stocks are changing as a result of both anthropogenic and natural drivers alike.
These products can thus serve as inputs to a wide range of applications including stock-change assessments, monitoring reporting, and verification frameworks.
arXiv Detail & Related papers (2023-04-05T17:55:00Z) - Country-wide Retrieval of Forest Structure From Optical and SAR
Satellite Imagery With Bayesian Deep Learning [74.94436509364554]
We propose a Bayesian deep learning approach to densely estimate forest structure variables at country-scale with 10-meter resolution.
Our method jointly transforms Sentinel-2 optical images and Sentinel-1 synthetic aperture radar images into maps of five different forest structure variables.
We train and test our model on reference data from 41 airborne laser scanning missions across Norway.
arXiv Detail & Related papers (2021-11-25T16:21:28Z) - A Semantic Segmentation Network for Urban-Scale Building Footprint
Extraction Using RGB Satellite Imagery [1.9400948599830012]
Urban areas consume over two-thirds of the world's energy and account for more than 70 percent of global CO2 emissions.
We propose a modified DeeplabV3+ module with a Dilated ResNet backbone to generate masks of building footprints from only three-channel RGB satellite imagery.
We achieve state-of-the-art performance across three standard benchmarks and demonstrate that our method is agnostic to the scale, resolution, and urban density of satellite imagery.
arXiv Detail & Related papers (2021-04-02T22:32:04Z) - Refer-it-in-RGBD: A Bottom-up Approach for 3D Visual Grounding in RGBD
Images [69.5662419067878]
Grounding referring expressions in RGBD image has been an emerging field.
We present a novel task of 3D visual grounding in single-view RGBD image where the referred objects are often only partially scanned due to occlusion.
Our approach first fuses the language and the visual features at the bottom level to generate a heatmap that localizes the relevant regions in the RGBD image.
Then our approach conducts an adaptive feature learning based on the heatmap and performs the object-level matching with another visio-linguistic fusion to finally ground the referred object.
arXiv Detail & Related papers (2021-03-14T11:18:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.