Unlocking Zero-Shot Plant Segmentation with Pl@ntNet Intelligence
- URL: http://arxiv.org/abs/2510.12579v1
- Date: Tue, 14 Oct 2025 14:38:32 GMT
- Title: Unlocking Zero-Shot Plant Segmentation with Pl@ntNet Intelligence
- Authors: Simon Ravé, Jean-Christophe Lombardo, Pejman Rasti, Alexis Joly, David Rousseau,
- Abstract summary: We present a zero-shot segmentation approach for agricultural imagery.<n>Our method exploits Plantnet's specialized plant representations to identify plant regions.<n>We show consistent performance gains when using Plantnet-fine-tuned DinoV2 over the base DinoV2 model.
- Score: 3.7603674895765766
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: We present a zero-shot segmentation approach for agricultural imagery that leverages Plantnet, a large-scale plant classification model, in conjunction with its DinoV2 backbone and the Segment Anything Model (SAM). Rather than collecting and annotating new datasets, our method exploits Plantnet's specialized plant representations to identify plant regions and produce coarse segmentation masks. These masks are then refined by SAM to yield detailed segmentations. We evaluate on four publicly available datasets of various complexity in terms of contrast including some where the limited size of the training data and complex field conditions often hinder purely supervised methods. Our results show consistent performance gains when using Plantnet-fine-tuned DinoV2 over the base DinoV2 model, as measured by the Jaccard Index (IoU). These findings highlight the potential of combining foundation models with specialized plant-centric models to alleviate the annotation bottleneck and enable effective segmentation in diverse agricultural scenarios.
Related papers
- DepthCropSeg++: Scaling a Crop Segmentation Foundation Model With Depth-Labeled Data [8.868203469534269]
DepthCropSeg++ is a foundation model for crop segmentation, capable of segmenting different crop species under open in-field environment.<n>We build upon a state-of-the-art semantic segmentation architecture ViT-Adapter architecture, enhance it with dynamic upAdapter architecture, and train the model with a two-stage selftraining pipeline.<n>Results demonstrate that DepthCropSeg++ achieves 93.11% moU on a comprehensive testing set, outperforming both supervised baselines and general vision foundation models.
arXiv Detail & Related papers (2026-01-18T11:51:09Z) - Zero-shot Hierarchical Plant Segmentation via Foundation Segmentation Models and Text-to-image Attention [19.2882360692347]
Foundation segmentation models achieve reasonable leaf instance extraction from top-view crop images without training.<n>We introduce ZeroPlantSeg, a zero-shot segmentation for rosette-shaped plant individuals from top-view images.
arXiv Detail & Related papers (2025-09-11T02:53:58Z) - CellPainTR: Generalizable Representation Learning for Cross-Dataset Cell Painting Analysis [51.56484100374058]
We introduce CellPainTR, a Transformer-based architecture designed to learn foundational representations of cellular morphology.<n>Our work represents a significant step towards creating truly foundational models for image-based profiling, enabling more reliable and scalable cross-study biological analysis.
arXiv Detail & Related papers (2025-09-02T03:30:07Z) - PlantSAM: An Object Detection-Driven Segmentation Pipeline for Herbarium Specimens [0.5339846068056558]
We introduce PlantSAM, an automated segmentation pipeline that integrates YOLOv10 for plant region detection and the Segment Anything Model (SAM2) for segmentation.<n>YOLOv10 generates bounding box prompts to guide SAM2, enhancing segmentation accuracy.<n>PlantSAM achieved state-of-the-art segmentation performance, with an IoU of 0.94 and a Dice coefficient of 0.97.
arXiv Detail & Related papers (2025-07-22T12:02:39Z) - Bringing SAM to new heights: Leveraging elevation data for tree crown segmentation from drone imagery [68.69685477556682]
Current monitoring methods involve ground measurements, requiring extensive cost, time and labor.<n>Drone remote sensing and computer vision offer great potential for mapping individual trees from aerial imagery at broad-scale.<n>We compare methods leveraging Segment Anything Model (SAM) for the task of automatic tree crown instance segmentation in high resolution drone imagery.<n>We also study the integration of elevation data into models, in the form of Digital Surface Model (DSM) information, which can readily be obtained at no additional cost from RGB drone imagery.
arXiv Detail & Related papers (2025-06-05T12:43:11Z) - Iterative Optimization Annotation Pipeline and ALSS-YOLO-Seg for Efficient Banana Plantation Segmentation in UAV Imagery [11.048503703669667]
We develop ALSS-YOLO-Seg, an efficient lightweight segmentation model optimized for UAV imagery.
The model's backbone includes an Adaptive Lightweight Channel Splitting and Shuffling (ALSS) module to improve information exchange between channels.
A Multi-Scale Channel Attention (MSCA) module combines multi-scale feature extraction with channel attention to tackle challenges of varying target sizes and complex ground backgrounds.
arXiv Detail & Related papers (2024-10-09T13:19:26Z) - Generating Diverse Agricultural Data for Vision-Based Farming Applications [74.79409721178489]
This model is capable of simulating distinct growth stages of plants, diverse soil conditions, and randomized field arrangements under varying lighting conditions.
Our dataset includes 12,000 images with semantic labels, offering a comprehensive resource for computer vision tasks in precision agriculture.
arXiv Detail & Related papers (2024-03-27T08:42:47Z) - Semantic-SAM: Segment and Recognize Anything at Any Granularity [83.64686655044765]
We introduce Semantic-SAM, a universal image segmentation model to enable segment and recognize anything at any desired granularity.
We consolidate multiple datasets across three granularities and introduce decoupled classification for objects and parts.
For the multi-granularity capability, we propose a multi-choice learning scheme during training, enabling each click to generate masks at multiple levels.
arXiv Detail & Related papers (2023-07-10T17:59:40Z) - Semantic Image Segmentation with Deep Learning for Vine Leaf Phenotyping [59.0626764544669]
In this study, we use Deep Learning methods to semantically segment grapevine leaves images in order to develop an automated object detection system for leaf phenotyping.
Our work contributes to plant lifecycle monitoring through which dynamic traits such as growth and development can be captured and quantified.
arXiv Detail & Related papers (2022-10-24T14:37:09Z) - Multi-Domain Few-Shot Learning and Dataset for Agricultural Applications [0.0]
We propose a method to learn from a few samples to automatically classify different pests, plants, and their diseases.
We learn a feature extractor to generate embeddings and then update the embeddings using Transformers.
We conduct 42 experiments in total to comprehensively analyze the model and it achieves up to 14% and 24% performance gains on few-shot image classification benchmarks.
arXiv Detail & Related papers (2021-09-21T04:20:18Z) - Two-View Fine-grained Classification of Plant Species [66.75915278733197]
We propose a novel method based on a two-view leaf image representation and a hierarchical classification strategy for fine-grained recognition of plant species.
A deep metric based on Siamese convolutional neural networks is used to reduce the dependence on a large number of training samples and make the method scalable to new plant species.
arXiv Detail & Related papers (2020-05-18T21:57:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.