FoSp: Focus and Separation Network for Early Smoke Segmentation
- URL: http://arxiv.org/abs/2306.04474v1
- Date: Wed, 7 Jun 2023 14:45:24 GMT
- Title: FoSp: Focus and Separation Network for Early Smoke Segmentation
- Authors: Lujian Yao, Haitao Zhao, Jingchao Peng, Zhongze Wang, Kaijie Zhao
- Abstract summary: Early smoke segmentation (ESS) enables the accurate identification of smoke sources, facilitating the prompt extinguishing of fires and preventing large-scale gas leaks.
ESS poses greater challenges than conventional object and regular smoke segmentation due to its small scale and transparent appearance.
We introduce a high-quality real-world dataset called SmokeSeg, which contains more small and transparent smoke than the existing datasets.
- Score: 0.6165605009782557
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Early smoke segmentation (ESS) enables the accurate identification of smoke
sources, facilitating the prompt extinguishing of fires and preventing
large-scale gas leaks. But ESS poses greater challenges than conventional
object and regular smoke segmentation due to its small scale and transparent
appearance, which can result in high miss detection rate and low precision. To
address these issues, a Focus and Separation Network (FoSp) is proposed. We
first introduce a Focus module employing bidirectional cascade which guides
low-resolution and high-resolution features towards mid-resolution to locate
and determine the scope of smoke, reducing the miss detection rate. Next, we
propose a Separation module that separates smoke images into a pure smoke
foreground and a smoke-free background, enhancing the contrast between smoke
and background fundamentally, improving segmentation precision. Finally, a
Domain Fusion module is developed to integrate the distinctive features of the
two modules which can balance recall and precision to achieve high F_beta.
Futhermore, to promote the development of ESS, we introduce a high-quality
real-world dataset called SmokeSeg, which contains more small and transparent
smoke than the existing datasets. Experimental results show that our model
achieves the best performance on three available datasets: SYN70K (mIoU:
83.00%), SMOKE5K (F_beta: 81.6%) and SmokeSeg (F_beta: 72.05%). Especially, our
FoSp outperforms SegFormer by 7.71% (F_beta) for early smoke segmentation on
SmokeSeg.
Related papers
- SALAD-Pan: Sensor-Agnostic Latent Adaptive Diffusion for Pan-Sharpening [50.44337053599724]
SALAD-Pan is a sensor-agnostic latent space diffusion method for efficient pansharpening.<n>It trains a band-wise single-channel VAE to encode high-resolution multispectral images into compact latent representations.<n>It achieves high-precision fusion in the diffusion process and robust zero-shot (cross-sensor) capability.
arXiv Detail & Related papers (2026-02-04T12:01:07Z) - SmokeBench: Evaluating Multimodal Large Language Models for Wildfire Smoke Detection [19.134309978060134]
SmokeBench is a benchmark to evaluate the ability of multimodal large language models (MLLMs) to recognize and localize wildfire smoke in images.<n>We evaluate several MLLMs, including Idefics2, Qwen2.5-VL, InternVL3, Unified-IO 2, Grounding DINO, GPT-4o, and Gemini-2.5 Pro.<n>Smoke volume is strongly correlated with model performance, whereas contrast plays a comparatively minor role.
arXiv Detail & Related papers (2025-12-12T01:47:28Z) - SGDFuse: SAM-Guided Diffusion for High-Fidelity Infrared and Visible Image Fusion [65.80051636480836]
This paper proposes a conditional diffusion model guided by the Segment Anything Model (SAM) to achieve high-fidelity and semantically-aware image fusion.<n>The framework operates in a two-stage process: it first performs a preliminary fusion of multi-modal features, and then utilizes the semantic masks as a condition to drive the diffusion model's coarse-to-fine denoising generation.<n>Extensive experiments demonstrate that SGDFuse achieves state-of-the-art performance in both subjective and objective evaluations.
arXiv Detail & Related papers (2025-08-07T10:58:52Z) - MFGDiffusion: Mask-Guided Smoke Synthesis for Enhanced Forest Fire Detection [6.307649189539342]
Smoke is the first visible indicator of a wildfire.<n>Current inpainting models exhibit limitations in generating high-quality smoke representations.<n>We propose a comprehensive framework for generating forest fire smoke images.
arXiv Detail & Related papers (2025-07-15T12:25:35Z) - DFVO: Learning Darkness-free Visible and Infrared Image Disentanglement and Fusion All at Once [57.15043822199561]
A Darkness-Free network is proposed to handle Visible and infrared image disentanglement and fusion all at Once (DFVO)<n>DFVO employs a cascaded multi-task approach to replace the traditional two-stage cascaded training (enhancement and fusion)<n>Our proposed approach outperforms state-of-the-art alternatives in terms of qualitative and quantitative evaluations.
arXiv Detail & Related papers (2025-05-07T15:59:45Z) - Fire and Smoke Detection with Burning Intensity Representation [14.738038604117364]
An effective Fire and Smoke Detection (FSD) and analysis system is of paramount importance due to the destructive potential of fire disasters.
Many existing FSD methods directly employ generic object detection techniques without considering the transparency of fire and smoke.
A new Attentive Fire and Smoke Detection Model (a-FSDM) is proposed to address this issue.
arXiv Detail & Related papers (2024-10-22T02:41:37Z) - High-Precision Dichotomous Image Segmentation via Probing Diffusion Capacity [69.32473738284374]
We propose DiffDIS, a diffusion-driven segmentation model that taps into the potential of the pre-trained U-Net within diffusion models.
By leveraging the robust generalization capabilities and rich, versatile image representation prior to the SD models, we significantly reduce the inference time while preserving high-fidelity, detailed generation.
Experiments on the DIS5K dataset demonstrate the superiority of DiffDIS, achieving state-of-the-art results through a streamlined inference process.
arXiv Detail & Related papers (2024-10-14T02:49:23Z) - Generative Edge Detection with Stable Diffusion [52.870631376660924]
Edge detection is typically viewed as a pixel-level classification problem mainly addressed by discriminative methods.
We propose a novel approach, named Generative Edge Detector (GED), by fully utilizing the potential of the pre-trained stable diffusion model.
We conduct extensive experiments on multiple datasets and achieve competitive performance.
arXiv Detail & Related papers (2024-10-04T01:52:23Z) - SIRST-5K: Exploring Massive Negatives Synthesis with Self-supervised
Learning for Robust Infrared Small Target Detection [53.19618419772467]
Single-frame infrared small target (SIRST) detection aims to recognize small targets from clutter backgrounds.
With the development of Transformer, the scale of SIRST models is constantly increasing.
With a rich diversity of infrared small target data, our algorithm significantly improves the model performance and convergence speed.
arXiv Detail & Related papers (2024-03-08T16:14:54Z) - CFDNet: A Generalizable Foggy Stereo Matching Network with Contrastive
Feature Distillation [11.655465312241699]
We introduce a framework based on contrastive feature distillation (CFD)
This strategy combines feature distillation from merged clean-fog features with contrastive learning, ensuring balanced dependence on fog depth hints and clean matching features.
arXiv Detail & Related papers (2024-02-28T09:12:01Z) - Wildfire Smoke Detection with Cross Contrast Patch Embedding [5.965059322800441]
The Transformer-based deep networks have increasingly shown significant advantages over CNNs.
Low-level information such as color, transparency and texture is very important for smoke recognition.
The fuzzy boundary of smoke makes the positive and negative label assignment for instances in a dilemma.
arXiv Detail & Related papers (2023-11-16T06:53:03Z) - FSDNet-An efficient fire detection network for complex scenarios based
on YOLOv3 and DenseNet [8.695064779659031]
This paper proposes a fire detection network called FSDNet (Fire Smoke Detection Network), which consists of a feature extraction module, a fire classification module, and a fire detection module.
The accuracy of FSDNet on the two benchmark datasets is 99.82% and 91.15%, respectively, and the average precision on MS-FS is 86.80%, which is better than the mainstream fire detection methods.
arXiv Detail & Related papers (2023-04-15T15:46:08Z) - Transmission-Guided Bayesian Generative Model for Smoke Segmentation [29.74065829663554]
Deep neural networks are prone to be overconfident for smoke segmentation due to its non-rigid shape and transparent appearance.
This is caused by both knowledge level uncertainty due to limited training data for accurate smoke segmentation and labeling level uncertainty representing the difficulty in labeling ground-truth.
We introduce a Bayesian generative model to simultaneously estimate the posterior distribution of model parameters and its predictions.
We also contribute a high-quality smoke segmentation dataset, SMOKE5K, consisting of 1,400 real and 4,000 synthetic images with pixel-wise annotation.
arXiv Detail & Related papers (2023-03-02T01:48:05Z) - Unified smoke and fire detection in an evolutionary framework with
self-supervised progressive data augment [5.8363672020565005]
In this study, we collect a large image data set to re-label them as a multi-label image classification problem.
We propose a data augment method by random image stitch to deploy resizing, deforming, position variation, and background altering.
Experiments show that the proposed method can effectively improve the generalization performance of the model for concurrent smoke and fire detection.
arXiv Detail & Related papers (2022-02-16T09:48:03Z) - Learning Selective Mutual Attention and Contrast for RGB-D Saliency
Detection [145.4919781325014]
How to effectively fuse cross-modal information is the key problem for RGB-D salient object detection.
Many models use the feature fusion strategy but are limited by the low-order point-to-point fusion methods.
We propose a novel mutual attention model by fusing attention and contexts from different modalities.
arXiv Detail & Related papers (2020-10-12T08:50:10Z) - InfoFocus: 3D Object Detection for Autonomous Driving with Dynamic
Information Modeling [65.47126868838836]
We propose a novel 3D object detection framework with dynamic information modeling.
Coarse predictions are generated in the first stage via a voxel-based region proposal network.
Experiments are conducted on the large-scale nuScenes 3D detection benchmark.
arXiv Detail & Related papers (2020-07-16T18:27:08Z) - A Single Stream Network for Robust and Real-time RGB-D Salient Object
Detection [89.88222217065858]
We design a single stream network to use the depth map to guide early fusion and middle fusion between RGB and depth.
This model is 55.5% lighter than the current lightest model and runs at a real-time speed of 32 FPS when processing a $384 times 384$ image.
arXiv Detail & Related papers (2020-07-14T04:40:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.