LiqD: A Dynamic Liquid Level Detection Model under Tricky Small
Containers
- URL: http://arxiv.org/abs/2403.08273v1
- Date: Wed, 13 Mar 2024 05:53:25 GMT
- Title: LiqD: A Dynamic Liquid Level Detection Model under Tricky Small
Containers
- Authors: Yukun Ma, Zikun Mao
- Abstract summary: This paper proposes a container dynamic liquid level detection model based on U2-Net.
A large number of experimental results show that the proposed model can effectively detect the dynamic liquid level changes of the liquid in the container.
- Score: 5.361320134021586
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: In daily life and industrial production, it is crucial to accurately detect
changes in liquid level in containers. Traditional contact measurement methods
have some limitations, while emerging non-contact image processing technology
shows good application prospects. This paper proposes a container dynamic
liquid level detection model based on U^2-Net. This model uses the SAM model to
generate an initial data set, and then evaluates and filters out high-quality
pseudo-label images through the SemiReward framework to build an exclusive data
set. The model uses U^2-Net to extract mask images of containers from the data
set, and uses morphological processing to compensate for mask defects.
Subsequently, the model calculates the grayscale difference between adjacent
video frame images at the same position, segments the liquid level change area
by setting a difference threshold, and finally uses a lightweight neural
network to classify the liquid level state. This approach not only mitigates
the impact of intricate surroundings, but also reduces the demand for training
data, showing strong robustness and versatility. A large number of experimental
results show that the proposed model can effectively detect the dynamic liquid
level changes of the liquid in the container, providing a novel and efficient
solution for related fields.
Related papers
- Stable Flow: Vital Layers for Training-Free Image Editing [74.52248787189302]
Diffusion models have revolutionized the field of content synthesis and editing.
Recent models have replaced the traditional UNet architecture with the Diffusion Transformer (DiT)
We propose an automatic method to identify "vital layers" within DiT, crucial for image formation.
Next, to enable real-image editing, we introduce an improved image inversion method for flow models.
arXiv Detail & Related papers (2024-11-21T18:59:51Z) - AssemAI: Interpretable Image-Based Anomaly Detection for Manufacturing Pipelines [0.0]
Anomaly detection in manufacturing pipelines remains a critical challenge, intensified by the complexity and variability of industrial environments.
This paper introduces AssemAI, an interpretable image-based anomaly detection system tailored for smart manufacturing pipelines.
arXiv Detail & Related papers (2024-08-05T01:50:09Z) - Robust CLIP-Based Detector for Exposing Diffusion Model-Generated Images [13.089550724738436]
Diffusion models (DMs) have revolutionized image generation, producing high-quality images with applications spanning various fields.
Their ability to create hyper-realistic images poses significant challenges in distinguishing between real and synthetic content.
This work introduces a robust detection framework that integrates image and text features extracted by CLIP model with a Multilayer Perceptron (MLP) classifier.
arXiv Detail & Related papers (2024-04-19T14:30:41Z) - Inpainting Computational Fluid Dynamics with Deep Learning [8.397730500554047]
An effective fluid data completion method reduces the required number of sensors in a fluid dynamics experiment.
The ill-posed nature of the fluid data completion problem makes it prohibitively difficult to obtain a theoretical solution.
We employ the vector quantization technique to map both complete and incomplete fluid data spaces onto discrete-valued lower-dimensional representations.
arXiv Detail & Related papers (2024-02-27T03:44:55Z) - Unpaired Downscaling of Fluid Flows with Diffusion Bridges [0.0]
We show how one can chain together two independent conditional diffusion models for use in domain translation.
The resulting transformation is a diffusion bridge between a low resolution and a high resolution dataset.
We demonstrate that the method enhances resolution and corrects context-dependent biases in geophysical fluid simulations.
arXiv Detail & Related papers (2023-05-02T23:13:44Z) - An Adversarial Active Sampling-based Data Augmentation Framework for
Manufacturable Chip Design [55.62660894625669]
Lithography modeling is a crucial problem in chip design to ensure a chip design mask is manufacturable.
Recent developments in machine learning have provided alternative solutions in replacing the time-consuming lithography simulations with deep neural networks.
We propose a litho-aware data augmentation framework to resolve the dilemma of limited data and improve the machine learning model performance.
arXiv Detail & Related papers (2022-10-27T20:53:39Z) - RetiFluidNet: A Self-Adaptive and Multi-Attention Deep Convolutional
Network for Retinal OCT Fluid Segmentation [3.57686754209902]
Quantification of retinal fluids is necessary for OCT-guided treatment management.
New convolutional neural architecture named RetiFluidNet is proposed for multi-class retinal fluid segmentation.
Model benefits from hierarchical representation learning of textural, contextual, and edge features.
arXiv Detail & Related papers (2022-09-26T07:18:00Z) - Visual-tactile sensing for Real-time liquid Volume Estimation in
Grasping [58.50342759993186]
We propose a visuo-tactile model for realtime estimation of the liquid inside a deformable container.
We fuse two sensory modalities, i.e., the raw visual inputs from the RGB camera and the tactile cues from our specific tactile sensor.
The robotic system is well controlled and adjusted based on the estimation model in real time.
arXiv Detail & Related papers (2022-02-23T13:38:31Z) - A Comprehensive Study of Image Classification Model Sensitivity to
Foregrounds, Backgrounds, and Visual Attributes [58.633364000258645]
We call this dataset RIVAL10 consisting of roughly $26k$ instances over $10$ classes.
We evaluate the sensitivity of a broad set of models to noise corruptions in foregrounds, backgrounds and attributes.
In our analysis, we consider diverse state-of-the-art architectures (ResNets, Transformers) and training procedures (CLIP, SimCLR, DeiT, Adversarial Training)
arXiv Detail & Related papers (2022-01-26T06:31:28Z) - Visual Saliency Transformer [127.33678448761599]
We develop a novel unified model based on a pure transformer, Visual Saliency Transformer (VST), for both RGB and RGB-D salient object detection (SOD)
It takes image patches as inputs and leverages the transformer to propagate global contexts among image patches.
Experimental results show that our model outperforms existing state-of-the-art results on both RGB and RGB-D SOD benchmark datasets.
arXiv Detail & Related papers (2021-04-25T08:24:06Z) - Optical Flow Estimation from a Single Motion-blurred Image [66.2061278123057]
Motion blur in an image may have practical interests in fundamental computer vision problems.
We propose a novel framework to estimate optical flow from a single motion-blurred image in an end-to-end manner.
arXiv Detail & Related papers (2021-03-04T12:45:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.