Improving Data-Efficient Fossil Segmentation via Model Editing
- URL: http://arxiv.org/abs/2210.03879v2
- Date: Mon, 10 Apr 2023 03:43:57 GMT
- Title: Improving Data-Efficient Fossil Segmentation via Model Editing
- Authors: Indu Panigrahi, Ryan Manzuk, Adam Maloof, Ruth Fong
- Abstract summary: We present a two-part paradigm to improve fossil segmentation with few labeled images.
We apply domain-informed image perturbations to expose the Mask R-CNN's inability to distinguish between different classes of fossils.
We extend an existing model-editing method for correcting systematic mistakes in image classification to image segmentation with no additional labeled data needed.
- Score: 4.683612295430956
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Most computer vision research focuses on datasets containing thousands of
images of commonplace objects. However, many high-impact datasets, such as
those in medicine and the geosciences, contain fine-grain objects that require
domain-expert knowledge to recognize and are time-consuming to collect and
annotate. As a result, these datasets contain few labeled images, and current
machine vision models cannot train intensively on them. Originally introduced
to correct large-language models, model-editing techniques in machine learning
have been shown to improve model performance using only small amounts of data
and additional training. Using a Mask R-CNN to segment ancient reef fossils in
rock sample images, we present a two-part paradigm to improve fossil
segmentation with few labeled images: we first identify model weaknesses using
image perturbations and then mitigate those weaknesses using model editing.
Specifically, we apply domain-informed image perturbations to expose the Mask
R-CNN's inability to distinguish between different classes of fossils and its
inconsistency in segmenting fossils with different textures. To address these
shortcomings, we extend an existing model-editing method for correcting
systematic mistakes in image classification to image segmentation with no
additional labeled data needed and show its effectiveness in decreasing
confusion between different kinds of fossils. We also highlight the best
settings for model editing in our situation: making a single edit using all
relevant pixels in one image (vs. using multiple images, multiple edits, or
fewer pixels). Though we focus on fossil segmentation, our approach may be
useful in other similar fine-grain segmentation problems where data is limited.
Related papers
- UnSeg: One Universal Unlearnable Example Generator is Enough against All Image Segmentation [64.01742988773745]
An increasing privacy concern exists regarding training large-scale image segmentation models on unauthorized private data.
We exploit the concept of unlearnable examples to make images unusable to model training by generating and adding unlearnable noise into the original images.
We empirically verify the effectiveness of UnSeg across 6 mainstream image segmentation tasks, 10 widely used datasets, and 7 different network architectures.
arXiv Detail & Related papers (2024-10-13T16:34:46Z) - Are Deep Learning Models Robust to Partial Object Occlusion in Visual Recognition Tasks? [4.9260675787714]
Image classification models, including convolutional neural networks (CNNs), perform well on a variety of classification tasks but struggle under partial occlusion.
We contribute the Image Recognition Under Occlusion (IRUO) dataset, based on the recently developed Occluded Video Instance (IRUO) dataset (arXiv:2102.01558)
We find that modern CNN-based models show improved recognition accuracy on occluded images compared to earlier CNN-based models, and ViT-based models are more accurate than CNN-based models on occluded images.
arXiv Detail & Related papers (2024-09-16T23:21:22Z) - Reinforcing Pre-trained Models Using Counterfactual Images [54.26310919385808]
This paper proposes a novel framework to reinforce classification models using language-guided generated counterfactual images.
We identify model weaknesses by testing the model using the counterfactual image dataset.
We employ the counterfactual images as an augmented dataset to fine-tune and reinforce the classification model.
arXiv Detail & Related papers (2024-06-19T08:07:14Z) - Deep Image Composition Meets Image Forgery [0.0]
Image forgery has been studied for many years.
Deep learning models require large amounts of labeled data for training.
We use state of the art image composition deep learning models to generate spliced images close to the quality of real-life manipulations.
arXiv Detail & Related papers (2024-04-03T17:54:37Z) - OsmLocator: locating overlapping scatter marks with a non-training
generative perspective [48.50108853199417]
Locating overlapping marks faces many difficulties such as no texture, less contextual information, hallow shape and tiny size.
Here, we formulate it as a optimization problem on clustering-based re-visualization from a non-training generative perspective.
We especially built a dataset named 2023 containing hundreds of scatter images with different markers and various levels of overlapping severity, and tested the proposed method and compared it to existing methods.
arXiv Detail & Related papers (2023-12-18T12:39:48Z) - The Change You Want to See (Now in 3D) [65.61789642291636]
The goal of this paper is to detect what has changed, if anything, between two "in the wild" images of the same 3D scene.
We contribute a change detection model that is trained entirely on synthetic data and is class-agnostic.
We release a new evaluation dataset consisting of real-world image pairs with human-annotated differences.
arXiv Detail & Related papers (2023-08-21T01:59:45Z) - Effective Data Augmentation With Diffusion Models [65.09758931804478]
We address the lack of diversity in data augmentation with image-to-image transformations parameterized by pre-trained text-to-image diffusion models.
Our method edits images to change their semantics using an off-the-shelf diffusion model, and generalizes to novel visual concepts from a few labelled examples.
We evaluate our approach on few-shot image classification tasks, and on a real-world weed recognition task, and observe an improvement in accuracy in tested domains.
arXiv Detail & Related papers (2023-02-07T20:42:28Z) - A Generalist Framework for Panoptic Segmentation of Images and Videos [61.61453194912186]
We formulate panoptic segmentation as a discrete data generation problem, without relying on inductive bias of the task.
A diffusion model is proposed to model panoptic masks, with a simple architecture and generic loss function.
Our method is capable of modeling video (in a streaming setting) and thereby learns to track object instances automatically.
arXiv Detail & Related papers (2022-10-12T16:18:25Z) - Terrain Classification using Transfer Learning on Hyperspectral Images:
A Comparative study [0.13999481573773068]
convolutional neural network (CNN) and the Multi-Layer Perceptron (MLP) have been proven to be an effective method of image classification.
However, they suffer from the issues of long training time and requirement of large amounts of the labeled data.
We propose using the method of transfer learning to decrease the training time and reduce the dependence on large labeled dataset.
arXiv Detail & Related papers (2022-06-19T14:36:33Z) - A Hierarchical Transformation-Discriminating Generative Model for Few
Shot Anomaly Detection [93.38607559281601]
We devise a hierarchical generative model that captures the multi-scale patch distribution of each training image.
The anomaly score is obtained by aggregating the patch-based votes of the correct transformation across scales and image regions.
arXiv Detail & Related papers (2021-04-29T17:49:48Z) - Rethinking Natural Adversarial Examples for Classification Models [43.87819913022369]
ImageNet-A is a famous dataset of natural adversarial examples.
We validated the hypothesis by reducing the background influence in ImageNet-A examples with object detection techniques.
Experiments showed that the object detection models with various classification models as backbones obtained much higher accuracy than their corresponding classification models.
arXiv Detail & Related papers (2021-02-23T14:46:48Z) - Syn2Real: Forgery Classification via Unsupervised Domain Adaptation [1.8229783460536682]
We propose to create a synthetic forged dataset using deep semantic image inpainting and copy-move forgery algorithm.
We use unsupervised domain adaptation networks to detect copy-move forgery in new domains by mapping the feature space from our synthetically generated dataset.
arXiv Detail & Related papers (2020-02-03T15:02:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.