Interactive Segmentation and Visualization for Tiny Objects in
Multi-megapixel Images
- URL: http://arxiv.org/abs/2204.10356v1
- Date: Thu, 21 Apr 2022 18:26:48 GMT
- Title: Interactive Segmentation and Visualization for Tiny Objects in
Multi-megapixel Images
- Authors: Chengyuan Xu, Boning Dong, Noah Stier, Curtis McCully, D. Andrew
Howell, Pradeep Sen, Tobias H\"ollerer
- Abstract summary: We introduce an interactive image segmentation and visualization framework for identifying, inspecting, and editing tiny objects in large multi-megapixel high-range images.
We developed an interactive toolkit that unifies inference model, HDR image visualization, segmentation mask inspection and editing into a single graphical user interface.
Our interface features mouse-controlled, synchronized, dual-window visualization of the image and the segmentation mask, a critical feature for locating tiny objects in multi-megapixel images.
- Score: 5.09193568605539
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce an interactive image segmentation and visualization framework
for identifying, inspecting, and editing tiny objects (just a few pixels wide)
in large multi-megapixel high-dynamic-range (HDR) images. Detecting cosmic rays
(CRs) in astronomical observations is a cumbersome workflow that requires
multiple tools, so we developed an interactive toolkit that unifies model
inference, HDR image visualization, segmentation mask inspection and editing
into a single graphical user interface. The feature set, initially designed for
astronomical data, makes this work a useful research-supporting tool for
human-in-the-loop tiny-object segmentation in scientific areas like
biomedicine, materials science, remote sensing, etc., as well as computer
vision. Our interface features mouse-controlled, synchronized, dual-window
visualization of the image and the segmentation mask, a critical feature for
locating tiny objects in multi-megapixel images. The browser-based tool can be
readily hosted on the web to provide multi-user access and GPU acceleration for
any device. The toolkit can also be used as a high-precision annotation tool,
or adapted as the frontend for an interactive machine learning framework. Our
open-source dataset, CR detection model, and visualization toolkit are
available at https://github.com/cy-xu/cosmic-conn.
Related papers
- Grad-CAMO: Learning Interpretable Single-Cell Morphological Profiles from 3D Cell Painting Images [0.0]
We introduce Grad-CAMO, a novel single-cell interpretability score for supervised feature extractors.
Grad-CAMO measures the proportion of a model's attention that is concentrated on the cell of interest versus the background.
arXiv Detail & Related papers (2024-03-26T11:48:37Z) - TIAViz: A Browser-based Visualization Tool for Computational Pathology
Models [0.6519788717471032]
We introduce TIAViz, a Python-based visualization tool built into TIAToolbox.
It allows flexible, interactive, fully zoomable overlay of a wide variety of information onto whole slide images.
arXiv Detail & Related papers (2024-02-15T14:54:46Z) - SeeBel: Seeing is Believing [0.9790236766474201]
We propose three visualizations that enable users to compare dataset statistics and AI performance for segmenting all images.
Our project tries to further increase the interpretability of the trained AI model for segmentation by visualizing its image attention weights.
We propose to conduct surveys on real users to study the efficacy of our visualization tool in computer vision and AI domain.
arXiv Detail & Related papers (2023-12-18T05:11:00Z) - Improving Human-Object Interaction Detection via Virtual Image Learning [68.56682347374422]
Human-Object Interaction (HOI) detection aims to understand the interactions between humans and objects.
In this paper, we propose to alleviate the impact of such an unbalanced distribution via Virtual Image Leaning (VIL)
A novel label-to-image approach, Multiple Steps Image Creation (MUSIC), is proposed to create a high-quality dataset that has a consistent distribution with real images.
arXiv Detail & Related papers (2023-08-04T10:28:48Z) - Detector Guidance for Multi-Object Text-to-Image Generation [61.70018793720616]
Detector Guidance (DG) integrates a latent object detection model to separate different objects during the generation process.
Human evaluations demonstrate that DG provides an 8-22% advantage in preventing the amalgamation of conflicting concepts.
arXiv Detail & Related papers (2023-06-04T02:33:12Z) - Interactive Visual Feature Search [8.255656003475268]
We introduce Visual Feature Search, a novel interactive visualization that is adaptable to any CNN.
Our tool allows a user to highlight an image region and search for images from a given dataset with the most similar model features.
We demonstrate how our tool elucidates different aspects of model behavior by performing experiments on a range of applications, such as in medical imaging and wildlife classification.
arXiv Detail & Related papers (2022-11-28T04:39:03Z) - A Simple Baseline for Multi-Camera 3D Object Detection [94.63944826540491]
3D object detection with surrounding cameras has been a promising direction for autonomous driving.
We present SimMOD, a Simple baseline for Multi-camera Object Detection.
We conduct extensive experiments on the 3D object detection benchmark of nuScenes to demonstrate the effectiveness of SimMOD.
arXiv Detail & Related papers (2022-08-22T03:38:01Z) - Vision Transformer with Convolutions Architecture Search [72.70461709267497]
We propose an architecture search method-Vision Transformer with Convolutions Architecture Search (VTCAS)
The high-performance backbone network searched by VTCAS introduces the desirable features of convolutional neural networks into the Transformer architecture.
It enhances the robustness of the neural network for object recognition, especially in the low illumination indoor scene.
arXiv Detail & Related papers (2022-03-20T02:59:51Z) - Multi-modal Transformers Excel at Class-agnostic Object Detection [105.10403103027306]
We argue that existing methods lack a top-down supervision signal governed by human-understandable semantics.
We develop an efficient and flexible MViT architecture using multi-scale feature processing and deformable self-attention.
We show the significance of MViT proposals in a diverse range of applications.
arXiv Detail & Related papers (2021-11-22T18:59:29Z) - Embedding-based Instance Segmentation of Microscopy Images [8.516639438995785]
We introduce EmbedSeg, an end-to-end trainable deep learning method based on the work by Neven et al.
While their approach embeds each pixel to the centroid of any given instance, in EmbedSeg, motivated by the complex shapes of biological objects, we propose to use the medoid instead.
We demonstrate that embedding-based instance segmentation achieves competitive results in comparison to state-of-the-art methods on diverse microscopy datasets.
arXiv Detail & Related papers (2021-01-25T12:06:44Z) - Learning RGB-D Feature Embeddings for Unseen Object Instance
Segmentation [67.88276573341734]
We propose a new method for unseen object instance segmentation by learning RGB-D feature embeddings from synthetic data.
A metric learning loss function is utilized to learn to produce pixel-wise feature embeddings.
We further improve the segmentation accuracy with a new two-stage clustering algorithm.
arXiv Detail & Related papers (2020-07-30T00:23:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.