Quantification and Classification of Carbon Nanotubes in Electron Micrographs using Vision Foundation Models
- URL: http://arxiv.org/abs/2601.06673v1
- Date: Sat, 10 Jan 2026 20:22:58 GMT
- Title: Quantification and Classification of Carbon Nanotubes in Electron Micrographs using Vision Foundation Models
- Authors: Sanjay Pradeep, Chen Wang, Matthew M. Dahm, Jeff D. Eldredge, Candace S. J. Tsai,
- Abstract summary: This work presents a unified framework leveraging vision foundation models to automate the quantification and classification of CNTs in electron microscopy images.<n>We introduce an interactive quantification tool built on the Segment Anything Model (SAM) that segments particles with near-perfect accuracy using minimal user input.<n>Second, we propose a novel classification pipeline that utilizes these segmentation masks to spatially constrain a DINOv2 vision transformer, extracting features exclusively from particle regions while suppressing background noise.<n> Evaluated on a dataset of 1,800 TEM images, this architecture achieves 95.5% accuracy in distinguishing between four different CNT morphologies, significantly outperforming the current baseline despite using a
- Score: 1.8969168959157112
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Accurate characterization of carbon nanotube morphologies in electron microscopy images is vital for exposure assessment and toxicological studies, yet current workflows rely on slow, subjective manual segmentation. This work presents a unified framework leveraging vision foundation models to automate the quantification and classification of CNTs in electron microscopy images. First, we introduce an interactive quantification tool built on the Segment Anything Model (SAM) that segments particles with near-perfect accuracy using minimal user input. Second, we propose a novel classification pipeline that utilizes these segmentation masks to spatially constrain a DINOv2 vision transformer, extracting features exclusively from particle regions while suppressing background noise. Evaluated on a dataset of 1,800 TEM images, this architecture achieves 95.5% accuracy in distinguishing between four different CNT morphologies, significantly outperforming the current baseline despite using a fraction of the training data. Crucially, this instance-level processing allows the framework to resolve mixed samples, correctly classifying distinct particle types co-existing within a single field of view. These results demonstrate that integrating zero-shot segmentation with self-supervised feature learning enables high-throughput, reproducible nanomaterial analysis, transforming a labor-intensive bottleneck into a scalable, data-driven process.
Related papers
- Zero-shot Shape Classification of Nanoparticles in SEM Images using Vision Foundation Models [0.9466841964978984]
Conventional deep learning methods for shape classification require extensive labeled datasets and computationally demanding training.<n>In this study, we introduce a zero-shot classification pipeline that leverages two vision foundation models.<n>We achieve high-precision shape classification across three morphologically diverse nanoparticle datasets.
arXiv Detail & Related papers (2025-08-05T09:03:56Z) - PathSegDiff: Pathology Segmentation using Diffusion model representations [63.20694440934692]
We propose PathSegDiff, a novel approach for histopathology image segmentation that leverages Latent Diffusion Models (LDMs) as pre-trained featured extractors.<n>Our method utilizes a pathology-specific LDM, guided by a self-supervised encoder, to extract rich semantic information from H&E stained histopathology images.<n>Our experiments demonstrate significant improvements over traditional methods on the BCSS and GlaS datasets.
arXiv Detail & Related papers (2025-04-09T14:58:21Z) - Machine Learning for Identifying Grain Boundaries in Scanning Electron Microscopy (SEM) Images of Nanoparticle Superlattices [0.0]
We present a machine learning workflow for automating grain segmentation in scanning electron microscopy (SEM) images of nanoparticles superlattices.<n>We transform the raw pixel data into explainable numerical representation of superlattice orientations for clustering.<n>This efficiency makes the workflow scalable to large datasets and makes it a valuable tool for integrating data-driven models into decision-making processes.
arXiv Detail & Related papers (2025-01-07T22:51:10Z) - Automated Grain Boundary (GB) Segmentation and Microstructural Analysis
in 347H Stainless Steel Using Deep Learning and Multimodal Microscopy [2.0445155106382797]
Austenitic 347H stainless steel offers superior mechanical properties and corrosion resistance required for extreme operating conditions.
CNN based deep-learning models is a powerful technique to detect features from material micrographs in an automated manner.
We combine scanning electron microscopy (SEM) images of 347H stainless steel as training data and electron backscatter diffraction (EBSD) micrographs as pixel-wise labels for grain boundary detection.
arXiv Detail & Related papers (2023-05-12T22:49:36Z) - Optimizations of Autoencoders for Analysis and Classification of
Microscopic In Situ Hybridization Images [68.8204255655161]
We propose a deep-learning framework to detect and classify areas of microscopic images with similar levels of gene expression.
The data we analyze requires an unsupervised learning model for which we employ a type of Artificial Neural Network - Deep Learning Autoencoders.
arXiv Detail & Related papers (2023-04-19T13:45:28Z) - Interpretable Joint Event-Particle Reconstruction for Neutrino Physics
at NOvA with Sparse CNNs and Transformers [124.29621071934693]
We present a novel neural network architecture that combines the spatial learning enabled by convolutions with the contextual learning enabled by attention.
TransformerCVN simultaneously classifies each event and reconstructs every individual particle's identity.
This architecture enables us to perform several interpretability studies which provide insights into the network's predictions.
arXiv Detail & Related papers (2023-03-10T20:36:23Z) - AMIGO: Sparse Multi-Modal Graph Transformer with Shared-Context
Processing for Representation Learning of Giga-pixel Images [53.29794593104923]
We present a novel concept of shared-context processing for whole slide histopathology images.
AMIGO uses the celluar graph within the tissue to provide a single representation for a patient.
We show that our model is strongly robust to missing information to an extent that it can achieve the same performance with as low as 20% of the data.
arXiv Detail & Related papers (2023-03-01T23:37:45Z) - A kinetic approach to consensus-based segmentation of biomedical images [39.58317527488534]
We apply a kinetic version of a bounded confidence consensus model to biomedical segmentation problems.
The large time behavior of the system is then computed with the aid of a surrogate Fokker-Planck approach.
We minimize the introduced segmentation metric for a relevant set of 2D gray-scale images.
arXiv Detail & Related papers (2022-11-08T09:54:34Z) - The Preliminary Results on Analysis of TAIGA-IACT Images Using
Convolutional Neural Networks [68.8204255655161]
The aim of the work is to study the possibility of the machine learning application to solve the tasks set for TAIGA-IACT.
The method of Convolutional Neural Networks (CNN) was applied to process and analyze Monte-Carlo events simulated with CORSIKA.
arXiv Detail & Related papers (2021-12-19T15:17:20Z) - Towards an Automatic Analysis of CHO-K1 Suspension Growth in
Microfluidic Single-cell Cultivation [63.94623495501023]
We propose a novel Machine Learning architecture, which allows us to infuse a neural deep network with human-powered abstraction on the level of data.
Specifically, we train a generative model simultaneously on natural and synthetic data, so that it learns a shared representation, from which a target variable, such as the cell count, can be reliably estimated.
arXiv Detail & Related papers (2020-10-20T08:36:51Z) - Improving the Segmentation of Scanning Probe Microscope Images using
Convolutional Neural Networks [0.9236074230806579]
We develop protocols for the segmentation of images of 2D assemblies of gold nanoparticles formed on silicon surfaces via deposition from an organic solvent.
The evaporation of the solvent drives far-from-equilibrium self-organisation of the particles, producing a wide variety of nano- and micro-structured patterns.
We show that a segmentation strategy using the U-Net convolutional neural network outperforms traditional automated approaches.
arXiv Detail & Related papers (2020-08-27T20:49:59Z) - Machine Learning Pipeline for Segmentation and Defect Identification
from High Resolution Transmission Electron Microscopy Data [0.0]
We demonstrate a flexible two step pipeline for analysis of high resolution transmission electron microscopy data.
Our trained U-Net is able to segment nanoparticles from amorphous background with a Dice coefficient of 0.8.
We are then able to classify whether nanoparticles contain a visible stacking fault with 86% accuracy.
arXiv Detail & Related papers (2020-01-14T19:49:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.