Cross-Model Image Annotation Platform with Active Learning
- URL: http://arxiv.org/abs/2008.02421v1
- Date: Thu, 6 Aug 2020 01:42:25 GMT
- Title: Cross-Model Image Annotation Platform with Active Learning
- Authors: Ng Hui Xian Lynnette, Henry Ng Siong Hock, Nguwi Yok Yen
- Abstract summary: This work presents an End-to-End pipeline tool for object annotation and recognition.
We have developed a modular image annotation platform which seamlessly incorporates assisted image annotation, active learning and model training and evaluation.
The highest accuracy achieved is 74%.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We have seen significant leapfrog advancement in machine learning in recent
decades. The central idea of machine learnability lies on constructing learning
algorithms that learn from good data. The availability of more data being made
publicly available also accelerates the growth of AI in recent years. In the
domain of computer vision, the quality of image data arises from the accuracy
of image annotation. Labeling large volume of image data is a daunting and
tedious task. This work presents an End-to-End pipeline tool for object
annotation and recognition aims at enabling quick image labeling. We have
developed a modular image annotation platform which seamlessly incorporates
assisted image annotation (annotation assistance), active learning and model
training and evaluation. Our approach provides a number of advantages over
current image annotation tools. Firstly, the annotation assistance utilizes
reference hierarchy and reference images to locate the objects in the images,
thus reducing the need for annotating the whole object. Secondly, images can be
annotated using polygon points allowing for objects of any shape to be
annotated. Thirdly, it is also interoperable across several image models, and
the tool provides an interface for object model training and evaluation across
a series of pre-trained models. We have tested the model and embeds several
benchmarking deep learning models. The highest accuracy achieved is 74%.
Related papers
- Helping Hands: An Object-Aware Ego-Centric Video Recognition Model [60.350851196619296]
We introduce an object-aware decoder for improving the performance of ego-centric representations on ego-centric videos.
We show that the model can act as a drop-in replacement for an ego-awareness video model to improve performance through visual-text grounding.
arXiv Detail & Related papers (2023-08-15T17:58:11Z) - DINOv2: Learning Robust Visual Features without Supervision [75.42921276202522]
This work shows that existing pretraining methods, especially self-supervised methods, can produce such features if trained on enough curated data from diverse sources.
Most of the technical contributions aim at accelerating and stabilizing the training at scale.
In terms of data, we propose an automatic pipeline to build a dedicated, diverse, and curated image dataset instead of uncurated data, as typically done in the self-supervised literature.
arXiv Detail & Related papers (2023-04-14T15:12:19Z) - MegaPose: 6D Pose Estimation of Novel Objects via Render & Compare [84.80956484848505]
MegaPose is a method to estimate the 6D pose of novel objects, that is, objects unseen during training.
We present a 6D pose refiner based on a render&compare strategy which can be applied to novel objects.
Second, we introduce a novel approach for coarse pose estimation which leverages a network trained to classify whether the pose error between a synthetic rendering and an observed image of the same object can be corrected by the refiner.
arXiv Detail & Related papers (2022-12-13T19:30:03Z) - EISeg: An Efficient Interactive Segmentation Tool based on PaddlePaddle [7.588694189597639]
We introduce EISeg, an Efficient Interactive SEGmentation annotation tool that can drastically improve image segmentation annotation efficiency.
We also provide various domain-specific models for remote sensing, medical imaging, industrial quality inspections, human segmentation, and temporal aware models for video segmentation.
arXiv Detail & Related papers (2022-10-17T07:12:13Z) - Robustar: Interactive Toolbox Supporting Precise Data Annotation for
Robust Vision Learning [53.900911121695536]
We introduce the initial release of our software Robustar.
It aims to improve the robustness of vision classification machine learning models through a data-driven perspective.
arXiv Detail & Related papers (2022-07-18T21:12:28Z) - Vision Models Are More Robust And Fair When Pretrained On Uncurated
Images Without Supervision [38.22842778742829]
Discriminative self-supervised learning allows training models on any random group of internet images.
We train models on billions of random images without any data pre-processing or prior assumptions about what we want the model to learn.
We extensively study and validate our model performance on over 50 benchmarks including fairness, to distribution shift, geographical diversity, fine grained recognition, image copy detection and many image classification datasets.
arXiv Detail & Related papers (2022-02-16T22:26:47Z) - AugNet: End-to-End Unsupervised Visual Representation Learning with
Image Augmentation [3.6790362352712873]
We propose AugNet, a new deep learning training paradigm to learn image features from a collection of unlabeled pictures.
Our experiments demonstrate that the method is able to represent the image in low dimensional space.
Unlike many deep-learning-based image retrieval algorithms, our approach does not require access to external annotated datasets.
arXiv Detail & Related papers (2021-06-11T09:02:30Z) - Towards Good Practices for Efficiently Annotating Large-Scale Image
Classification Datasets [90.61266099147053]
We investigate efficient annotation strategies for collecting multi-class classification labels for a large collection of images.
We propose modifications and best practices aimed at minimizing human labeling effort.
Simulated experiments on a 125k image subset of the ImageNet100 show that it can be annotated to 80% top-1 accuracy with 0.35 annotations per image on average.
arXiv Detail & Related papers (2021-04-26T16:29:32Z) - Data Augmentation for Object Detection via Differentiable Neural
Rendering [71.00447761415388]
It is challenging to train a robust object detector when annotated data is scarce.
Existing approaches to tackle this problem include semi-supervised learning that interpolates labeled data from unlabeled data.
We introduce an offline data augmentation method for object detection, which semantically interpolates the training data with novel views.
arXiv Detail & Related papers (2021-03-04T06:31:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.