YOLIC: An Efficient Method for Object Localization and Classification on
Edge Devices
- URL: http://arxiv.org/abs/2307.06689v3
- Date: Sun, 30 Jul 2023 09:29:43 GMT
- Title: YOLIC: An Efficient Method for Object Localization and Classification on
Edge Devices
- Authors: Kai Su, Yoichi Tomioka, Qiangfu Zhao, Yong Liu
- Abstract summary: You Only Look at Interested Cells" (YOLIC) is an efficient method for object localization and classification on edge devices.
By adopting Cells of Interest for classification instead of individual pixels, YOLIC encapsulates relevant information, reduces computational load, and enables rough object shape inference.
This paper presents extensive experiments on multiple datasets to demonstrate that YOLIC achieves detection performance comparable to the state-of-the-art YOLO algorithms.
- Score: 10.058627390826967
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the realm of Tiny AI, we introduce ``You Only Look at Interested Cells"
(YOLIC), an efficient method for object localization and classification on edge
devices. Through seamlessly blending the strengths of semantic segmentation and
object detection, YOLIC offers superior computational efficiency and precision.
By adopting Cells of Interest for classification instead of individual pixels,
YOLIC encapsulates relevant information, reduces computational load, and
enables rough object shape inference. Importantly, the need for bounding box
regression is obviated, as YOLIC capitalizes on the predetermined cell
configuration that provides information about potential object location, size,
and shape. To tackle the issue of single-label classification limitations, a
multi-label classification approach is applied to each cell for effectively
recognizing overlapping or closely situated objects. This paper presents
extensive experiments on multiple datasets to demonstrate that YOLIC achieves
detection performance comparable to the state-of-the-art YOLO algorithms while
surpassing in speed, exceeding 30fps on a Raspberry Pi 4B CPU. All resources
related to this study, including datasets, cell designer, image annotation
tool, and source code, have been made publicly available on our project website
at https://kai3316.github.io/yolic.github.io
Related papers
- LRSAA: Large-scale Remote Sensing Image Target Recognition and Automatic Annotation [0.0]
This paper presents a method for object recognition and automatic labeling in large-area remote sensing images called LRSAA.
The method integrates YOLOv11 and MobileNetV3-SSD object detection algorithms through ensemble learning to enhance model performance.
arXiv Detail & Related papers (2024-11-24T12:30:12Z) - Large-scale Remote Sensing Image Target Recognition and Automatic Annotation [0.0]
This paper presents a method for object recognition and automatic labeling in large-area remote sensing images called LRSAA.
The method integrates YOLOv11 and MobileNetV3-SSD object detection algorithms through ensemble learning to enhance model performance.
arXiv Detail & Related papers (2024-11-12T13:57:13Z) - Unsupervised Learning of Object-Centric Embeddings for Cell Instance
Segmentation in Microscopy Images [3.039768384237206]
We introduce object-centric embeddings (OCEs)
OCEs embed image patches such that the offsets between patches cropped from the same object are preserved.
We show theoretically that OCEs can be learnt through a self-supervised task that predicts the spatial offset between image patches.
arXiv Detail & Related papers (2023-10-12T16:59:50Z) - Background Activation Suppression for Weakly Supervised Object
Localization and Semantic Segmentation [84.62067728093358]
Weakly supervised object localization and semantic segmentation aim to localize objects using only image-level labels.
New paradigm has emerged by generating a foreground prediction map to achieve pixel-level localization.
This paper presents two astonishing experimental observations on the object localization learning process.
arXiv Detail & Related papers (2023-09-22T15:44:10Z) - YUDO: YOLO for Uniform Directed Object Detection [0.0]
This paper presents an efficient way of detecting directed objects by predicting their center coordinates and direction angle.
Since the objects are of uniform size, the proposed model works without predicting the object's width and height.
arXiv Detail & Related papers (2023-08-08T19:18:20Z) - Weakly-supervised Contrastive Learning for Unsupervised Object Discovery [52.696041556640516]
Unsupervised object discovery is promising due to its ability to discover objects in a generic manner.
We design a semantic-guided self-supervised learning model to extract high-level semantic features from images.
We introduce Principal Component Analysis (PCA) to localize object regions.
arXiv Detail & Related papers (2023-07-07T04:03:48Z) - SupeRGB-D: Zero-shot Instance Segmentation in Cluttered Indoor
Environments [67.34330257205525]
In this work, we explore zero-shot instance segmentation (ZSIS) from RGB-D data to identify unseen objects in a semantic category-agnostic manner.
We present a method that uses annotated objects to learn the objectness'' of pixels and generalize to unseen object categories in cluttered indoor environments.
arXiv Detail & Related papers (2022-12-22T17:59:48Z) - An advanced YOLOv3 method for small object detection [2.906551456030129]
This paper introduces an improved YOLOv3 algorithm for small object detection.
In the proposed method, the dilated convolutions mish (DCM) module is introduced into the backbone network of YOLOv3.
In the neck network of YOLOv3, the convolutional block attention module (CBAM) and multi-level fusion module are introduced.
arXiv Detail & Related papers (2022-12-06T07:58:21Z) - VPIT: Real-time Embedded Single Object 3D Tracking Using Voxel Pseudo Images [90.60881721134656]
We propose a novel voxel-based 3D single object tracking (3D SOT) method called Voxel Pseudo Image Tracking (VPIT)
Experiments on KITTI Tracking dataset show that VPIT is the fastest 3D SOT method and maintains competitive Success and Precision values.
arXiv Detail & Related papers (2022-06-06T14:02:06Z) - ImpDet: Exploring Implicit Fields for 3D Object Detection [74.63774221984725]
We introduce a new perspective that views bounding box regression as an implicit function.
This leads to our proposed framework, termed Implicit Detection or ImpDet.
Our ImpDet assigns specific values to points in different local 3D spaces, thereby high-quality boundaries can be generated.
arXiv Detail & Related papers (2022-03-31T17:52:12Z) - Scale Normalized Image Pyramids with AutoFocus for Object Detection [75.71320993452372]
A scale normalized image pyramid (SNIP) is generated that, like human vision, only attends to objects within a fixed size range at different scales.
We propose an efficient spatial sub-sampling scheme which only operates on fixed-size sub-regions likely to contain objects.
The resulting algorithm is referred to as AutoFocus and results in a 2.5-5 times speed-up during inference when used with SNIP.
arXiv Detail & Related papers (2021-02-10T18:57:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.