DeepSperm: A robust and real-time bull sperm-cell detection in densely
populated semen videos
- URL: http://arxiv.org/abs/2003.01395v1
- Date: Tue, 3 Mar 2020 09:05:05 GMT
- Title: DeepSperm: A robust and real-time bull sperm-cell detection in densely
populated semen videos
- Authors: Priyanto Hidayatullah, Xueting Wang, Toshihiko Yamasaki, Tati L.E.R.
Mengko, Rinaldi Munir, Anggraini Barlian, Eros Sukmawati, Supraptono
Supraptono
- Abstract summary: This study proposes an architecture, called DeepSperm, that solves the challenges and is more accurate and faster than state-of-the-art architectures.
In our experiment, we achieve 86.91 mAP on the test dataset and a processing speed of 50.3 fps.
- Score: 26.494850349599528
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Background and Objective: Object detection is a primary research interest in
computer vision. Sperm-cell detection in a densely populated bull semen
microscopic observation video presents challenges such as partial occlusion,
vast number of objects in a single video frame, tiny size of the object,
artifacts, low contrast, and blurry objects because of the rapid movement of
the sperm cells. This study proposes an architecture, called DeepSperm, that
solves the aforementioned challenges and is more accurate and faster than
state-of-the-art architectures. Methods: In the proposed architecture, we use
only one detection layer, which is specific for small object detection. For
handling overfitting and increasing accuracy, we set a higher network
resolution, use a dropout layer, and perform data augmentation on hue,
saturation, and exposure. Several hyper-parameters are tuned to achieve better
performance. We compare our proposed method with those of a conventional image
processing-based object-detection method, you only look once (YOLOv3), and mask
region-based convolutional neural network (Mask R-CNN). Results: In our
experiment, we achieve 86.91 mAP on the test dataset and a processing speed of
50.3 fps. In comparison with YOLOv3, we achieve an increase of 16.66 mAP point,
3.26 x faster on testing, and 1.4 x faster on training with a small training
dataset, which contains 40 video frames. The weights file size was also reduced
significantly, with 16.94 x smaller than that of YOLOv3. Moreover, it requires
1.3 x less graphical processing unit (GPU) memory than YOLOv3. Conclusions:
This study proposes DeepSperm, which is a simple, effective, and efficient
architecture with its hyper-parameters and configuration to detect bull sperm
cells robustly in real time. In our experiment, we surpass the state of the art
in terms of accuracy, speed, and resource needs.
Related papers
- YOLO-MS: Rethinking Multi-Scale Representation Learning for Real-time
Object Detection [80.11152626362109]
We provide an efficient and performant object detector, termed YOLO-MS.
We train our YOLO-MS on the MS COCO dataset from scratch without relying on any other large-scale datasets.
Our work can also be used as a plug-and-play module for other YOLO models.
arXiv Detail & Related papers (2023-08-10T10:12:27Z) - Wearing face mask detection using deep learning through COVID-19
pandemic [0.0]
In this paper, we do an investigation on the capability of three state-of-the-art object detection neural networks on face mask detection for real-time applications.
According to the performance of different models, the best model that can be suitable for use in real-world and mobile device applications was the YOLOv4-tiny model.
arXiv Detail & Related papers (2023-04-28T19:39:32Z) - Pushing the Limits of Asynchronous Graph-based Object Detection with
Event Cameras [62.70541164894224]
We introduce several architecture choices which allow us to scale the depth and complexity of such models while maintaining low computation.
Our method runs 3.7 times faster than a dense graph neural network, taking only 8.4 ms per forward pass.
arXiv Detail & Related papers (2022-11-22T15:14:20Z) - SuperYOLO: Super Resolution Assisted Object Detection in Multimodal
Remote Sensing Imagery [36.216230299131404]
We propose SuperYOLO, which fuses multimodal data and performs high-resolution (HR) object detection on multiscale objects.
Our proposed model shows a favorable accuracy and speed tradeoff compared to the state-of-the-art models.
arXiv Detail & Related papers (2022-09-27T12:58:58Z) - SALISA: Saliency-based Input Sampling for Efficient Video Object
Detection [58.22508131162269]
We propose SALISA, a novel non-uniform SALiency-based Input SAmpling technique for video object detection.
We show that SALISA significantly improves the detection of small objects.
arXiv Detail & Related papers (2022-04-05T17:59:51Z) - A lightweight and accurate YOLO-like network for small target detection
in Aerial Imagery [94.78943497436492]
We present YOLO-S, a simple, fast and efficient network for small target detection.
YOLO-S exploits a small feature extractor based on Darknet20, as well as skip connection, via both bypass and concatenation.
YOLO-S has an 87% decrease of parameter size and almost one half FLOPs of YOLOv3, making practical the deployment for low-power industrial applications.
arXiv Detail & Related papers (2022-04-05T16:29:49Z) - EResFD: Rediscovery of the Effectiveness of Standard Convolution for
Lightweight Face Detection [13.357235715178584]
We re-examine the effectiveness of the standard convolutional block as a lightweight backbone architecture for face detection.
We show that heavily channel-pruned standard convolution layers can achieve better accuracy and inference speed.
Our proposed detector EResFD obtained 80.4% mAP on WIDER FACE Hard subset which only takes 37.7 ms for VGA image inference on CPU.
arXiv Detail & Related papers (2022-04-04T02:30:43Z) - EAutoDet: Efficient Architecture Search for Object Detection [110.99532343155073]
EAutoDet framework can discover practical backbone and FPN architectures for object detection in 1.4 GPU-days.
We propose a kernel reusing technique by sharing the weights of candidate operations on one edge and consolidating them into one convolution.
In particular, the discovered architectures surpass state-of-the-art object detection NAS methods and achieve 40.1 mAP with 120 FPS and 49.2 mAP with 41.3 FPS on COCO test-dev set.
arXiv Detail & Related papers (2022-03-21T05:56:12Z) - YOLObile: Real-Time Object Detection on Mobile Devices via
Compression-Compilation Co-Design [38.98949683262209]
We propose YOLObile framework, a real-time object detection on mobile devices via compression-compilation co-design.
A novel block-punched pruning scheme is proposed for any kernel size.
Under our YOLObile framework, we achieve 17 FPS inference speed using GPU on Samsung Galaxy S20.
arXiv Detail & Related papers (2020-09-12T01:41:08Z) - Expandable YOLO: 3D Object Detection from RGB-D Images [64.14512458954344]
This paper aims at constructing a light-weight object detector that inputs a depth and a color image from a stereo camera.
By extending the network architecture of YOLOv3 to 3D in the middle, it is possible to output in the depth direction.
Intersection over Uninon (IoU) in 3D space is introduced to confirm the accuracy of region extraction results.
arXiv Detail & Related papers (2020-06-26T07:32:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.