Efficient Palm-Line Segmentation with U-Net Context Fusion Module
- URL: http://arxiv.org/abs/2102.12127v1
- Date: Wed, 24 Feb 2021 08:42:52 GMT
- Title: Efficient Palm-Line Segmentation with U-Net Context Fusion Module
- Authors: Toan Pham Van, Son Trung Nguyen, Linh Bao Doan, Ngoc N. Tran and Ta
Minh Thanh
- Abstract summary: We propose an algorithm to extract principle palm lines from an image of a person's hand.
Our method applies deep learning networks (DNNs) to improve performance.
Based on the UNet segmentation neural network architecture, we propose a highly efficient architecture to detect palm-lines.
- Score: 1.1024591739346292
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Many cultures around the world believe that palm reading can be used to
predict the future life of a person. Palmistry uses features of the hand such
as palm lines, hand shape, or fingertip position. However, the research on
palm-line detection is still scarce, many of them applied traditional image
processing techniques. In most real-world scenarios, images usually are not in
well-conditioned, causing these methods to severely under-perform. In this
paper, we propose an algorithm to extract principle palm lines from an image of
a person's hand. Our method applies deep learning networks (DNNs) to improve
performance. Another challenge of this problem is the lack of training data. To
deal with this issue, we handcrafted a dataset from scratch. From this dataset,
we compare the performance of readily available methods with ours. Furthermore,
based on the UNet segmentation neural network architecture and the knowledge of
attention mechanism, we propose a highly efficient architecture to detect
palm-lines. We proposed the Context Fusion Module to capture the most important
context feature, which aims to improve segmentation accuracy. The experimental
results show that it outperforms the other methods with the highest F1 Score
about 99.42% and mIoU is 0.584 for the same dataset.
Related papers
- TrueDeep: A systematic approach of crack detection with less data [0.0]
We show that by incorporating domain knowledge along with deep learning architectures, we can achieve similar performance with less data.
Our algorithms, developed with 23% of the overall data, have a similar performance on the test data and significantly better performance on multiple blind datasets.
arXiv Detail & Related papers (2023-05-30T14:51:58Z) - Interacting Hand-Object Pose Estimation via Dense Mutual Attention [97.26400229871888]
3D hand-object pose estimation is the key to the success of many computer vision applications.
We propose a novel dense mutual attention mechanism that is able to model fine-grained dependencies between the hand and the object.
Our method is able to produce physically plausible poses with high quality and real-time inference speed.
arXiv Detail & Related papers (2022-11-16T10:01:33Z) - ALADIN: Distilling Fine-grained Alignment Scores for Efficient
Image-Text Matching and Retrieval [51.588385824875886]
Cross-modal retrieval consists in finding images related to a given query text or vice-versa.
Many recent methods proposed effective solutions to the image-text matching problem, mostly using recent large vision-language (VL) Transformer networks.
This paper proposes an ALign And DIstill Network (ALADIN) to fill in the gap between effectiveness and efficiency.
arXiv Detail & Related papers (2022-07-29T16:01:48Z) - Geometric Synthesis: A Free lunch for Large-scale Palmprint Recognition
Model Pretraining [27.81138870690135]
We introduce an intuitive geometric model which represents palmar creases with parameterized B'ezier curves.
By randomly sampling B'ezier parameters, we can synthesize massive training samples of diverse identities.
Experimental results demonstrate that such synthetically pretrained models have a very strong generalization ability.
arXiv Detail & Related papers (2022-03-11T01:20:22Z) - Learning Co-segmentation by Segment Swapping for Retrieval and Discovery [67.6609943904996]
The goal of this work is to efficiently identify visually similar patterns from a pair of images.
We generate synthetic training pairs by selecting object segments in an image and copy-pasting them into another image.
We show our approach provides clear improvements for artwork details retrieval on the Brueghel dataset.
arXiv Detail & Related papers (2021-10-29T16:51:16Z) - Understanding Egocentric Hand-Object Interactions from Hand Pose
Estimation [24.68535915849555]
We propose a method to label a dataset which contains the egocentric images pair-wisely.
We also use the collected pairwise data to train our encoder-decoder style network which has been proven efficient in.
arXiv Detail & Related papers (2021-09-29T18:34:06Z) - Revisiting Point Cloud Shape Classification with a Simple and Effective
Baseline [111.3236030935478]
We find that auxiliary factors like different evaluation schemes, data augmentation strategies, and loss functions make a large difference in performance.
A projection-based method, which we refer to as SimpleView, performs surprisingly well.
It achieves on par or better results than sophisticated state-of-the-art methods on ModelNet40 while being half the size of PointNet++.
arXiv Detail & Related papers (2021-06-09T18:01:11Z) - Group-Wise Semantic Mining for Weakly Supervised Semantic Segmentation [49.90178055521207]
This work addresses weakly supervised semantic segmentation (WSSS), with the goal of bridging the gap between image-level annotations and pixel-level segmentation.
We formulate WSSS as a novel group-wise learning task that explicitly models semantic dependencies in a group of images to estimate more reliable pseudo ground-truths.
In particular, we devise a graph neural network (GNN) for group-wise semantic mining, wherein input images are represented as graph nodes.
arXiv Detail & Related papers (2020-12-09T12:40:13Z) - MVHM: A Large-Scale Multi-View Hand Mesh Benchmark for Accurate 3D Hand
Pose Estimation [32.12879364117658]
Estimating 3D hand poses from a single RGB image is challenging because depth ambiguity leads the problem ill-posed.
We design a spin match algorithm that enables a rigid mesh model matching with any target mesh ground truth.
We present a multi-view hand pose estimation approach to verify that training a hand pose estimator with our generated dataset greatly enhances the performance.
arXiv Detail & Related papers (2020-12-06T07:55:08Z) - Palm-GAN: Generating Realistic Palmprint Images Using Total-Variation
Regularized GAN [7.119324327867636]
We present a deep learning framework based on generative adversarial networks (GAN), which is able to generate realistic palmprint images.
We apply this framework to a popular palmprint databases, and generate images which look very realistic, and similar to the samples in this database.
arXiv Detail & Related papers (2020-03-21T03:24:36Z) - Image Matching across Wide Baselines: From Paper to Practice [80.9424750998559]
We introduce a comprehensive benchmark for local features and robust estimation algorithms.
Our pipeline's modular structure allows easy integration, configuration, and combination of different methods.
We show that with proper settings, classical solutions may still outperform the perceived state of the art.
arXiv Detail & Related papers (2020-03-03T15:20:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.