DeepTag: A General Framework for Fiducial Marker Design and Detection
- URL: http://arxiv.org/abs/2105.13731v1
- Date: Fri, 28 May 2021 10:54:59 GMT
- Title: DeepTag: A General Framework for Fiducial Marker Design and Detection
- Authors: Zhuming Zhang, Yongtao Hu, Guoxing Yu, Jingwen Dai
- Abstract summary: We propose a general deep learning based framework, DeepTag, for fiducial marker design and detection.
DeepTag supports detection of a wide variety of existing marker families and makes it possible to design new marker families with customized local patterns.
Experiments show that DeepTag well supports different marker families and greatly outperforms the existing methods in terms of both detection robustness and pose accuracy.
- Score: 1.2180122937388957
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: A fiducial marker system usually consists of markers, a detection algorithm,
and a coding system. The appearance of markers and the detection robustness are
generally limited by the existing detection algorithms, which are hand-crafted
with traditional low-level image processing techniques. Furthermore, a
sophisticatedly designed coding system is required to overcome the shortcomings
of both markers and detection algorithms. To improve the flexibility and
robustness in various applications, we propose a general deep learning based
framework, DeepTag, for fiducial marker design and detection. DeepTag not only
supports detection of a wide variety of existing marker families, but also
makes it possible to design new marker families with customized local patterns.
Moreover, we propose an effective procedure to synthesize training data on the
fly without manual annotations. Thus, DeepTag can easily adapt to existing and
newly-designed marker families. To validate DeepTag and existing methods,
beside existing datasets, we further collect a new large and challenging
dataset where markers are placed in different view distances and angles.
Experiments show that DeepTag well supports different marker families and
greatly outperforms the existing methods in terms of both detection robustness
and pose accuracy. Both code and dataset are available at
\url{https://herohuyongtao.github.io/research/publications/deep-tag/}.
Related papers
- New Efficient Visual OILU Markers [0.5120567378386615]
We will exploit basic patterns to develop new efficient visual markers.
The proposed markers allow producing rich panel of unique identifiers.
The robustness of the markers against acquisition and geometric distortions is validated.
arXiv Detail & Related papers (2024-04-12T13:55:05Z) - Token-Specific Watermarking with Enhanced Detectability and Semantic Coherence for Large Language Models [31.062753031312006]
Large language models generate high-quality responses with potential misinformation.
Watermarking is pivotal in this context, which involves embedding hidden markers in texts.
We introduce a novel multi-objective optimization (MOO) approach for watermarking.
Our method simultaneously achieves detectability and semantic integrity.
arXiv Detail & Related papers (2024-02-28T05:43:22Z) - An Unforgeable Publicly Verifiable Watermark for Large Language Models [84.2805275589553]
Current watermark detection algorithms require the secret key used in the watermark generation process, making them susceptible to security breaches and counterfeiting during public detection.
We propose an unforgeable publicly verifiable watermark algorithm named UPV that uses two different neural networks for watermark generation and detection, instead of using the same key at both stages.
arXiv Detail & Related papers (2023-07-30T13:43:27Z) - The Treasure Beneath Multiple Annotations: An Uncertainty-aware Edge
Detector [70.43599299422813]
Existing methods fuse multiple annotations using a simple voting process, ignoring the inherent ambiguity of edges and labeling bias of annotators.
We propose a novel uncertainty-aware edge detector (UAED), which employs uncertainty to investigate the subjectivity and ambiguity of diverse annotations.
UAED achieves superior performance consistently across multiple edge detection benchmarks.
arXiv Detail & Related papers (2023-03-21T13:14:36Z) - DeepFormableTag: End-to-end Generation and Recognition of Deformable
Fiducial Markers [27.135078472097895]
Existing detection methods assume that markers are printed on ideally planar surfaces.
A fiducial marker generator creates a set of free-form color patterns to encode significantly large-scale information.
A differentiable image simulator creates a training dataset of photorealistic scene images with the deformed markers.
A trained marker detector seeks the regions of interest and recognizes multiple marker patterns simultaneously.
arXiv Detail & Related papers (2022-06-16T09:29:26Z) - Application of Ghost-DeblurGAN to Fiducial Marker Detection [1.1470070927586016]
This paper develops a lightweight generative adversarial network, named Ghost-DeGAN, for real-time motion deblurring.
A new large-scale dataset, YorkTag, is proposed that provides pairs of sharp/blurred images containing fiducial markers.
With the proposed model trained and tested on YorkTag, it is demonstrated that when applied along with fiducial marker systems to motion-blurred images, Ghost-DeblurGAN improves the marker detection significantly.
arXiv Detail & Related papers (2021-09-08T00:59:10Z) - SOLD2: Self-supervised Occlusion-aware Line Description and Detection [95.8719432775724]
We introduce the first joint detection and description of line segments in a single deep network.
Our method does not require any annotated line labels and can therefore generalize to any dataset.
We evaluate our approach against previous line detection and description methods on several multi-view datasets.
arXiv Detail & Related papers (2021-04-07T19:27:17Z) - Dense Label Encoding for Boundary Discontinuity Free Rotation Detection [69.75559390700887]
This paper explores a relatively less-studied methodology based on classification.
We propose new techniques to push its frontier in two aspects.
Experiments and visual analysis on large-scale public datasets for aerial images show the effectiveness of our approach.
arXiv Detail & Related papers (2020-11-19T05:42:02Z) - Deep Soft Procrustes for Markerless Volumetric Sensor Alignment [81.13055566952221]
In this work, we improve markerless data-driven correspondence estimation to achieve more robust multi-sensor spatial alignment.
We incorporate geometric constraints in an end-to-end manner into a typical segmentation based model and bridge the intermediate dense classification task with the targeted pose estimation one.
Our model is experimentally shown to achieve similar results with marker-based methods and outperform the markerless ones, while also being robust to the pose variations of the calibration structure.
arXiv Detail & Related papers (2020-03-23T10:51:32Z) - Weakly-Supervised Salient Object Detection via Scribble Annotations [54.40518383782725]
We propose a weakly-supervised salient object detection model to learn saliency from scribble labels.
We present a new metric, termed saliency structure measure, to measure the structure alignment of the predicted saliency maps.
Our method not only outperforms existing weakly-supervised/unsupervised methods, but also is on par with several fully-supervised state-of-the-art models.
arXiv Detail & Related papers (2020-03-17T12:59:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.