Historical Astronomical Diagrams Decomposition in Geometric Primitives
- URL: http://arxiv.org/abs/2403.08721v1
- Date: Wed, 13 Mar 2024 17:20:25 GMT
- Title: Historical Astronomical Diagrams Decomposition in Geometric Primitives
- Authors: Syrine Kalleli, Scott Trigg, S\'egol\`ene Albouy, Mathieu Husson, and
Mathieu Aubry
- Abstract summary: We introduce a unique dataset of 303 astronomical diagrams from diverse traditions, ranging from the XIIth to the XVIIIth century.
We develop a model that builds on DINO-DETR to enable the prediction of multiple geometric primitives.
Our approach widely improves over the LETR baseline, which is restricted to lines, by introducing a meaningful parametrization for multiple primitives.
- Score: 13.447991818689463
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Automatically extracting the geometric content from the hundreds of thousands
of diagrams drawn in historical manuscripts would enable historians to study
the diffusion of astronomical knowledge on a global scale. However,
state-of-the-art vectorization methods, often designed to tackle modern data,
are not adapted to the complexity and diversity of historical astronomical
diagrams. Our contribution is thus twofold. First, we introduce a unique
dataset of 303 astronomical diagrams from diverse traditions, ranging from the
XIIth to the XVIIIth century, annotated with more than 3000 line segments,
circles and arcs. Second, we develop a model that builds on DINO-DETR to enable
the prediction of multiple geometric primitives. We show that it can be trained
solely on synthetic data and accurately predict primitives on our challenging
dataset. Our approach widely improves over the LETR baseline, which is
restricted to lines, by introducing a meaningful parametrization for multiple
primitives, jointly training for detection and parameter refinement, using
deformable attention and training on rich synthetic data. Our dataset and code
are available on our webpage.
Related papers
- EarthView: A Large Scale Remote Sensing Dataset for Self-Supervision [72.84868704100595]
This paper presents a dataset specifically designed for self-supervision on remote sensing data, intended to enhance deep learning applications on Earth monitoring tasks.
The dataset spans 15 tera pixels of global remote-sensing data, combining imagery from a diverse range of sources, including NEON, Sentinel, and a novel release of 1m spatial resolution data from Satellogic.
Accompanying the dataset is EarthMAE, a tailored Masked Autoencoder developed to tackle the distinct challenges of remote sensing data.
arXiv Detail & Related papers (2025-01-14T13:42:22Z) - AstroM$^3$: A self-supervised multimodal model for astronomy [0.0]
We propose AstroM$3$, a self-supervised pre-training approach that enables a model to learn from multiple modalities simultaneously.
Specifically, we extend the CLIP (Contrastive Language-Image Pretraining) model to a trimodal setting, allowing the integration of time-series photometry data, spectra, and astrophysical metadata.
Results demonstrate that CLIP pre-training improves classification performance for time-series photometry, where accuracy increases from 84.6% to 91.5%.
arXiv Detail & Related papers (2024-11-13T18:20:29Z) - Geometry-Aware Generative Autoencoders for Warped Riemannian Metric Learning and Generative Modeling on Data Manifolds [18.156807299614503]
We introduce Geometry-Aware Generative Autoencoder (GAGA), a novel framework that combines manifold learning with generative modeling.
GAGA shows competitive performance in simulated and real-world datasets, including a 30% improvement over the state-of-the-art methods in single-cell population-level trajectory inference.
arXiv Detail & Related papers (2024-10-16T17:53:26Z) - Dissecting embedding method: learning higher-order structures from data [0.0]
Geometric deep learning methods for data learning often include set of assumptions on the geometry of the feature space.
These assumptions together with data being discrete and finite can cause some generalisations, which are likely to create wrong interpretations of the data and models outputs.
arXiv Detail & Related papers (2024-10-14T08:19:39Z) - Diffusion Models as Data Mining Tools [87.77999285241219]
This paper demonstrates how to use generative models trained for image synthesis as tools for visual data mining.
We show that after finetuning conditional diffusion models to synthesize images from a specific dataset, we can use these models to define a typicality measure.
This measure assesses how typical visual elements are for different data labels, such as geographic location, time stamps, semantic labels, or even the presence of a disease.
arXiv Detail & Related papers (2024-07-20T17:14:31Z) - GeoWizard: Unleashing the Diffusion Priors for 3D Geometry Estimation from a Single Image [94.56927147492738]
We introduce GeoWizard, a new generative foundation model designed for estimating geometric attributes from single images.
We show that leveraging diffusion priors can markedly improve generalization, detail preservation, and efficiency in resource usage.
We propose a simple yet effective strategy to segregate the complex data distribution of various scenes into distinct sub-distributions.
arXiv Detail & Related papers (2024-03-18T17:50:41Z) - Exploring Data Geometry for Continual Learning [64.4358878435983]
We study continual learning from a novel perspective by exploring data geometry for the non-stationary stream of data.
Our method dynamically expands the geometry of the underlying space to match growing geometric structures induced by new data.
Experiments show that our method achieves better performance than baseline methods designed in Euclidean space.
arXiv Detail & Related papers (2023-04-08T06:35:25Z) - PolarMOT: How Far Can Geometric Relations Take Us in 3D Multi-Object
Tracking? [62.997667081978825]
We encode 3D detections as nodes in a graph, where spatial and temporal pairwise relations among objects are encoded via localized polar coordinates on graph edges.
This allows our graph neural network to learn to effectively encode temporal and spatial interactions.
We establish a new state-of-the-art on nuScenes dataset and, more importantly, show that our method, PolarMOT, generalizes remarkably well across different locations.
arXiv Detail & Related papers (2022-08-03T10:06:56Z) - First Full-Event Reconstruction from Imaging Atmospheric Cherenkov
Telescope Real Data with Deep Learning [55.41644538483948]
The Cherenkov Telescope Array is the future of ground-based gamma-ray astronomy.
Its first prototype telescope built on-site, the Large Size Telescope 1, is currently under commissioning and taking its first scientific data.
We present for the first time the development of a full-event reconstruction based on deep convolutional neural networks and its application to real data.
arXiv Detail & Related papers (2021-05-31T12:51:42Z) - Extendable and invertible manifold learning with geometry regularized
autoencoders [9.742277703732187]
A fundamental task in data exploration is to extract simplified low dimensional representations that capture intrinsic geometry in data.
Common approaches to this task use kernel methods for manifold learning.
We present a new method for integrating both approaches by incorporating a geometric regularization term in the bottleneck of the autoencoder.
arXiv Detail & Related papers (2020-07-14T15:59:10Z) - Extracting the main trend in a dataset: the Sequencer algorithm [0.0]
One-dimensional trends, often called sequences, provide insights into simple phenomena.
We present the Sequencer, an algorithm designed to generically identify the main trend in a dataset.
We show that, in a number of cases, it outperforms the popular t-SNE and UMAP dimensionality reduction techniques.
arXiv Detail & Related papers (2020-06-24T18:00:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.