APEX-Net: Automatic Plot Extractor Network
- URL: http://arxiv.org/abs/2101.06217v3
- Date: Thu, 11 Feb 2021 05:03:01 GMT
- Title: APEX-Net: Automatic Plot Extractor Network
- Authors: Aalok Gangopadhyay, Prajwal Singh, Shanmuganathan Raman
- Abstract summary: We propose APEX-Net, a deep learning based framework with novel loss functions for solving the plot extraction problem.
We introduce APEX-1M, a new large scale dataset which contains both the plot images and the raw data.
We show visual results of our network on unseen plot images and demonstrate that it extracts the shape of the plots to a great extent.
- Score: 24.299931323012757
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Automatic extraction of raw data from 2D line plot images is a problem of
great importance having many real-world applications. Several algorithms have
been proposed for solving this problem. However, these algorithms involve a
significant amount of human intervention. To minimize this intervention, we
propose APEX-Net, a deep learning based framework with novel loss functions for
solving the plot extraction problem. We introduce APEX-1M, a new large scale
dataset which contains both the plot images and the raw data. We demonstrate
the performance of APEX-Net on the APEX-1M test set and show that it obtains
impressive accuracy. We also show visual results of our network on unseen plot
images and demonstrate that it extracts the shape of the plots to a great
extent. Finally, we develop a GUI based software for plot extraction that can
benefit the community at large. For dataset and more information visit
https://sites.google.com/view/apexnetpaper/.
Related papers
- Leveraging Vision Capabilities of Multimodal LLMs for Automated Data Extraction from Plots [0.0]
We show that current large language models, with proper instructions and engineered prompts, are capable of accurately extracting data from plots.
This capability is inherent to the pretrained models and can be achieved with a chain-of-thought sequence of zero-shot engineered prompts.
arXiv Detail & Related papers (2025-03-16T02:41:43Z) - Depth Anything: Unleashing the Power of Large-Scale Unlabeled Data [87.61900472933523]
This work presents Depth Anything, a highly practical solution for robust monocular depth estimation.
We scale up the dataset by designing a data engine to collect and automatically annotate large-scale unlabeled data.
We evaluate its zero-shot capabilities extensively, including six public datasets and randomly captured photos.
arXiv Detail & Related papers (2024-01-19T18:59:52Z) - Learnable Graph Matching: A Practical Paradigm for Data Association [74.28753343714858]
We propose a general learnable graph matching method to address these issues.
Our method achieves state-of-the-art performance on several MOT datasets.
For image matching, our method outperforms state-of-the-art methods on a popular indoor dataset, ScanNet.
arXiv Detail & Related papers (2023-03-27T17:39:00Z) - I2P-Rec: Recognizing Images on Large-scale Point Cloud Maps through
Bird's Eye View Projections [18.7557037030769]
Place recognition is an important technique for autonomous cars to achieve full autonomy.
We propose the I2P-Rec method to solve the problem by transforming the cross-modal data into the same modality.
With only a small set of training data, I2P-Rec achieves recall rates at Top-1% over 80% and 90%, when localizing monocular and stereo images on point cloud maps.
arXiv Detail & Related papers (2023-03-02T07:56:04Z) - Supervised and Contrastive Self-Supervised In-Domain Representation
Learning for Dense Prediction Problems in Remote Sensing [0.0]
This paper explores the effectiveness of in-domain representations in both supervised and self-supervised forms to solve the domain difference between remote sensing and the ImageNet dataset.
For self-supervised pre-training, we have utilized the SimSiam algorithm as it is simple and does not need huge computational resources.
Our results have demonstrated that using datasets with a high spatial resolution for self-supervised representation learning leads to high performance in downstream tasks.
arXiv Detail & Related papers (2023-01-29T20:56:51Z) - Let's Enhance: A Deep Learning Approach to Extreme Deblurring of Text
Images [3.441021278275805]
This work presents a novel deep-learning-based pipeline for the inverse problem of image deblurring.
Our results build on our winning submission to the recent Helsinki Deblurring Challenge 2021, whose goal was to explore the limits of state-of-the-art deblurring algorithms.
arXiv Detail & Related papers (2022-11-18T09:06:56Z) - Large-scale Unsupervised Semantic Segmentation [163.3568726730319]
We propose a new problem of large-scale unsupervised semantic segmentation (LUSS) with a newly created benchmark dataset to track the research progress.
Based on the ImageNet dataset, we propose the ImageNet-S dataset with 1.2 million training images and 40k high-quality semantic segmentation annotations for evaluation.
arXiv Detail & Related papers (2021-06-06T15:02:11Z) - Plot2API: Recommending Graphic API from Plot via Semantic Parsing Guided
Neural Network [13.936788648883068]
Plot-based Graphic API recommendation (Plot2API) is an unstudied but meaningful issue.
We present a novel deep multi-task learning approach named Semantic Parsing Guided Neural Network (SPGNN)
In SPGNN, the recently advanced Convolutional Neural Network (CNN) named EfficientNet is employed as the backbone network for API recommendation.
arXiv Detail & Related papers (2021-04-02T13:08:56Z) - MAGNeto: An Efficient Deep Learning Method for the Extractive Tags
Summarization Problem [0.0]
We study a new image annotation task named Extractive Tags Summarization (ETS)
The goal is to extract important tags from the context lying in an image and its corresponding tags.
Our proposed solution consists of different widely used blocks like convolutional and self-attention layers.
arXiv Detail & Related papers (2020-11-09T11:34:21Z) - Adaptive Context-Aware Multi-Modal Network for Depth Completion [107.15344488719322]
We propose to adopt the graph propagation to capture the observed spatial contexts.
We then apply the attention mechanism on the propagation, which encourages the network to model the contextual information adaptively.
Finally, we introduce the symmetric gated fusion strategy to exploit the extracted multi-modal features effectively.
Our model, named Adaptive Context-Aware Multi-Modal Network (ACMNet), achieves the state-of-the-art performance on two benchmarks.
arXiv Detail & Related papers (2020-08-25T06:00:06Z) - Scaling Graph Neural Networks with Approximate PageRank [64.92311737049054]
We present the PPRGo model which utilizes an efficient approximation of information diffusion in GNNs.
In addition to being faster, PPRGo is inherently scalable, and can be trivially parallelized for large datasets like those found in industry settings.
We show that training PPRGo and predicting labels for all nodes in this graph takes under 2 minutes on a single machine, far outpacing other baselines on the same graph.
arXiv Detail & Related papers (2020-07-03T09:30:07Z) - JHU-CROWD++: Large-Scale Crowd Counting Dataset and A Benchmark Method [92.15895515035795]
We introduce a new large scale unconstrained crowd counting dataset (JHU-CROWD++) that contains "4,372" images with "1.51 million" annotations.
We propose a novel crowd counting network that progressively generates crowd density maps via residual error estimation.
arXiv Detail & Related papers (2020-04-07T14:59:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.