A Convolutional Neural Network-based Patent Image Retrieval Method for
Design Ideation
- URL: http://arxiv.org/abs/2003.08741v3
- Date: Tue, 19 May 2020 22:58:48 GMT
- Title: A Convolutional Neural Network-based Patent Image Retrieval Method for
Design Ideation
- Authors: Shuo Jiang, Jianxi Luo, Guillermo Ruiz Pava, Jie Hu, Christopher L.
Magee
- Abstract summary: We propose a convolutional neural network (CNN)-based patent image retrieval method.
The core of this approach is a novel neural network architecture named Dual-VGG.
The accuracy of both training tasks and patent image embedding space are evaluated to show the performance of our model.
- Score: 5.195924252155368
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The patent database is often used in searches of inspirational stimuli for
innovative design opportunities because of its large size, extensive variety
and rich design information in patent documents. However, most patent mining
research only focuses on textual information and ignores visual information.
Herein, we propose a convolutional neural network (CNN)-based patent image
retrieval method. The core of this approach is a novel neural network
architecture named Dual-VGG that is aimed to accomplish two tasks: visual
material type prediction and international patent classification (IPC) class
label prediction. In turn, the trained neural network provides the deep
features in the image embedding vectors that can be utilized for patent image
retrieval and visual mapping. The accuracy of both training tasks and patent
image embedding space are evaluated to show the performance of our model. This
approach is also illustrated in a case study of robot arm design retrieval.
Compared to traditional keyword-based searching and Google image searching, the
proposed method discovers more useful visual information for engineering
design.
Related papers
- Early screening of potential breakthrough technologies with enhanced interpretability: A patent-specific hierarchical attention network model [4.779196219827507]
We propose an interpretable machine learning approach to predicting future citation counts from patent texts.
A case study of 35,376 pharmaceutical patents demonstrates the effectiveness of our approach.
It is expected that the proposed approach will enhance expert-machine collaboration in identifying breakthrough technologies.
arXiv Detail & Related papers (2024-07-24T02:17:10Z) - Exploring Explainability in Video Action Recognition [5.7782784592048575]
Video Action Recognition and Image Classification are foundational tasks in computer vision.
Video-TCAV aims to quantify the importance of specific concepts in the decision-making process of Video Action Recognition models.
We propose a machine-assisted approach to generate spatial andtemporal concepts relevant to Video Action Recognition for testing Video-TCAV.
arXiv Detail & Related papers (2024-04-13T19:34:14Z) - Unveiling Black-boxes: Explainable Deep Learning Models for Patent
Classification [48.5140223214582]
State-of-the-art methods for multi-label patent classification rely on deep opaque neural networks (DNNs)
We propose a novel deep explainable patent classification framework by introducing layer-wise relevance propagation (LRP)
Considering the relevance score, we then generate explanations by visualizing relevant words for the predicted patent class.
arXiv Detail & Related papers (2023-10-31T14:11:37Z) - Classification of Visualization Types and Perspectives in Patents [9.123089032348311]
We adopt state-of-the-art deep learning methods for the classification of visualization types and perspectives in patent images.
We derive a set of hierarchical classes from a dataset that provides weakly-labeled data for image perspectives.
arXiv Detail & Related papers (2023-07-19T21:45:07Z) - Review of Large Vision Models and Visual Prompt Engineering [50.63394642549947]
Review aims to summarize the methods employed in the computer vision domain for large vision models and visual prompt engineering.
We present influential large models in the visual domain and a range of prompt engineering methods employed on these models.
arXiv Detail & Related papers (2023-07-03T08:48:49Z) - Semantic Prompt for Few-Shot Image Recognition [76.68959583129335]
We propose a novel Semantic Prompt (SP) approach for few-shot learning.
The proposed approach achieves promising results, improving the 1-shot learning accuracy by 3.67% on average.
arXiv Detail & Related papers (2023-03-24T16:32:19Z) - Experimental Design for Multi-Channel Imaging via Task-Driven Feature Selection [2.662628670752034]
We present a new paradigm for experimental design that simultaneously optimize the design (set of image channels) and trains a machine-learning model to execute a user-specified image-analysis task.
Results show substantial improvement over classical experimental design, two recent application-specific methods within the new paradigm, and state-of-the-art approaches in supervised feature selection.
arXiv Detail & Related papers (2022-10-13T10:36:24Z) - Few-Shot Object Detection by Knowledge Distillation Using
Bag-of-Visual-Words Representations [58.48995335728938]
We design a novel knowledge distillation framework to guide the learning of the object detector.
We first present a novel Position-Aware Bag-of-Visual-Words model for learning a representative bag of visual words.
We then perform knowledge distillation based on the fact that an image should have consistent BoVW representations in two different feature spaces.
arXiv Detail & Related papers (2022-07-25T10:40:40Z) - Generative Design Ideation: A Natural Language Generation Approach [7.807713821263175]
This paper aims to explore a generative approach for knowledge-based design ideation by applying the latest pre-trained language models in artificial intelligence (AI)
The AI-generated ideas are not only in concise and understandable language but also able to synthesize the target design with external knowledge sources with controllable knowledge distance.
arXiv Detail & Related papers (2022-03-28T08:11:29Z) - NAS-DIP: Learning Deep Image Prior with Neural Architecture Search [65.79109790446257]
Recent work has shown that the structure of deep convolutional neural networks can be used as a structured image prior.
We propose to search for neural architectures that capture stronger image priors.
We search for an improved network by leveraging an existing neural architecture search algorithm.
arXiv Detail & Related papers (2020-08-26T17:59:36Z) - Ventral-Dorsal Neural Networks: Object Detection via Selective Attention [51.79577908317031]
We propose a new framework called Ventral-Dorsal Networks (VDNets)
Inspired by the structure of the human visual system, we propose the integration of a "Ventral Network" and a "Dorsal Network"
Our experimental results reveal that the proposed method outperforms state-of-the-art object detection approaches.
arXiv Detail & Related papers (2020-05-15T23:57:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.