Frontiers of Deep Learning: From Novel Application to Real-World Deployment
- URL: http://arxiv.org/abs/2407.14386v1
- Date: Fri, 19 Jul 2024 15:11:55 GMT
- Title: Frontiers of Deep Learning: From Novel Application to Real-World Deployment
- Authors: Rui Xie,
- Abstract summary: This report studies two research papers that represent recent progress on deep learning.
The first paper applied the transformer networks, which are typically used in language models, to improve the quality of synthetic aperture radar image.
The second paper presents an in-storage computing design solution to enable cost-efficient and high-performance implementations of deep learning recommendation systems.
- Score: 3.3813152538225135
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning continues to re-shape numerous fields, from natural language processing and imaging to data analytics and recommendation systems. This report studies two research papers that represent recent progress on deep learning from two largely different aspects: The first paper applied the transformer networks, which are typically used in language models, to improve the quality of synthetic aperture radar image by effectively reducing the speckle noise. The second paper presents an in-storage computing design solution to enable cost-efficient and high-performance implementations of deep learning recommendation systems. In addition to summarizing each paper in terms of motivation, key ideas and techniques, and evaluation results, this report also presents thoughts and discussions about possible future research directions. By carrying out in-depth study on these two representative papers and related references, this doctoral candidate has developed better understanding on the far-reaching impact and efficient implementation of deep learning models.
Related papers
- SciPIP: An LLM-based Scientific Paper Idea Proposer [30.670219064905677]
This paper proposes a scientific paper idea proposer (SciPIP)
Based on a user-provided research background, SciPIP retrieves helpful papers from a literature database.
We then introduce dual-path idea proposal strategies, where one path infers solutions from retrieved literature and the other path generates original ideas.
arXiv Detail & Related papers (2024-10-30T16:18:22Z) - Learning the Bitter Lesson: Empirical Evidence from 20 Years of CVPR Proceedings [1.3812010983144802]
This study examines the alignment of emphConference on Computer Vision and Pattern Recognition (CVPR) research with the principles of the "bitter lesson" proposed by Rich Sutton.
We analyze two decades of CVPR abstracts and titles using large language models (LLMs) to assess the field's embracement of these principles.
arXiv Detail & Related papers (2024-10-12T21:06:13Z) - Towards Coarse-to-Fine Evaluation of Inference Efficiency for Large Language Models [95.96734086126469]
Large language models (LLMs) can serve as the assistant to help users accomplish their jobs, and also support the development of advanced applications.
For the wide application of LLMs, the inference efficiency is an essential concern, which has been widely studied in existing work.
We perform a detailed coarse-to-fine analysis of the inference performance of various code libraries.
arXiv Detail & Related papers (2024-04-17T15:57:50Z) - Computation-efficient Deep Learning for Computer Vision: A Survey [121.84121397440337]
Deep learning models have reached or even exceeded human-level performance in a range of visual perception tasks.
Deep learning models usually demand significant computational resources, leading to impractical power consumption, latency, or carbon emissions in real-world scenarios.
New research focus is computationally efficient deep learning, which strives to achieve satisfactory performance while minimizing the computational cost during inference.
arXiv Detail & Related papers (2023-08-27T03:55:28Z) - Deep networks for system identification: a Survey [56.34005280792013]
System identification learns mathematical descriptions of dynamic systems from input-output data.
Main aim of the identified model is to predict new data from previous observations.
We discuss architectures commonly adopted in the literature, like feedforward, convolutional, and recurrent networks.
arXiv Detail & Related papers (2023-01-30T12:38:31Z) - Multilingual Name Entity Recognition and Intent Classification Employing
Deep Learning Architectures [2.9115403886004807]
We explore the effectiveness of two separate families of Deep Learning networks for named entity recognition and intent classification.
The models were trained and tested on the ATIS benchmark dataset for both English and Greek languages.
arXiv Detail & Related papers (2022-11-04T12:42:29Z) - Deep Depth Completion: A Survey [26.09557446012222]
We provide a comprehensive literature review that helps readers better grasp the research trends and clearly understand the current advances.
We investigate the related studies from the design aspects of network architectures, loss functions, benchmark datasets, and learning strategies.
We present a quantitative comparison of model performance on two widely used benchmark datasets, including an indoor and an outdoor dataset.
arXiv Detail & Related papers (2022-05-11T08:24:00Z) - Enhancing Identification of Structure Function of Academic Articles
Using Contextual Information [6.28532577139029]
This paper takes articles of the ACL conference as the corpus to identify the structure function of academic articles.
We employ the traditional machine learning models and deep learning models to construct the classifiers based on various feature input.
Inspired by (2), this paper introduces contextual information into the deep learning models and achieved significant results.
arXiv Detail & Related papers (2021-11-28T11:21:21Z) - Deep Learning Schema-based Event Extraction: Literature Review and
Current Trends [60.29289298349322]
Event extraction technology based on deep learning has become a research hotspot.
This paper fills the gap by reviewing the state-of-the-art approaches, focusing on deep learning-based models.
arXiv Detail & Related papers (2021-07-05T16:32:45Z) - A Survey on Neural Recommendation: From Collaborative Filtering to
Content and Context Enriched Recommendation [70.69134448863483]
Research in recommendation has shifted to inventing new recommender models based on neural networks.
In recent years, we have witnessed significant progress in developing neural recommender models.
arXiv Detail & Related papers (2021-04-27T08:03:52Z) - Video Summarization Using Deep Neural Networks: A Survey [72.98424352264904]
Video summarization technologies aim to create a concise and complete synopsis by selecting the most informative parts of the video content.
This work focuses on the recent advances in the area and provides a comprehensive survey of the existing deep-learning-based methods for generic video summarization.
arXiv Detail & Related papers (2021-01-15T11:41:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.