Ranking and Rejecting of Pre-Trained Deep Neural Networks in Transfer
Learning based on Separation Index
- URL: http://arxiv.org/abs/2012.13717v1
- Date: Sat, 26 Dec 2020 11:14:12 GMT
- Title: Ranking and Rejecting of Pre-Trained Deep Neural Networks in Transfer
Learning based on Separation Index
- Authors: Mostafa Kalhor, Ahmad Kalhor, and Mehdi Rahmani
- Abstract summary: We introduce an algorithm to rank pre-trained Deep Neural Networks (DNNs) by applying a distance-based complexity measure named Separation Index (SI) to the target dataset.
The efficiency of the proposed algorithm is evaluated by using three challenging datasets including Linnaeus 5, Breast Cancer Images, and COVID-CT.
- Score: 0.16058099298620418
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Automated ranking of pre-trained Deep Neural Networks (DNNs) reduces the
required time for selecting optimal pre-trained DNN and boost the
classification performance in transfer learning. In this paper, we introduce a
novel algorithm to rank pre-trained DNNs by applying a straightforward
distance-based complexity measure named Separation Index (SI) to the target
dataset. For this purpose, at first, a background about the SI is given and
then the automated ranking algorithm is explained. In this algorithm, the SI is
computed for the target dataset which passes from the feature extracting parts
of pre-trained DNNs. Then, by descending sort of the computed SIs, the
pre-trained DNNs are ranked, easily. In this ranking method, the best DNN makes
maximum SI on the target dataset and a few pre-trained DNNs may be rejected in
the case of their sufficiently low computed SIs. The efficiency of the proposed
algorithm is evaluated by using three challenging datasets including Linnaeus
5, Breast Cancer Images, and COVID-CT. For the two first case studies, the
results of the proposed algorithm exactly match with the ranking of the trained
DNNs by the accuracy on the target dataset. For the third case study, despite
using different preprocessing on the target data, the ranking of the algorithm
has a high correlation with the ranking resulted from classification accuracy.
Related papers
- Unveiling the Power of Sparse Neural Networks for Feature Selection [60.50319755984697]
Sparse Neural Networks (SNNs) have emerged as powerful tools for efficient feature selection.
We show that SNNs trained with dynamic sparse training (DST) algorithms can achieve, on average, more than $50%$ memory and $55%$ FLOPs reduction.
Our findings show that feature selection with SNNs trained with DST algorithms can achieve, on average, more than $50%$ memory and $55%$ FLOPs reduction.
arXiv Detail & Related papers (2024-08-08T16:48:33Z) - SA-CNN: Application to text categorization issues using simulated
annealing-based convolutional neural network optimization [0.0]
Convolutional neural networks (CNNs) are a representative class of deep learning algorithms.
We introduce SA-CNN neural networks for text classification tasks based on Text-CNN neural networks.
arXiv Detail & Related papers (2023-03-13T14:27:34Z) - Unsupervised Learning of Initialization in Deep Neural Networks via
Maximum Mean Discrepancy [74.34895342081407]
We propose an unsupervised algorithm to find good initialization for input data.
We first notice that each parameter configuration in the parameter space corresponds to one particular downstream task of d-way classification.
We then conjecture that the success of learning is directly related to how diverse downstream tasks are in the vicinity of the initial parameters.
arXiv Detail & Related papers (2023-02-08T23:23:28Z) - RankDNN: Learning to Rank for Few-shot Learning [70.49494297554537]
This paper introduces a new few-shot learning pipeline that casts relevance ranking for image retrieval as binary ranking relation classification.
It provides a new perspective on few-shot learning and is complementary to state-of-the-art methods.
arXiv Detail & Related papers (2022-11-28T13:59:31Z) - Towards Better Out-of-Distribution Generalization of Neural Algorithmic
Reasoning Tasks [51.8723187709964]
We study the OOD generalization of neural algorithmic reasoning tasks.
The goal is to learn an algorithm from input-output pairs using deep neural networks.
arXiv Detail & Related papers (2022-11-01T18:33:20Z) - Training Algorithm Matters for the Performance of Neural Network
Potential [4.774810604472842]
We compare the performance of two popular training algorithms, the adaptive moment estimation algorithm (Adam) and the extended Kalman filter algorithm (EKF)
It is found that NNPs trained with EKF are more transferable and less sensitive to the value of the learning rate, as compared to Adam.
arXiv Detail & Related papers (2021-09-08T16:48:33Z) - Adaptive Nearest Neighbor Machine Translation [60.97183408140499]
kNN-MT combines pre-trained neural machine translation with token-level k-nearest-neighbor retrieval.
Traditional kNN algorithm simply retrieves a same number of nearest neighbors for each target token.
We propose Adaptive kNN-MT to dynamically determine the number of k for each target token.
arXiv Detail & Related papers (2021-05-27T09:27:42Z) - Evaluating Deep Neural Network Ensembles by Majority Voting cum
Meta-Learning scheme [3.351714665243138]
We propose an ensemble of seven independent Deep Neural Networks (DNNs) for a new data instance.
One-seventh of the data is deleted and replenished by bootstrap sampling from the remaining samples.
All the algorithms in this paper have been tested on five benchmark datasets.
arXiv Detail & Related papers (2021-05-09T03:10:56Z) - Gradient-only line searches to automatically determine learning rates
for a variety of stochastic training algorithms [0.0]
We study the application of the Gradient-Only Line Search that is Inexact (GOLS-I) to determine the learning rate schedule for a selection of popular neural network training algorithms.
GOLS-I's learning rate schedules are competitive with manually tuned learning rates, over seven optimization algorithms, three types of neural network architecture, 23 datasets and two loss functions.
arXiv Detail & Related papers (2020-06-29T08:59:31Z) - File Classification Based on Spiking Neural Networks [0.5065947993017157]
We propose a system for file classification in large data sets based on spiking neural networks (SNNs)
The proposed system may represent a valid alternative to classical machine learning algorithms for inference tasks.
arXiv Detail & Related papers (2020-04-08T11:50:29Z) - Learning to Hash with Graph Neural Networks for Recommender Systems [103.82479899868191]
Graph representation learning has attracted much attention in supporting high quality candidate search at scale.
Despite its effectiveness in learning embedding vectors for objects in the user-item interaction network, the computational costs to infer users' preferences in continuous embedding space are tremendous.
We propose a simple yet effective discrete representation learning framework to jointly learn continuous and discrete codes.
arXiv Detail & Related papers (2020-03-04T06:59:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.