Source Model Selection for Transfer Learning of Image Classification using Supervised Contrastive Loss

Author(s):  
Young-Seong Cho ◽  
Samuel Kim ◽  
Jee-Hyong Lee
Entropy ◽  
2019 ◽  
Vol 21 (5) ◽  
pp. 456 ◽  
Author(s):  
Hao Cheng ◽  
Dongze Lian ◽  
Shenghua Gao ◽  
Yanlin Geng

Inspired by the pioneering work of the information bottleneck (IB) principle for Deep Neural Networks’ (DNNs) analysis, we thoroughly study the relationship among the model accuracy, I ( X ; T ) and I ( T ; Y ) , where I ( X ; T ) and I ( T ; Y ) are the mutual information of DNN’s output T with input X and label Y. Then, we design an information plane-based framework to evaluate the capability of DNNs (including CNNs) for image classification. Instead of each hidden layer’s output, our framework focuses on the model output T. We successfully apply our framework to many application scenarios arising in deep learning and image classification problems, such as image classification with unbalanced data distribution, model selection, and transfer learning. The experimental results verify the effectiveness of the information plane-based framework: Our framework may facilitate a quick model selection and determine the number of samples needed for each class in the unbalanced classification problem. Furthermore, the framework explains the efficiency of transfer learning in the deep learning area.


2020 ◽  
Author(s):  
Parul Awasthy ◽  
Bishwaranjan Bhattacharjee ◽  
John Kender ◽  
Radu Florian

IEEE Access ◽  
2020 ◽  
Vol 8 ◽  
pp. 6190-6200
Author(s):  
Amiel Meiseles ◽  
Lior Rokach

2021 ◽  
Author(s):  
Tushar Semwal ◽  
Haofan Wang ◽  
Chinnakotla Krishna Teja Reddy

Transfer Learning (TL) has achieved significant developments in the past few years. However, the majority of work on TL assume implicit access to both the \textit{target} and \textit{source} datasets, which limits its application in the context of Federated Learning (FL), where target (client) datasets are usually not directly accessible. In this paper, we address the problem of source model selection in TL for federated scenarios. We propose a simple framework, called Selective Federated Transfer Learning (SFTL), to select the best pre-trained models which provide a positive performance gain when their parameters are transferred on to a new task. We leverage the concepts from representation similarity to compare the similarity of the client model and the source models and provide a method which could be augmented to existing FL algorithms to improve both the communication cost and the accuracy of client models.


Sign in / Sign up

Export Citation Format

Share Document