KNOWLEDGE DISTILLATION INSPIRED FINE-TUNING OF TUCKER DECOMPOSED CNNs AND ADVERSARIAL ROBUSTNESS ANALYSIS
Ranajoy Sadhukhan, Avinab Saha, Jayanta Mukhopadhyay, Amit Patra
-
SPS
IEEE Members: $11.00
Non-members: $15.00Length: 13:11
The recent works in the tensor decomposition of convolutional neural networks have paid little attention to finetuning the decomposed models more effectively. We propose to improve the accuracy as well as the adversarial robustness of decomposed networks over existing noniterative methods by distilling knowledge from the computationally intensive undecomposed (teacher) model to the decomposed (student) model. Through a series of experiments, we demonstrate the effectiveness of knowledge distillation with different loss functions and compare it to the existing fine-tuning strategy of minimizing cross-entropy loss with ground truth labels. Finally, we conclude that the student networks obtained by the proposed approach are superior not only in terms of accuracy but also adversarial robustness, which is often compromised in the existing methods.