The University of Southampton
University of Southampton Institutional Repository

GrOD: deep learning with gradients orthogonal decomposition for knowledge transfer, distillation, and adversarial training

GrOD: deep learning with gradients orthogonal decomposition for knowledge transfer, distillation, and adversarial training
GrOD: deep learning with gradients orthogonal decomposition for knowledge transfer, distillation, and adversarial training

Regularization that incorporates the linear combination of empirical loss and explicit regularization terms as the loss function has been frequently used for many machine learning tasks. The explicit regularization term is designed in different types, depending on its applications. While regularized learning often boost the performance with higher accuracy and faster convergence, the regularization would sometimes hurt the empirical loss minimization and lead to poor performance. To deal with such issues in this work, we propose a novel strategy, namely Gradients Orthogonal Decomposition (GrOD), that improves the training procedure of regularized deep learning. Instead of linearly combining gradients of the two terms, GrOD re-estimates a new direction for iteration that does not hurt the empirical loss minimization while preserving the regularization affects, through orthogonal decomposition. We have performed extensive experiments to use GrOD improving the commonly used algorithms of transfer learning [2], knowledge distillation [3], and adversarial learning [4]. The experiment results based on large datasets, including Caltech 256 [5], MIT indoor 67 [6], CIFAR-10 [7], and ImageNet [8], show significant improvement made by GrOD for all three algorithms in all cases.

Deep neural networks, gradient-based learning, regularized deep learning
1556-4681
1-25
Xiong, Haoyi
ce4ad3c5-7887-4830-941c-02e593f20dae
Wan, Ruosi
d6cdf439-669f-4e34-b736-90fb87044a80
Zhao, Jian
5ef19b9f-bc56-4bfd-97ae-fdaa7327f14e
Chen, Zeyu
9dca53b3-37a6-4a30-a72a-d73151dd904e
Li, Xingjian
b8318628-5f1e-4334-8bff-ea1e489a1cea
Zhu, Zhanxing
e55e7385-8ba2-4a85-8bae-e00defb7d7f0
Huan, Jun
0ea4757d-fe12-44b7-9928-6f94e70117ae
Xiong, Haoyi
ce4ad3c5-7887-4830-941c-02e593f20dae
Wan, Ruosi
d6cdf439-669f-4e34-b736-90fb87044a80
Zhao, Jian
5ef19b9f-bc56-4bfd-97ae-fdaa7327f14e
Chen, Zeyu
9dca53b3-37a6-4a30-a72a-d73151dd904e
Li, Xingjian
b8318628-5f1e-4334-8bff-ea1e489a1cea
Zhu, Zhanxing
e55e7385-8ba2-4a85-8bae-e00defb7d7f0
Huan, Jun
0ea4757d-fe12-44b7-9928-6f94e70117ae

Xiong, Haoyi, Wan, Ruosi, Zhao, Jian, Chen, Zeyu, Li, Xingjian, Zhu, Zhanxing and Huan, Jun (2022) GrOD: deep learning with gradients orthogonal decomposition for knowledge transfer, distillation, and adversarial training. ACM Transactions on Knowledge Discovery from Data, 16 (6), 1-25, [117]. (doi:10.1145/3530836).

Record type: Article

Abstract

Regularization that incorporates the linear combination of empirical loss and explicit regularization terms as the loss function has been frequently used for many machine learning tasks. The explicit regularization term is designed in different types, depending on its applications. While regularized learning often boost the performance with higher accuracy and faster convergence, the regularization would sometimes hurt the empirical loss minimization and lead to poor performance. To deal with such issues in this work, we propose a novel strategy, namely Gradients Orthogonal Decomposition (GrOD), that improves the training procedure of regularized deep learning. Instead of linearly combining gradients of the two terms, GrOD re-estimates a new direction for iteration that does not hurt the empirical loss minimization while preserving the regularization affects, through orthogonal decomposition. We have performed extensive experiments to use GrOD improving the commonly used algorithms of transfer learning [2], knowledge distillation [3], and adversarial learning [4]. The experiment results based on large datasets, including Caltech 256 [5], MIT indoor 67 [6], CIFAR-10 [7], and ImageNet [8], show significant improvement made by GrOD for all three algorithms in all cases.

Text
3530836 - Version of Record
Available under License Creative Commons Attribution.
Download (1MB)

More information

Accepted/In Press date: 1 March 2022
e-pub ahead of print date: 18 April 2022
Published date: 8 September 2022
Additional Information: Funding Information: This work was partially supported by the National Key R&D Program of China (No. 2021ZD0110303) to Haoyi Xiong, Zeyu Chen and Xingjian Li, and National Science Foundation of China (No. 62006244), Young Elite Scientist Sponsorship Program of China Association for Science and Technology (No. YESS20200140) to Jian Zhao. An earlier version of this work that focuses on the regularization for deep transfer learning only has been accepted for publication as “Towards Making Deep Transfer Learning Never Hurt” [1] in the 2019 IEEE International Conference on Data Mining (ICDM-19).
Keywords: Deep neural networks, gradient-based learning, regularized deep learning

Identifiers

Local EPrints ID: 486135
URI: http://eprints.soton.ac.uk/id/eprint/486135
ISSN: 1556-4681
PURE UUID: 136cdafa-5b4b-483f-9caa-731449c46f99

Catalogue record

Date deposited: 10 Jan 2024 17:42
Last modified: 05 Jun 2024 20:00

Export record

Altmetrics

Contributors

Author: Haoyi Xiong
Author: Ruosi Wan
Author: Jian Zhao
Author: Zeyu Chen
Author: Xingjian Li
Author: Zhanxing Zhu
Author: Jun Huan

Download statistics

Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.

View more statistics

Atom RSS 1.0 RSS 2.0

Contact ePrints Soton: eprints@soton.ac.uk

ePrints Soton supports OAI 2.0 with a base URL of http://eprints.soton.ac.uk/cgi/oai2

This repository has been built using EPrints software, developed at the University of Southampton, but available to everyone to use.

We use cookies to ensure that we give you the best experience on our website. If you continue without changing your settings, we will assume that you are happy to receive cookies on the University of Southampton website.

×