TY - GEN
T1 - A Novel Self-Knowledge Distillation Approach with Siamese Representation Learning for Action Recognition
AU - Vu, Duc Quang
AU - Phung, Thi Thu Trang
AU - Wang, Jia Ching
N1 - Publisher Copyright:
© 2021 IEEE.
PY - 2021
Y1 - 2021
N2 - Knowledge distillation is an effective transfer of knowledge from a heavy network (teacher) to a small network (student) to boost students' performance. Self-knowledge dis-tillation, the special case of knowledge distillation, has been proposed to remove the large teacher network training process while preserving the student's performance. This paper intro-duces a novel Self-knowledge distillation approach via Siamese representation learning, which minimizes the difference between two representation vectors of the two different views from a given sample. Our proposed method, SKD-SRL, utilizes both soft label distillation and the similarity of representation vectors. Therefore, SKD-SRL can generate more consistent predictions and representations in various views of the same data point. Our benchmark has been evaluated on various standard datasets. The experimental results have shown that SKD-SRL significantly improves the accuracy compared to existing supervised learning and knowledge distillation methods regardless of the networks.
AB - Knowledge distillation is an effective transfer of knowledge from a heavy network (teacher) to a small network (student) to boost students' performance. Self-knowledge dis-tillation, the special case of knowledge distillation, has been proposed to remove the large teacher network training process while preserving the student's performance. This paper intro-duces a novel Self-knowledge distillation approach via Siamese representation learning, which minimizes the difference between two representation vectors of the two different views from a given sample. Our proposed method, SKD-SRL, utilizes both soft label distillation and the similarity of representation vectors. Therefore, SKD-SRL can generate more consistent predictions and representations in various views of the same data point. Our benchmark has been evaluated on various standard datasets. The experimental results have shown that SKD-SRL significantly improves the accuracy compared to existing supervised learning and knowledge distillation methods regardless of the networks.
UR - http://www.scopus.com/inward/record.url?scp=85125278790&partnerID=8YFLogxK
U2 - 10.1109/VCIP53242.2021.9675335
DO - 10.1109/VCIP53242.2021.9675335
M3 - 會議論文篇章
AN - SCOPUS:85125278790
T3 - 2021 International Conference on Visual Communications and Image Processing, VCIP 2021 - Proceedings
BT - 2021 International Conference on Visual Communications and Image Processing, VCIP 2021 - Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2021 International Conference on Visual Communications and Image Processing, VCIP 2021
Y2 - 5 December 2021 through 8 December 2021
ER -