TY - JOUR
T1 - Specific Expert Learning
T2 - Enriching Ensemble Diversity via Knowledge Distillation
AU - Kao, Wei Cheng
AU - Xie, Hong Xia
AU - Lin, Chih Yang
AU - Cheng, Wen Huang
N1 - Publisher Copyright:
© 2013 IEEE.
PY - 2023/4/1
Y1 - 2023/4/1
N2 - In recent years, ensemble methods have shown sterling performance and gained popularity in visual tasks. However, the performance of an ensemble is limited by the paucity of diversity among the models. Thus, to enrich the diversity of the ensemble, we present the distillation approach - learning from experts (LFEs). Such method involves a novel knowledge distillation (KD) method that we present, specific expert learning (SEL), which can reduce class selectivity and improve the performance on specific weaker classes and overall accuracy. Through SEL, models can acquire different knowledge from distinct networks with various areas of expertise, and a highly diverse ensemble can be obtained afterward. Our experimental results demonstrate that, on CIFAR-10, the accuracy of the ResNet-32 increases 0.91% with SEL, and that the ensemble trained by SEL increases accuracy by 1.13%. Compared to state-of-the-art approaches, for example, DML only improves accuracy by 0.3% and 1.02% on single ResNet-32 and the ensemble, respectively. Furthermore, our proposed architecture also can be applied to ensemble distillation (ED), which applies KD on the ensemble model. In conclusion, our experimental results show that our proposed SEL not only improves the accuracy of a single classifier but also boosts the diversity of the ensemble model.
AB - In recent years, ensemble methods have shown sterling performance and gained popularity in visual tasks. However, the performance of an ensemble is limited by the paucity of diversity among the models. Thus, to enrich the diversity of the ensemble, we present the distillation approach - learning from experts (LFEs). Such method involves a novel knowledge distillation (KD) method that we present, specific expert learning (SEL), which can reduce class selectivity and improve the performance on specific weaker classes and overall accuracy. Through SEL, models can acquire different knowledge from distinct networks with various areas of expertise, and a highly diverse ensemble can be obtained afterward. Our experimental results demonstrate that, on CIFAR-10, the accuracy of the ResNet-32 increases 0.91% with SEL, and that the ensemble trained by SEL increases accuracy by 1.13%. Compared to state-of-the-art approaches, for example, DML only improves accuracy by 0.3% and 1.02% on single ResNet-32 and the ensemble, respectively. Furthermore, our proposed architecture also can be applied to ensemble distillation (ED), which applies KD on the ensemble model. In conclusion, our experimental results show that our proposed SEL not only improves the accuracy of a single classifier but also boosts the diversity of the ensemble model.
KW - Deep learning
KW - ensemble diversity
KW - knowledge distillation (KD)
UR - http://www.scopus.com/inward/record.url?scp=85120038588&partnerID=8YFLogxK
U2 - 10.1109/TCYB.2021.3125320
DO - 10.1109/TCYB.2021.3125320
M3 - 期刊論文
C2 - 34793316
AN - SCOPUS:85120038588
SN - 2168-2267
VL - 53
SP - 2494
EP - 2505
JO - IEEE Transactions on Cybernetics
JF - IEEE Transactions on Cybernetics
IS - 4
ER -