Weighted Mutual Learning with Diversity-Driven Model Compression.

Miao Zhang, Li Wang, David Campos, Wei Huang, Chenjuan Guo, Bin Yang

Publikation: Konferencebidrag uden forlag/tidsskriftPaper uden forlag/tidsskriftForskningpeer review


Online distillation attracts attention from the community as it simplifies the traditional two-stage knowledge distillation process into a single stage. Online distillation collaboratively trains a group of peer models, which are treated as students, and all students gain extra knowledge from each other. However, memory consumption and diversity among students are two key challenges to the scalability and quality of online distillation. To address the two challenges, this paper presents a framework called Weighted Mutual Learning with Diversity-Driven Model Compression (WML) for online distillation. First, at the base of a hierarchical structure where students share different parts, we leverage the structured network pruning to generate diversified students with different models sizes, thus also helping reduce the memory requirements. Second, rather than taking the average of students, this paper, for the first time, leverages a bi-level formulation to estimate the relative importance of students with a close-form, to further boost the effectiveness of the distillation from each other. Extensive experiments show the generalization of the proposed framework, which outperforms existing online distillation methods on a variety of deep neural networks. More interesting, as a byproduct, WML produces
a series of students with different model sizes in a single run, which also achieves
competitive results compared with existing channel pruning methods.
StatusUdgivet - 2022


Dyk ned i forskningsemnerne om 'Weighted Mutual Learning with Diversity-Driven Model Compression.'. Sammen danner de et unikt fingeraftryk.