A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2020; you can also visit the original URL.
The file type is application/pdf
.
Deep Mutual Learning
2018
2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition
Model distillation is an effective and widely used technique to transfer knowledge from a teacher to a student network. The typical application is to transfer from a powerful large network or ensemble to a small network, in order to meet the low-memory or fast execution requirements. In this paper, we present a deep mutual learning (DML) strategy. Different from the one-way transfer between a static pre-defined teacher and a student in model distillation, with DML, an ensemble of students learn
doi:10.1109/cvpr.2018.00454
dblp:conf/cvpr/ZhangXHL18
fatcat:gfarpispcfgsni7aslq7gk4yvi