Abstract: Knowledge distillation (KD) shows a bright promise as a powerful regularization strategy to boost generalization ability by leveraging learned sample-level soft targets. Yet, employing a ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results