Be Your Own Teacher: Improve the Performance of Convolutional Neural Networks via Self Distillation

Loss Source 1: Cross entropy loss,各個階段的分類器都有 Loss Source 2: KL loss,深層的分類器作爲淺層分類器的teacher Loss Source 3: L2 loss from hints,深層分類器的特徵和淺層分類器的特徵做L2 loss,bottleneck即feature adaptation,爲了使student和teacher一樣
相關文章
相關標籤/搜索