当前位置:网站首页>Toward a Unified Model

Toward a Unified Model

2022-08-11 06:16:00 zhSunw


insert image description here
Two-stage Knowledge Learning (Collation-Examination) is designed on the basis of knowledge ledgedistillation and the method of contrastive learning is introduced.

Method

Collaborative Knowledge Transfer (CKT)

Progressive Feature Projector (PFP)

Project the features of the teacher and student networks into a common feature space, and calculate the L1 loss of the two projected features: Q represents the number of coding layers, Ti represents the ith teacher network, and S represents the student network
Projection is achieved by a small network consisting of several convolutional blocks with stride 1 and a ReLU activation function
insert image description here
insert image description here

Bidirectional Feature Matching

Project the projected features of the teacher network back to the original input space, and calculate the L1 loss of the two features to ensure the validity of the projection
insert image description here
insert image description here

Multi-contrastive Regularization (Two-stage Knowledge Learning)

insert image description here

Contrastive Regularization: v, v+, v- represent predicted samples, positive samples (target samples), and negative samples respectively; Ψ represents the features extracted by VGG-19; R is the number of negative samples.

Soft Contrastive Regularization. (Knowledge Collation (KC).)

Soft: Use teacher's predictions as positive samples to avoid the challenges of early learning
hereinsert image description
insert image description here

Hard Contrastive Regularization.(Knowledge Examination (KE).)

Hard: Use groundtruth as a positive sample to learn directly after training for a period of time
Insert image description here
insert image description here

原网站

版权声明
本文为[zhSunw]所创,转载请带上原文链接,感谢
https://yzsam.com/2022/223/202208110514323321.html