[1] vs [2]

Untitled

[2]

Untitled

[1]

same idea, same architecture, different networks

  1. expert / teacher network: pre-trained VGG-16 vs Resnet18 on ImageNet
  2. cloner / student network: a simplified VGG-16 vs the same Resnet18, both of them are randomly initialized.
  3. both a multiresolution scheme, i.e. transfer knowledge from intermediate layers
  4. same loss: cosine lose between corresponding layers.
    1. [1]是逐像素的, [2]的cosine是对整个layer的,所以定位的时候要用gradient

different network location mechanism

[1] is simpler and better.

[2] via gradient of loss function

AUC-ROC / AR for anomaly detection on MVTecAD: 0.955 vs 0.8774

AUC-ROC /AR for anomaly localization on MVTecAD: 0.970 vs 0.9071

References

[1] Student-Teacher Feature Pyramid Matching for Anomaly Detection, BMVC21

  1. https://github.com/hcw-00/STPM_anomaly_detection
    1. Unofficial pytorch implementation, 和论文效果一致。
  2. 张浚然
    1. anomaly detection和location的threshold如何确定。
  3. fig 4 说明pre-trained on imagenet比用其他小数据集好很多.

[2] MKD, cvpr21