Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
125 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

S2SD: Simultaneous Similarity-based Self-Distillation for Deep Metric Learning (2009.08348v3)

Published 17 Sep 2020 in cs.CV

Abstract: Deep Metric Learning (DML) provides a crucial tool for visual similarity and zero-shot applications by learning generalizing embedding spaces, although recent work in DML has shown strong performance saturation across training objectives. However, generalization capacity is known to scale with the embedding space dimensionality. Unfortunately, high dimensional embeddings also create higher retrieval cost for downstream applications. To remedy this, we propose \emph{Simultaneous Similarity-based Self-distillation (S2SD). S2SD extends DML with knowledge distillation from auxiliary, high-dimensional embedding and feature spaces to leverage complementary context during training while retaining test-time cost and with negligible changes to the training time. Experiments and ablations across different objectives and standard benchmarks show S2SD offers notable improvements of up to 7% in Recall@1, while also setting a new state-of-the-art. Code available at https://github.com/MLforHealth/S2SD.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Karsten Roth (36 papers)
  2. Timo Milbich (15 papers)
  3. Björn Ommer (72 papers)
  4. Joseph Paul Cohen (50 papers)
  5. Marzyeh Ghassemi (96 papers)
Citations (17)

Summary

We haven't generated a summary for this paper yet.