Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Limitations of weak labels for embedding and tagging (2002.01687v4)

Published 5 Feb 2020 in cs.SD, cs.AI, cs.LG, and eess.AS

Abstract: Many datasets and approaches in ambient sound analysis use weakly labeled data.Weak labels are employed because annotating every data sample with a strong label is too expensive.Yet, their impact on the performance in comparison to strong labels remains unclear.Indeed, weak labels must often be dealt with at the same time as other challenges, namely multiple labels per sample, unbalanced classes and/or overlapping events.In this paper, we formulate a supervised learning problem which involves weak labels.We create a dataset that focuses on the difference between strong and weak labels as opposed to other challenges. We investigate the impact of weak labels when training an embedding or an end-to-end classifier.Different experimental scenarios are discussed to provide insights into which applications are most sensitive to weakly labeled data.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Nicolas Turpault (10 papers)
  2. Romain Serizel (53 papers)
  3. Emmanuel Vincent (44 papers)
Citations (9)

Summary

We haven't generated a summary for this paper yet.