PosCUDA: Position based Convolution for Unlearnable Audio Datasets (2401.02135v1)
Abstract: Deep learning models require large amounts of clean data to acheive good performance. To avoid the cost of expensive data acquisition, researchers use the abundant data available on the internet. This raises significant privacy concerns on the potential misuse of personal data for model training without authorisation. Recent works such as CUDA propose solutions to this problem by adding class-wise blurs to make datasets unlearnable, i.e a model can never use the acquired dataset for learning. However these methods often reduce the quality of the data making it useless for practical applications. We introduce PosCUDA, a position based convolution for creating unlearnable audio datasets. PosCUDA uses class-wise convolutions on small patches of audio. The location of the patches are based on a private key for each class, hence the model learns the relations between positional blurs and labels, while failing to generalize. We empirically show that PosCUDA can achieve unlearnability while maintaining the quality of the original audio datasets. Our proposed method is also robust to different audio feature representations such as MFCC, raw audio and different architectures such as transformers, convolutional networks etc.
- Kashmir Hill, “The secretive company that might end privacy as we know it,” in Ethics of Data and Analytics, pp. 170–177. Auerbach Publications, 2022.
- Eric Drott, “Copyright, compensation, and commons in the music ai industry,” Creative Industries Journal, vol. 14, no. 2, pp. 190–207, 2021.
- “Artificial intelligence and music: open questions of copyright law and engineering praxis,” in Arts. MDPI, 2019, vol. 8, p. 115.
- “Adversarial examples make strong poisons,” Advances in Neural Information Processing Systems, vol. 34, pp. 30339–30351, 2021.
- “Unlearnable examples: Making personal data unexploitable,” arXiv preprint arXiv:2101.04898, 2021.
- “Neural tangent generalization attacks,” in International Conference on Machine Learning. PMLR, 2021, pp. 12230–12240.
- “Robust unlearnable examples: Protecting data against adversarial learning,” arXiv preprint arXiv:2203.14533, 2022.
- “Cuda: Convolution-based unlearnable datasets,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 3862–3871.
- “Targeted backdoor attacks on deep learning systems using data poisoning,” arXiv preprint arXiv:1712.05526, 2017.
- “Neural attention distillation: Erasing backdoor triggers from deep neural networks,” arXiv preprint arXiv:2101.05930, 2021.
- “Reflection backdoor: A natural backdoor attack on deep neural networks,” in Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part X 16. Springer, 2020, pp. 182–199.
- “Input-aware dynamic backdoor attack,” Advances in Neural Information Processing Systems, vol. 33, pp. 3454–3464, 2020.
- Cynthia Dwork, “Differential privacy,” in International colloquium on automata, languages, and programming. Springer, 2006, pp. 1–12.
- “Deep learning with differential privacy,” in Proceedings of the 2016 ACM SIGSAC conference on computer and communications security, 2016, pp. 308–318.
- “Dpd-infogan: Differentially private distributed infogan,” in Proceedings of the 1st Workshop on Machine Learning and Systems, 2021, pp. 1–6.
- “Towards generalizable data protection with transferable unlearnable examples,” arXiv preprint arXiv:2305.11191, 2023.
- “Very deep convolutional neural networks for raw waveforms,” in 2017 IEEE international conference on acoustics, speech and signal processing (ICASSP). IEEE, 2017, pp. 421–425.
- “Attention is all you need,” Advances in neural information processing systems, vol. 30, 2017.
- “Fréchet audio distance: A reference-free metric for evaluating music enhancement algorithms.,” in INTERSPEECH, 2019, pp. 2350–2354.
- “Gans trained by a two time-scale update rule converge to a local nash equilibrium,” Advances in neural information processing systems, vol. 30, 2017.
- “Panns: Large-scale pretrained audio neural networks for audio pattern recognition,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 28, pp. 2880–2894, 2020.