Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 60 tok/s
Gemini 2.5 Pro 51 tok/s Pro
GPT-5 Medium 18 tok/s Pro
GPT-5 High 14 tok/s Pro
GPT-4o 77 tok/s Pro
Kimi K2 159 tok/s Pro
GPT OSS 120B 456 tok/s Pro
Claude Sonnet 4 38 tok/s Pro
2000 character limit reached

Tensor network compressibility of convolutional models (2403.14379v2)

Published 21 Mar 2024 in cs.CV, cs.LG, and quant-ph

Abstract: Convolutional neural networks (CNNs) are one of the most widely used neural network architectures, showcasing state-of-the-art performance in computer vision tasks. Although larger CNNs generally exhibit higher accuracy, their size can be effectively reduced by tensorization'' while maintaining accuracy, namely, replacing the convolution kernels with compact decompositions such as Tucker, Canonical Polyadic decompositions, or quantum-inspired decompositions such as matrix product states, and directly training the factors in the decompositions to bias the learning towards low-rank decompositions. But why doesn't tensorization seem to impact the accuracy adversely? We explore this by assessing how \textit{truncating} the convolution kernels of \textit{dense} (untensorized) CNNs impact their accuracy. Specifically, we truncated the kernels of (i) a vanilla four-layer CNN and (ii) ResNet-50 pre-trained for image classification on CIFAR-10 and CIFAR-100 datasets. We found that kernels (especially those inside deeper layers) could often be truncated along several cuts resulting in significant loss in kernel norm but not in classification accuracy. This suggests that suchcorrelation compression'' (underlying tensorization) is an intrinsic feature of how information is encoded in dense CNNs. We also found that aggressively truncated models could often recover the pre-truncation accuracy after only a few epochs of re-training, suggesting that compressing the internal correlations of convolution layers does not often transport the model to a worse minimum. Our results can be applied to tensorize and compress CNN models more effectively.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (18)
  1. K. Simonyan and A. Zisserman,   (2014), arXiv:1409.1556 .
  2. F. Chollet,   (2016), arXiv:1610.02357 .
  3. V. O. Hinton, Geoffrey and J. Dean,  (2015), arXiv:1503.02531 .
  4. L. R. Tucker, Psychometrika 1966 31:3 31, 279 (1966).
  5. R. A. Harshman, UCLA Working Papers in Phonetics 16, 1 (1970).
  6. T. G. Kolda and B. W. Bader, SIAM Review 51, 455 (2009).
  7. R. R. Orús, Annals of Physics 349, 117 (2014), arXiv:1306.2164 .
  8. M. D. Zeiler and R. Fergus,   (2013), arXiv:1311.2901 .
  9. V. Dumoulin and F. Visin,   (2016), arXiv:1603.07285 .
  10. The softmax layer is not essential in shallow CNNs.
  11. MATLAB, https://www.mathworks.com/help/ images/ref/im2col.html .
  12. F. Dangel,  (2023), arXiv:2307.02275 .
  13. We use reshape to mean the operation implemented by the eponymous NumPy function.
  14. The compression ratio is the ratio of the size of the dense and tensorized models.
  15. C. Hillar and L.-H. Lim,   (2009), arXiv:0911.1393 .
  16. V. d. Silva and L.-H. Lim, SIAM Journal on Matrix Analysis and Applications 30 (2008).
  17. See Theorem 4.2 on https://www.tensors.net/tutorial-4.
  18. A graphical calculus for tensors was first popularized by Penrose. The current graphical calculus of quantum tensor networks can be formalized as string diagrams in a suitable category.
Citations (1)

Summary

We haven't generated a summary for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Lightbulb On Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com
Youtube Logo Streamline Icon: https://streamlinehq.com