Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 167 tok/s
Gemini 2.5 Pro 47 tok/s Pro
GPT-5 Medium 39 tok/s Pro
GPT-5 High 29 tok/s Pro
GPT-4o 92 tok/s Pro
Kimi K2 188 tok/s Pro
GPT OSS 120B 429 tok/s Pro
Claude Sonnet 4.5 34 tok/s Pro
2000 character limit reached

Parametrized Complexity of Quantum Inspired Algorithms (2112.11686v1)

Published 22 Dec 2021 in quant-ph and cs.DS

Abstract: Motivated by recent progress in quantum technologies and in particular quantum software, research and industrial communities have been trying to discover new applications of quantum algorithms such as quantum optimization and machine learning. Regardless of which hardware platform these novel algorithms operate on, whether it is adiabatic or gate based, from theoretical point of view, they are performing drastically better than their classical counterparts. Two promising areas of quantum algorithms quantum machine learning and quantum optimization. These are based on performing matrix operations using quantum states and operation, in order to speed up data analysis where quantum computing can efficiently work with high dimensional vectors. Motivated by that, quantum inspired algorithms (e.g. for recommendation systems and principal component analysis) are developed to cope with high dimensionality using probabilistic techniques that are inspire from quantum computing. In this paper we review recent progress in the area of quantum inspired algorithms for low rank matrix approximation. We further explore the possibility of using parametrized complexity for such algorithms to refine practical complexity analysis. Finally, we conjecture that quantum inspired algorithms that use low rank approximation and also sample and query technique for input representations are Fixed Parameter Tractable (FPT).

Summary

We haven't generated a summary for this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.