Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 134 tok/s
Gemini 2.5 Pro 41 tok/s Pro
GPT-5 Medium 27 tok/s Pro
GPT-5 High 29 tok/s Pro
GPT-4o 119 tok/s Pro
Kimi K2 180 tok/s Pro
GPT OSS 120B 418 tok/s Pro
Claude Sonnet 4.5 37 tok/s Pro
2000 character limit reached

Towards Antisymmetric Neural Ansatz Separation (2208.03264v3)

Published 5 Aug 2022 in cs.LG

Abstract: We study separations between two fundamental models (or \emph{Ans\"atze}) of antisymmetric functions, that is, functions $f$ of the form $f(x_{\sigma(1)}, \ldots, x_{\sigma(N)}) = \text{sign}(\sigma)f(x_1, \ldots, x_N)$, where $\sigma$ is any permutation. These arise in the context of quantum chemistry, and are the basic modeling tool for wavefunctions of Fermionic systems. Specifically, we consider two popular antisymmetric Ans\"atze: the Slater representation, which leverages the alternating structure of determinants, and the Jastrow ansatz, which augments Slater determinants with a product by an arbitrary symmetric function. We construct an antisymmetric function in $N$ dimensions that can be efficiently expressed in Jastrow form, yet provably cannot be approximated by Slater determinants unless there are exponentially (in $N2$) many terms. This represents the first explicit quantitative separation between these two Ans\"atze.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (30)
  1. Solving the quantum many-body problem with artificial neural networks. Science, 355(6325):602–606, 2017.
  2. Spherical cnns. arXiv preprint arXiv:1801.10130, 2018.
  3. Amit Daniely. Depth separation for neural networks. In Conference on Learning Theory, pp.  690–696. PMLR, 2017.
  4. The power of depth for feedforward neural networks. In Conference on learning theory, pp.  907–940. PMLR, 2016.
  5. RP Feynman and Michael Cohen. Energy spectrum of the excitations in liquid helium. Physical Review, 102(5):1189, 1956.
  6. Universal approximation of symmetric and anti-symmetric functions. arXiv preprint arXiv:1912.01765, 2019a.
  7. Solving many-electron schrödinger equation using deep neural networks. Journal of Computational Physics, 399:108929, 2019b.
  8. Deep-neural-network solution of the electronic schrödinger equation. Nature Chemistry, 12(10):891–897, 2020.
  9. Geometry of backflow transformation ansatz for quantum many-body fermionic wavefunctions. arXiv preprint arXiv:2111.10314, 2021.
  10. Marcus Hutter. On representing (anti) symmetric functions. arXiv preprint arXiv:2007.15298, 2020.
  11. Generalizations of cauchy’s determinant and schur’s pfaffian. Advances in Applied Mathematics, 36(3):251–287, 2006.
  12. Robert Jastrow. Many-body problem with strong forces. Physical Review, 98(5):1479, 1955.
  13. Highly accurate protein structure prediction with alphafold. Nature, 596(7873):583–589, 2021.
  14. Imagenet classification with deep convolutional neural networks. Communications of the ACM, 60(6):84–90, 2017.
  15. Edwin Langmann. A method to derive explicit formulas for an elliptic generalization of the jack polynomials. arXiv preprint math-ph/0511015, 2005.
  16. Convolutional networks for images, speech, and time series. The handbook of brain theory and neural networks, 3361(10):1995, 1995.
  17. Set transformer: A framework for attention-based permutation-invariant neural networks. In International Conference on Machine Learning, pp. 3744–3753. PMLR, 2019.
  18. Di Luo and Bryan K Clark. Backflow transformations via neural networks for quantum many-body wave functions. Physical review letters, 122(22):226401, 2019.
  19. Approximation bounds for smooth functions in c (r/sup d/) by neural and mixture networks. IEEE Transactions on Neural Networks, 9(5):969–978, 1998.
  20. Attila Maróti. On elementary lower bounds for the partition function. Integers: Electronic J. Comb. Number Theory, 3:A10, 2003.
  21. Fermionic wave functions from neural-network constrained hidden states. arXiv preprint arXiv:2111.10420, 2021.
  22. Ab initio solution of the many-electron schrödinger equation with deep neural networks. Physical Review Research, 2(3):033429, 2020.
  23. Optimization-based separations for neural networks. In Conference on Learning Theory, pp.  3–64. PMLR, 2022.
  24. A simple neural network module for relational reasoning. Advances in neural information processing systems, 30, 2017.
  25. Thomas Sundquist. Two variable pfaffian identities and symmetric functions. Journal of Algebraic Combinatorics, 5(2):135–148, 1996.
  26. Modern quantum chemistry: introduction to advanced electronic structure theory. Courier Corporation, 2012.
  27. On the limitations of representing functions on sets. In International Conference on Machine Learning, pp. 6487–6494. PMLR, 2019.
  28. Universal approximation of functions on sets. Journal of Machine Learning Research, 23(151):1–56, 2022.
  29. Deep sets. Advances in neural information processing systems, 30, 2017.
  30. Exponential separations in symmetric neural networks. arXiv preprint arXiv:2206.01266, 2022.
Citations (4)

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

Authors (2)

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.