Neural networks for abstraction and reasoning: Towards broad generalization in machines (2402.03507v1)
Abstract: For half a century, artificial intelligence research has attempted to reproduce the human qualities of abstraction and reasoning - creating computer systems that can learn new concepts from a minimal set of examples, in settings where humans find this easy. While specific neural networks are able to solve an impressive range of problems, broad generalisation to situations outside their training data has proved elusive.In this work, we look at several novel approaches for solving the Abstraction & Reasoning Corpus (ARC), a dataset of abstract visual reasoning tasks introduced to test algorithms on broad generalization. Despite three international competitions with $100,000 in prizes, the best algorithms still fail to solve a majority of ARC tasks and rely on complex hand-crafted rules, without using machine learning at all. We revisit whether recent advances in neural networks allow progress on this task. First, we adapt the DreamCoder neurosymbolic reasoning solver to ARC. DreamCoder automatically writes programs in a bespoke domain-specific language to perform reasoning, using a neural network to mimic human intuition. We present the Perceptual Abstraction and Reasoning Language (PeARL) language, which allows DreamCoder to solve ARC tasks, and propose a new recognition model that allows us to significantly improve on the previous best implementation.We also propose a new encoding and augmentation scheme that allows LLMs to solve ARC tasks, and find that the largest models can solve some ARC tasks. LLMs are able to solve a different group of problems to state-of-the-art solvers, and provide an interesting way to complement other approaches. We perform an ensemble analysis, combining models to achieve better results than any system alone. Finally, we publish the arckit Python library to make future research on ARC easier.
- Spartan Books.
- Chollet F (2019). On the measure of intelligence. ArXiv: 1911.01547.
- Nature 596: 583–589.
- Nature 529: 484–489.
- Brandom R (2018). Self-driving cars are headed toward an ai roadblock. URL https://www.theverge.com/2018/7/3/17530232/self-driving-ai-winter-full-autonomy-waymo-tesla-uber.
- Azulay A, Weiss Y (2019) Why do deep convolutional networks generalize so poorly to small image transformations? J Mach Learn Res 20: 184:1–184:25.
- Abstraction and reasoning challenge. URL https://kaggle.com/competitions/abstraction-and-reasoning-challenge.
- Wind JS (2022). 1st place solution + code and official documentation. Kaggle Forums. URL https://www.kaggle.com/c/abstraction-and-reasoning-challenge/discussion/154597.
- Alford S (2021) A Neurosymbolic Approach to Abstraction and Reasoning. Master’s thesis, Massachusetts Institute of Technology.
- In: Larochelle H, Ranzato M, Hadsell R, Balcan M, Lin H, editors, 33rd Annual Conference on Neural Information Processing Systems (NeurIPS).
- OpenAI (2023). GPT-4 technical report. ArXiv:2303.08774.
- LLaMA: Open and efficient foundation language models. ArXiv:2302.13971.
- Basic Books New York.
- Foundalis H (2007). Index of bongard problems. URL https://www.foundalis.com/res/bps/bpidx.htm.
- Halme A (2020). Bongard solvers. URL https://notes.fringeling.com/BongardSolvers/.
- Kharagorgiev S (2018). Solving Bongard problems with deep learning. URL https://k10v.github.io/2018/02/25/Solving-Bongard-problems-with-deep-learning/.
- Solving Bongard problems with a visual language and pragmatic reasoning. ArXiv:1804.04452.
- Foundalis HE (2006) Phaeaco: A Cognitive Architecture Inspired by Bongard’s problems. Ph.D. thesis, Indiana University.
- In: Bengio S, Wallach HM, Larochelle H, Grauman K, Cesa-Bianchi N, et al., editors, 31st Annual Conference on Neural Information Processing Systems (NeurIPS) 2018. pp. 7549–7561.
- Legg S, Hutter M (2007). A collection of definitions of intelligence. ArXiv:0706.3639.
- Gershgorn D (2017). The data that changed the direction of AI research and possibly the world. URL https://qz.com/1034972/the-data-that-changed-the-direction-of-ai-research-and-possibly-the-world.
- Raven JC (1936) Mental tests used in genetic studies: The performance of related individuals on tests mainly educative and mainly reproductive. Master’s thesis, University of London.
- Graphs, constraints, and search for the abstraction and reasoning corpus. ArXiv:arXiv.2210.09880.
- Golubev V (2019). 3rd place[short preview+code]. Kaggle Forums. URL https://www.kaggle.com/competitions/abstraction-and-reasoning-challenge/discussion/154305.
- Golubev V (2019). 7 solved tasks via trees. Kaggle Kernels. URL https://www.kaggle.com/code/golubev/7-solved-tasks-via-trees/notebook.
- Technical report, Center for Brains, Minds and Machines (CBMM). URL https://dspace.mit.edu/handle/1721.1/128607.
- DreamCoder: Growing generalizable, interpretable knowledge with wake-sleep Bayesian program learning. ArXiv:2006.08381.
- Muggleton SH (1991) Inductive logic programming. New Generation Computing 8: 295–318.
- Machine Learning 111: 147–172.
- Neural Computation 7: 889–904.
- Ellis K (2022). Dreamcoder official code. GitHub. URL https://github.com/ellisk42/ec.
- Alford S (2021). bidir-synth: DreamCoder and bidirectional program synthesis on ARC. URL https://github.com/simonalford42/bidir-synth.git.
- IEEE Transactions on Pattern Analysis and Machine Intelligence 37: 1904–1916.
- In: Bengio Y, LeCun Y, editors, 2nd International Conference on Learning Representations, ICLR 2014, Banff, AB, Canada, April 14-16, 2014.
- In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2015, Boston, MA, USA, June 7-12, 2015. IEEE Computer Society, pp. 3431–3440.
- Yu F, Koltun V (2016) Multi-scale context aggregation by dilated convolutions. In: Bengio Y, LeCun Y, editors, 4th International Conference on Learning Representations, ICLR 2016, San Juan, Puerto Rico, May 2-4, 2016.
- Kingma DP, Ba J (2015) Adam: A method for stochastic optimization. In: Bengio Y, LeCun Y, editors, 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015. URL http://arxiv.org/abs/1412.6980.
- (2023). Python reference: Full grammar specification. URL https://docs.python.org/3/reference/grammar.html.
- Training language models to follow instructions with human feedback. ArXiv: 2203.02155.
- ArXiv: 2107.03374.
- Emergent Analogical Reasoning in Large Language Models. ArXiv:2212.09196.
- Mitchell M (2023). On analogy-making in large language models. URL https://aiguide.substack.com/p/on-analogy-making-in-large-language.
- Mitchell M (2020). Can GPT-3 make analogies? URL https://medium.com/@melaniemitchell.me/can-gpt-3-make-analogies-16436605c446.
- Mitchell M (1993) Analogy-Making as Perception. MIT Press.
- MIT Technology Review .
- GPTQ: Accurate post-training quantization for generative pre-trained transformers. ArXiv: 2210.17323.
- qwopqwop200 (2023). GPTQ for LLaMa. URL https://github.com/qwopqwop200/GPTQ-for-LLaMa.
- OpenAI (2023). TikToken. URL https://github.com/openai/tiktoken.
- OpenAI (2023). Introducing ChatGPT. URL https://openai.com/blog/chatgpt.
- ArXiv: 2203.15556.
- Sequential modeling enables scalable learning for large vision models. ArXiv:2312.00785.
- In: Conference on Computer Vision and Pattern Recognition, CVPR 2017, Youtube-8M Workshop.
- Vijaymeena M, Kavitha K (2016) A survey on similarity measures in text mining. Machine Learning and Applications: An International Journal 3: 19–28.
- In: The Tenth International Conference on Learning Representations, ICLR.
- In: Thirty-sixth Conference on Neural Information Processing Systems (NeurIPS).
- Lab42 (2023). ARCreate: Crowdsourcing ARC 2. URL https://arc-editor.lab42.global/.
- Comparing Humans, GPT-4, and GPT-4V on abstraction and reasoning tasks. ArXiv:2311.09247v2.
- Proceedings of the 43rd Annual Meeting of the Cognitive Science Society: Comparative Cognition: Animal Minds, CogSci 2021 : 2471-2477.
- Wind JS. DSL solution to the ARC challenge. URL https://github.com/top-quarks/ARC-solution.