Do Large Language Models Solve ARC Visual Analogies Like People Do?

Open Access
Authors
Publication date 2024
Host editors
  • L. Samuelson
  • S. Frank
  • M. Toneva
  • A. Mackey
  • E. Hazeltine
Book title 46th Annual Meeting of the Cognitive Science Society (CogSci 2024)
Book subtitle Dynamics of Cognition : Rotterdam, the Netherlands, 24-27 July 2024
ISBN
  • 9798331309060
Series Proceedings of the Annual Meeting of the Cognitive Science Society
Event 46th Annual Meeting of the Cognitive Science Society
Volume | Issue number 1
Pages (from-to) 579-586
Publisher Austin, TX: Cognitive Science Society
Organisations
  • Faculty of Social and Behavioural Sciences (FMG) - Psychology Research Institute (PsyRes)
Abstract
The Abstraction Reasoning Corpus (ARC) is a visual analogical reasoning test designed for humans and machines (Chollet, 2019). We compared human and large language model (LLM) performance on a new child-friendly set of ARC items. Results show that both children and adults outperform most LLMs on these tasks. Error analysis revealed a similar "fallback" solution strategy in LLMs and young children, where part of the analogy is simply copied. In addition, we found two other error types, one based on seemingly grasping key concepts (e.g., Inside-Outside) and the other based on simple combinations of analogy input matrices. On the whole, "concept" errors were more common in humans, and "matrix" errors were more common in LLMs. This study sheds new light on LLM reasoning ability and the extent to which we can use error analyses and comparisons with human development to understand how LLMs solve visual analogies.
Document type Conference contribution
Language English
Published at https://escholarship.org/uc/item/4bp4m6cf
Other links https://github.com/cstevenson-uva/kidsARC https://www.proceedings.com/77494.html
Downloads
qt4bp4m6cf (Final published version)
Permalink to this page
Back