Does GPT Really Get It? A Hierarchical Scale to Quantify Human and AI’s Understanding of Algorithms

Authors

  • Mirabel Reid Georgia Institute of Technology
  • Santosh S. Vempala Georgia Institute of Technology

DOI:

https://doi.org/10.1609/aaai.v39i2.32140

Abstract

As Large Language Models (LLMs) are used for increasingly complex cognitive tasks, a natural question is whether AI really understands. The study of understanding in LLMs is in its infancy, and the community has yet to incorporate research and insights from philosophy, psychology, and education. Here we focus on understanding algorithms, and propose a hierarchy of levels of understanding. We validate the hierarchy using a study with human subjects (undergraduate and graduate students). Following this, we apply the hierarchy to large language models (generations of GPT), revealing interesting similarities and differences with humans. We expect that our rigorous criteria for algorithm understanding will help monitor and quantify AI's progress in such cognitive domains.

Published

2025-04-11

How to Cite

Reid, M., & Vempala, S. S. (2025). Does GPT Really Get It? A Hierarchical Scale to Quantify Human and AI’s Understanding of Algorithms. Proceedings of the AAAI Conference on Artificial Intelligence, 39(2), 1492–1500. https://doi.org/10.1609/aaai.v39i2.32140

Issue

Section

AAAI Technical Track on Cognitive Modeling & Cognitive Systems