Sida Liu

A Learner in the Complex World.


Why LLMs Aren’t Good at Math

Large language models (LLMs) are trained by processing huge amounts of text. They “learn” by reading, much like people do through reading and listening.

When kids develop thinking skills, they first learn language, and only much later—around third grade or so—do they begin to understand math and logical reasoning. Thinking in precise, step-by-step ways is challenging for young kids and likely tied to stages of brain development.

In Daniel Kahneman’s terms, “System 1” (fast, intuitive thinking) develops earlier, while “System 2” (slow, analytical thinking) comes later and requires more effort—and, in the context of LLMs, greater computational power.

Similarly, LLMs struggle with math. While they can understand scientific facts to some degree, they aren’t yet able to apply this knowledge with the exact, logical thinking needed to solve problems accurately.

To illustrate this issue more clearly, consider how we solve algebra. We hold expressions in mind and mentally manipulate them step by step. In contrast, LLMs function more like someone who talks through each step out loud, relying on what they just said to reason through the next step. If we tried this approach ourselves, it would slow down and disrupt our ability to perform math effectively.

For now, it’s best not to rely on LLMs for math or tasks that require precise logical reasoning—they’re simply not there yet.



2 responses to “Why LLMs Aren’t Good at Math”

  1. […] our previous discussion on Why LLMs Aren’t Good at Math, we explored some limitations of large language models (LLMs) in mathematical reasoning. But what […]

    Like

  2. I notice that OpenAI’s O1 and O3 models are getting better and better at solving math problems, likely due to heavy use of reinforcement learning. So, maybe I shouldn’t say that LLMs aren’t good at math anymore.

    Like

Leave a comment