( ・_・)ノ Ritchot's Corner

GPT-5 has come a long way in Mathematics

Key Points

Last December, I wrote an article titled o1 pro mode still has a long way to go for Mathematics. At the time, OpenAI’s new “reasoning” models were being heavily marketed as smarter, more careful, and better at logic. In my own tests, the reality in a secondary math context was much less impressive. The model was better than its predecessors, but it still missed enough questions that I could reasonably say my mathematics classroom was largely safe.

With the release of GPT-5 in August 2025, I wanted to revisit those tests with a new set of questions to see how far LLMs have actually come. The leaps forward have been astounding. Not even a year later, I can no longer assert that these tools “have a long way to go” in the way I meant back then.

Compared to last year’s o1 pro mode run, where the model cleared the “4/4 reliability” bar on only 40 of 60 questions (67 percent) and answered 177 of 240 individual trials correctly (74 percent), GPT-5 represents a substantial jump in consistency. In this new round of testing, it achieved 4/4 reliability on 48 of 50 questions with the “use code” prompt and 48 of 50 without it, and produced 392 correct answers out of 400 total attempts, for a 98 percent success rate.

Put differently, under essentially the same problem framework provided by the University of Waterloo’s Centre for Education in Mathematics and Computing (CEMC), the system has moved from missing roughly one in four attempts to missing about one in fifty. The single stubborn failure was the same question type regardless of whether I explicitly encouraged it to call on tools such as writing code. As before, it seemed more like an issue with the image model rather than with mathematics ability.

Last Year’s Test

In my original article, I had a relatively simple question: how good, really, are these models at the kind of mathematics my students actually do?

OpenAI’s o1 and o1 pro modes were marketed as “thinking” models that reasoned through problems, especially in mathematics and science. They were multimodal, which meant they could allegedly handle text and images in a more unified way. For an educator previously in the secondary math trenches, this raised obvious questions about assessment, homework, and what it means to do authentic work in a world where any question can be fed into ChatGPT.

To stress test those claims in a classroom-relevant way, I used the CEMC Problems of the Week from the University of Waterloo. These are not competition-only questions for Olympiad students, but structured problems that span grades 3 to 12, with a mix of number sense, algebra, geometry, and word problems. I took 12 questions at each of five difficulty levels for a total of 60 questions, converted the PDFs into images, and evaluated o1 pro mode on four trials per question using a simple, consistent prompt.

The model reached “4/4 reliability” on 67 percent of questions and got 74 percent of individual attempts correct. More interesting than the aggregate stats, though, was the pattern of failure: o1 pro mode struggled badly with vision. It often misread diagrams, dropped key paths from network problems, or simply failed to extract basic information from images, especially in more visual problems at the Grade 5–6 level.

My conclusion at the time was that students could still not reliably offload standard homework or test questions to an AI, and multimodal problems were often especially safe.

That is no longer the case.

This Year’s Setup

For the GPT-5 run, I wanted continuity with last year’s experiment, but I also wanted to close off some obvious loopholes. If the model really has improved, it should perform well even when I am not going out of my way to “prompt engineer” its success.

Question Source

I again used the CEMC Problems of the Week, drawing from a new set of questions from the current academic year. Instead of 60 questions, I used 50, spread across multiple grade bands and topics, including geometry, number sense, and algebraic reasoning as they were released.

For someone with no budget and lack of time, this is my only realistically feasible way to try to ensure that these exact questions were not already in the training data, though at this point I have to assume that questions with similar patterns and structures are present.

Prompts, Images, And “Thinking Mode”

Last time, the vision model was clearly the weak link (and I had a weaker understanding of how LLM inputs worked), so I wanted to test whether that had changed while making my input generally better for an LLM. To do that, I:

Solve the following. Clearly explain how you arrived at your result. Use code if necessary.

I exclusively used the new “thinking” mode (and later, extended thinking) allowed by GPT-5. If OpenAI is going to claim that extended deliberation leads to better reasoning, it seemed only fair to force the model to think.

More on The Code Prompt vs No-Code Prompt

One practical question for educators is whether it matters if a student explicitly asks the model to use tools like a code interpreter.

To probe that, I set up two conditions:

  1. A “use code” prompt where I explicitly encouraged the model to call upon code or other tools if helpful.
  2. A “no code mentioned” prompt, where I did not tell it to use tools at all.

For each of the 50 questions, GPT-5 answered it four times under each condition. That gave me 200 attempts with “use code” and 200 without, 400 total.

There was also one small, but hopefully not substantial, change near the end of testing. When GPT-5.1 rolled out in mid-November, my default “robotic” personality setting shifted to “efficient” with two questions left. If the model’s persona had a substantial effect on its actual mathematical performance, that would be worth knowing. In practice, I saw no obvious impact.

The Results

The headline numbers bear repeating, because they are the heart of why this follow-up exists at all. They are also why I am never going to bother with this particular test again.

4/4 reliability:

Per-attempt accuracy:

In practice, only two questions ever produced wrong answers. As before, they were questions that likely stumped the image model used by GPT-5. One involved calculating the area of different spaces in a rectangular park; the other involved “writing a program” (think pseudocode) to direct a robot through a visual maze.

If you care about what students can do in real classrooms, this shift matters more than the jump from, say, 74 percent to 80 percent. Last year, letting an LLM do your math homework was like using an unreliable calculator that gives you the wrong answer about one time in four. This year, it is more like a calculator that flickers once every few dozen questions.

Ethan Mollick has pointed out that, despite being weak at math a year ago, current models now dominate hard STEM contests like the International Mathematical Olympiad and other Olympiads. Robert Ghrist, a professor of mathematics and engineering at Penn, has commented that he now has to work hard to design unambiguous numerical problems that GPT-5-Pro cannot solve, something that was “totally different even 4–6 months ago.” My CEMC experiment is a small, classroom-scale echo of that broader step change.

That level of reliability is absolutely enough to undermine the integrity of any assessment that assumes students cannot trivially get correct answers on demand.

Tool Calls, Prompting, And Why Our Training Is Out Of Date

One of the most interesting findings was what did not matter.

I saw no meaningful difference between runs where I explicitly prompted “use code if helpful” and runs where I said nothing about tools at all. GPT-5 seemed to decide independently whether or not to call its own tools during the hidden thinking stage. In at least one instance, I saw it “debate” tool use internally: it would think for a while, consider using code, decide against it, then proceed with a manual derivation anyway.

Overall, for usability this is impressive. It means that “tool use” is now something the model can manage on its own, rather than something we gate through prompts. Students do not need to understand when code is appropriate. The system can allocate tool use for them.

This aligns very closely with Mollick’s argument that as models get larger and better, they become more capable of inferring intent, and the detailed “prompt formulas” we have been teaching become less relevant. He has been blunt about the fact that many organizations are now heavily invested in training practices that were appropriate for models from six months ago, but not for the ones we are actually using today. Reasoning models make chain-of-thought prompting less important. What matters more is context, clear goals, and giving the AI a well-defined job.

A lot of teacher PD and corporate “AI workshops” are still organized around magic acronyms, rigid templates, and the promise that if you follow a particular prompt recipe, the AI will finally work. At 98 percent accuracy on non-trivial problems, my experience matches Mollick’s. The hard part is no longer “how do I get this to function at all,” but “how do I design tasks and systems around the fact that this mostly just works.”

Vision And Geometry Are Still A Lagging Edge

Even with model improvements, some of the old weaknesses persist, although they are now narrower and more subtle.

A few patterns stood out:

One of the small but telling shifts was how often it got the parts of visualization right. Last year, I would not have trusted an LLM to correctly illustrate a number line with labeled fractions for my students without careful checking. This year, I watched it place numbers correctly and explain the reasoning in ways that were usable for teaching.

Using GPT-5 As A Teaching Tool

If GPT-5 can now clear a non-trivial, curriculum-aligned math benchmark with 98 percent accuracy, the question is not just “can students cheat with this” but “what would it look like to use this responsibly in instruction.”

A few possibilities are already clear:

As these are still probabilistic models, quality is not uniform across responses, even for the same question. Sometimes the explanations are clumsy or overly verbose. Sometimes it skips a pedagogically important step that may confuse a younger learner. There may well be differences in how it explains things depending on which personality or mode is selected, or whether you nudge it to “explain this to a Grade 8 student” versus “show all formal steps.”

Further, GPT-5.1 now exposes personality controls that let you pick from predefined styles or define your own, with OpenAI’s own documentation encouraging you to “customize your ChatGPT personality.” In practice, I have seen examples where different personality settings give fundamentally different styles of advice, including different suggested breathing patterns for a presenter and different role expectations. As a teacher, I really want more clarity on the functional implications of AI personality. If one student uses “Warm and Encouraging” and another uses “Efficient and Direct,” are they getting subtly different mathematical norms and expectations from the same underlying model?

The baseline, though, is now that students have access to a free or low-cost AI tool which can already act as a reasonably competent math tutor. It might not replace a skilled teacher, but it will absolutely replace a large share of what homework, extra practice, and worked examples used to look like.

AI As A Mathematical Collaborator

If this were only about middle school or high school math, we might still tell ourselves a comforting story: “Sure, it can do worksheets, but serious mathematics is safe.”

Scott Aaronson—an American theoretical computer scientist best known for his work on quantum computing and computational complexity theory—recently described, for the first time, a research paper where a key technical step in the proof of the main result was supplied by an AI, using GPT-5-Thinking. He was clear that if a student had handed him the same argument, he would have called it clever. In his longer write-up, Aaronson points out that an AI that “merely” fills in the insights that should have been obvious to you is still a huge deal for real research, because it speeds up the actual discovery process, not just the LaTeX or bibliography.

Terence Tao—widely regarded as one of the greatest living mathematicians—has written about using extended conversations with an AI to help answer a nontrivial MathOverflow question. He had already done theoretical work suggesting a particular answer, but used AI-assisted heuristic calculations to locate feasible parameters for a counterexample, then verified them with a simple program. Without AI, he suggests he probably would not even have attempted that numerical search.

At multiple levels of mathematical practice, from contest problems to classroom exercises to research-level work, there is growing evidence that modern models are not just “good at math for a chatbot.” They are becoming meaningful collaborators.

When people operating at the frontiers of mathematical research are saying “this noticeably sped up my work” and “this suggested a clever key step,” it becomes very hard to maintain the fiction that high school algebra is out of reach.

AI And Academic Scholarship More Broadly

The same pattern shows up outside mathematics. A recent paper on AI and academic publishing, summarized by Jay Van Bavel, found that researchers using generative AI published substantially more papers and that the quality of those papers, as measured by journal impact factors, also rose. The productivity gap between AI users and non-users grew from about 15 percent in 2023 to over a third in 2024. There were also disproportionate gains for early-career researchers and authors from non-English-speaking countries, suggesting that AI is not just increasing output, but also helping level parts of the playing field.

In other words, AI is not just good enough to help my Grade 8 students fake their homework. It is already altering the trajectory of academic careers and research output.

We can argue about whether this is good or bad, or about what “authorship” and “scholarship” should mean in this context. But we can no longer argue, in good faith, that these tools are marginal or that we have ample time before they matter.

Institutions, Benchmarks, And “Job Interviewing” Your Models

All of this creates a problem for institutions that like clean policies and slow cycles…which is most of education.

Mollick has argued that as AI models get better and more embedded in work, organizations need to stop relying on vibes and generic benchmarks. In his “giving your AI a job interview” piece, he points to research like GDPval, which shows performance varying significantly by task even among top models, and to cases like “GuacaDrone,” where different models offer systematically different advice on ambiguous, judgment-heavy questions.

It is not enough to know that a model scores well on some broad benchmark like MMLU. You need to know what your model does on your tasks, including the ways it might be systematically better or worse, more or less risk-seeking, more or less conservative. That requires realistic scenarios, repeated trials, and expert review, and it is not a one-time effort. You need to do it multiple times a year as new models are released and old ones drift.

My CEMC experiment is, in a very modest way, an example of this kind of local benchmarking. I took a real question source that actually appears in the lives of my students, defined a simple reliability framework, and ran repeated trials. The result was not “GPT-5 scored 98 percent on a leaderboard,” but “GPT-5 will almost always get the problems my students do in class correct, using minimal prompting.”

Last year, I wrote that I did not know a single school, district, or board meaningfully allocating time and personnel to rigorously test emerging models against internal, context-specific benchmarks. That is still true. What has changed is that the models have leapt forward while the institutional response has mostly stayed frozen.

At this point, saying “we need to wait and see how good these things get” is not a serious position. For most of the math our students do, they are already good enough.

Limitations and Caveats

None of this is a peer-reviewed study. I am still a classroom teacher with severe time constraints, running tests in the margins of a very busy job.

There are real limitations here:

Despite all that, the direction of travel is clear enough. The worrying part to me is not that GPT-5 can ace a CEMC worksheet. It is that institutions are still writing policies and designing assessments as if last December’s performance is the ceiling.

Where This Leaves The Classroom

So where does this leave a classroom now?

For one, we can no longer responsibly tell ourselves that standard secondary math homework is a strong measure of student understanding. If a student wants to outsource everything to an AI, the friction is now vanishingly small and the error rate is low enough that they can coast for quite a while before it catches up to them.

It also means that “just use images” is no longer a robust defense for take-home tasks. Vision remains imperfect, especially for complex geometry, but it is not a safe harbor. If something can be cleanly described in text and solved with symbolic or numeric reasoning, GPT-5 is probably already good at it.

Personality customization adds another wrinkle. If different students pick different AI “vibes” and get different types of explanations, hints, and levels of hand-holding, we will need to think carefully about equity, scaffolding, and what we count as independent work. The same underlying model might behave like a patient tutor for one student and an efficiency-obsessed problem solver for another.

None of this means it is time to give up on math education. It means we have to be much more intentional about:

Last year, I ended by telling my students that they still had to do the heavy lifting. This year, I think the more honest message is that the heavy lifting has shifted. They may not need to grind as many practice problems by hand, but they absolutely need to learn how to question, interpret, and extend the work that an AI hands them.

If we do not teach that, someone else—or something else—will.

If you want to chat, shoot me an email. If you would like to get updates, subscribe to my blog via email or RSS feed. You can also follow me at LinkedIn, and X.

  1. You can find an archive of the questions I used at the CEMC website. It is the first 10 questions for each level set from 2025/2026. I also have an archive, which you can contact me to obtain.

  2. You can download a recording of my results as an XLSX file.