📱

Read on Your E-Reader

Thousands of readers get articles like this delivered straight to their e-reader. Works with Kindle, Boox, and any device that syncs with Google Drive or Dropbox.

Learn More

This is a preview. The full article is published at techcrunch.com.

AI models are starting to crack high-level math problems

AI models are starting to crack high-level math problems

By Russell BrandomTechCrunch

Over the weekend, Neel Somani, who is a software engineer, former quant researcher, and a startup founder, was testing the math skills of OpenAI’s new model when he made an unexpected discovery. After pasting the problem into ChatGPT and letting it think for 15 minutes, he came back to a full solution. He evaluated the proof and formalized it with a tool called Harmonic - but it all checked out. “I was curious to establish a baseline for when LLMs are effectively able to solve open math problems compared to where they struggle,” Somani said. The surprise was that, using the latest model, the frontier started to push forward a bit. ChatGPT’s chain of thought is even more impressive, rattling off mathematical axioms like Legendre’s formula , Bertrand’s postulate , and the Star of David theorum . Eventually, the model found a Math Overflow post from 2013 , where Harvard mathematician Noam Elkies had given an elegant solution to a similar problem. But ChatGPT’s final proof differed from Elkies’ work in important ways, and gave a more complete solution to a version of the problem posed by legendary mathematician Paul Erdős, whose vast collection of unsolved problems has become a proving ground for AI. For anyone skeptical of machine intelligence, it’s a surprising result - and it’s not the only one. AI tools have become ubiquitous in mathematics, from formalization-oriented LLMs like Harmonic’s Aristotle to literature review tools like OpenAI’s deep research. But since the release of GPT 5.2 - which Somani describes as “anecdotally more skilled at mathematical reasoning than previous iterations” - the sheer volume of solved problems has become difficult to ignore, raising new questions about large language models’ ability to push the frontiers of human knowledge. Somani was looking at the Erdős problems, a set of over one thousand conjectures by the Hungarian mathematician that are maintained online . The problems have become a tempting target for AI-driven mathematics, varying significantly in both subject matter and difficulty. The first batch of autonomous solutions came in November from a Gemini-powered model called AlphaEvolve - but more recently, Somani and others have found GPT 5.2 to be remarkably adept with high-level math. Since Christmas, 15 problems have been moved from “open” to “solved” on the Erdős website - and 11 of the solutions have specifically credited AI models as involved in the process. The revered mathematician Terence Tao has a more nuanced look at the progress on his GitHub page , counting eight different problems where AI models made meaningful autonomous progress on an Erdős problem, with six other cases where progress was made by locating and building on previous research. It’s a long way from AI systems being able to do math without human intervention, but it’s clear that there’s an important role for large models to play. Join the Disrupt 2026 Waitlist Add yourself to the Disrupt 2026 waitlist to be first in line when Early Bird tickets drop. Past Disrupts have brought Google Cloud, Netflix,...

Preview: ~500 words

Continue reading at Techcrunch

Read Full Article

More from TechCrunch

Subscribe to get new articles from this feed on your e-reader.

View feed

This preview is provided for discovery purposes. Read the full article at techcrunch.com. LibSpace is not affiliated with Techcrunch.

AI models are starting to crack high-level math problems | Read on Kindle | LibSpace