ChatGPT 5.5 Pro Solves PhD-Level Math in Under an Hour: Is $5/M Input Worth It for Research Tasks?
May 11, 2026 · 7 min read
A Fields Medalist's Experiment Changes the ROI Conversation
Timothy Gowers, a Fields Medal-winning mathematician at Cambridge, recently ran an experiment that should reshape how developers and researchers think about AI model pricing. He gave ChatGPT 5.5 Pro a set of open problems in number theory, the kind of problems that typically occupy PhD students for months. The model improved an exponential bound to a polynomial bound in under an hour.
An MIT researcher reviewing the work called the core idea "fully original", meaning it was not a recombination of known techniques but a genuinely novel mathematical insight. This is not a benchmark score or a cherry-picked demo. This is a premium AI model producing PhD-caliber work on unsolved problems in a fraction of the time and cost of human research.
The question this raises is not whether AI can do research. It is whether the premium price tag of frontier models is justified when the alternative is human expert time at $50-200 per hour.
The Premium Model Pricing Landscape
GPT-5.5 standard is already expensive at $5.00/$30.00 per million tokens. But the real frontier reasoning happens on the premium tier. Here is how the top-tier models compare against budget alternatives:
| Model | Input (per 1M) | Output (per 1M) | Best For |
|---|---|---|---|
| GPT-5.5 | $5.00 | $30.00 | Deep reasoning, research |
| Claude Opus 4.7 | $5.00 | $25.00 | Complex coding, analysis |
| Gemini 3.1 Pro | $2.00 | $12.00 | Large context tasks |
| DeepSeek R1 | $0.70 | $2.50 | Budget reasoning |
| DeepSeek V4 Flash | $0.14 | $0.28 | Simple tasks, high volume |
| Grok 4.20 | $1.25 | $2.50 | Mid-tier reasoning |
GPT-5.5 costs 35x more per input token and 107x more per output token than DeepSeek V4 Flash. On paper, that sounds indefensible. But the Gowers experiment reveals why raw token cost is the wrong metric for complex tasks.
Calculating the ROI: $0.85 vs Two Hours of PhD Time
Let us run the actual numbers on what a session like Gowers' experiment might cost. A complex research reasoning session with GPT-5.5 typically involves:
- Input tokens: ~100K (the problem statement, context, previous attempts, relevant definitions)
- Output tokens: ~15K (the model's reasoning chain, proof steps, and final result)
At GPT-5.5 rates: (0.1M x $5.00) + (0.015M x $30.00) = $0.50 + $0.45 = $0.95. Call it roughly a dollar for a session that produces a novel mathematical insight.
Now consider the human alternative. A PhD researcher or postdoc working on the same problem costs approximately $50-80 per hour (salary plus overhead at a research university). If the AI saves even 2 hours of human research time, the ROI calculation is straightforward:
- Human cost saved: 2 hours x $50/hr = $100
- AI cost incurred: $0.95
- Net savings: $99.05
- ROI: 105x return on the AI investment
Even if you are conservative and assume the AI only saves 30 minutes of human time, the ROI is still ($25 / $0.95) = 26x. The economics are not even close. For complex reasoning tasks where human expertise is expensive, premium model pricing is essentially a rounding error.
When Does Paying 35x More Make Economic Sense?
The Gowers experiment is an extreme example, but the principle applies broadly. The key variable is task complexity. Here is a framework for when premium models justify their cost:
Premium models win when the task is complex enough that cheap models would need infinite retries. If DeepSeek V4 Flash cannot solve a problem at all, then its $0.14/M input price is irrelevant. You could run it 1,000 times and still not get a correct answer. The effective cost is infinite.
Consider three tiers of task complexity and the model economics for each:
| Task Complexity | Example | Best Model Tier | Typical Cost |
|---|---|---|---|
| Simple | Format JSON, write boilerplate | DeepSeek V4 Flash ($0.14/$0.28) | $0.01-0.05 |
| Moderate | Implement a feature, debug code | Claude Sonnet 4.6 ($3/$15) | $0.50-5.00 |
| Complex | Novel algorithm, research proof | GPT-5.5 / Opus 4.7 ($5/$25-30) | $0.50-2.00 |
| Impossible for cheap models | PhD-level math, open problems | GPT-5.5 ($5/$30) | $0.50-1.50 |
Notice that the "complex" and "impossible" tiers cost roughly the same per session ($0.50-2.00). The difference is not in the AI cost; it is in the human cost of not using the right model. A developer spending 4 hours debugging a subtle concurrency issue could have spent $1.00 on GPT-5.5 or Claude Opus 4.7 and gotten the answer in minutes.
The Developer Parallel: Complex Coding Tasks Follow the Same Logic
You do not need to be proving number theory theorems to benefit from this analysis. The same economics apply to everyday development work. Consider these scenarios where premium models pay for themselves:
- Debugging a production outage: Your site is down, costing $500/hr in lost revenue. GPT-5.5 analyzes your logs and identifies a race condition in 3 minutes for $0.40. The cheaper model misidentifies the issue, adding 30 minutes of downtime.
- Architecture review: You need to evaluate whether your system should use event sourcing vs CQRS. A senior architect charges $200/hr for a 2-hour consultation. Claude Opus 4.7 produces a detailed analysis with tradeoffs specific to your codebase for under $2.00.
- Complex refactoring: Migrating a 10,000-line module from callbacks to async/await. A cheap model introduces subtle bugs that take hours to surface. Opus handles the entire migration correctly on the first pass, saving a full day of debugging.
In each case, the question is not "is GPT-5.5 at $5.00/M too expensive?" It is "is $1-2 per session too expensive compared to the alternative?" When framed this way, the answer is almost always no.
The Budget Reasoning Alternative: DeepSeek R1
For teams that need reasoning capability on a tighter budget, DeepSeek R1 at $0.70/$2.50 per million tokens offers a middle path. It is purpose-built for chain-of-thought reasoning and performs well on mathematical and logical tasks. At roughly 7x cheaper than GPT-5.5 on input and 12x cheaper on output, it is worth testing on your specific use case.
However, DeepSeek R1 has not demonstrated the ability to produce "fully original" mathematical insights the way GPT-5.5 did in the Gowers experiment. For truly novel reasoning, there appears to be a capability threshold that only the most expensive models cross. Below that threshold, you are paying for "pretty good" reasoning. Above it, you are paying for genuine breakthroughs.
The practical strategy for most developers: use DeepSeek R1 or Grok 4.20 ($1.25/$2.50) as your default reasoning model, and escalate to GPT-5.5 or Claude Opus 4.7 only when the cheaper model fails or when the task is too important to risk a wrong answer.
Putting It All Together: A Cost-Aware Research and Coding Strategy
The Gowers experiment proves that the ceiling of what AI models can accomplish is rising faster than their pricing. A session that costs less than a cup of coffee can now produce research-grade output that would take a human expert hours or days. The economic question is no longer "can I afford premium AI?" It is "can I afford not to use it when the task demands it?"
Here is the optimal approach for cost-conscious developers and researchers:
- Default to budget models for routine work: DeepSeek V4 Flash ($0.14/$0.28) handles 70% of coding tasks perfectly well.
- Step up to mid-tier for standard development: Claude Sonnet 4.6 ($3.00/$15.00) for feature work, refactoring, and debugging.
- Reserve premium models for high-complexity tasks: GPT-5.5 ($5.00/$30.00) or Claude Opus 4.7 ($5.00/$25.00) for research, architecture, and problems where getting it wrong costs real money or time.
- Always compare against human time cost: If your hourly rate is $50+ and the AI session costs under $2, the ROI math speaks for itself.
Want to model these costs for your specific project? The AI Cost Estimator lets you compare pricing across all major models and calculate your expected spend based on project size, task complexity, and tooling choice.
Want to calculate exact costs for your project?
Estimate Your AI Coding Costs →