logoalt Hacker News

energy123last Wednesday at 7:03 AM2 repliesview on HN

It's the best model pound for pound, but I find GPT 5.2 Thinking/Pro to be more useful for serious work when run with xhigh effort. I can get it to think for 20 minutes, but Gemini 3.0 Pro is like 2.5 minutes max. Obviously I lack full visibility because tok/s and token efficiency likely differs between them, but I take it as a proxy of how much compute they're giving us per inference, and it matches my subjective judgement of output quality. Maybe Google nerfs the reasoning effort in the Gemini subscription to save money and that's why I am experiencing this.


Replies

knowrijulast Wednesday at 10:24 AM

When ChatGPT takes 20 minutes to reason, is it actually spending all the time burning tokens or does a bulk of the time go into 'scheduling' waits. If someone specifically selected xhigh reasoning, I am guessing it can be processed with high batch count.

show 1 reply
cjlast Wednesday at 7:28 AM

I'm curious, what types of prompts are you running that benefit from 10+ minutes of think time?

Whats the quality difference between default ChatGPT and Thinking? Is it an extra 20% quality boost or is the difference night/day?

I've often imagined it would be great to have some kind of chrome extension or 3rd party tool to always run prompts in multiple thinking tiers so you can get an immediate response to read while you wait for the thinking models to think.

show 1 reply