Comment by jasonjmcghee
Comment by jasonjmcghee 2 days ago
Idk if I'm just holding it wrong, but calling Gemini 3 "the best model in the world" doesn't line up with my experience at all.
It seems to just be worse at actually doing what you ask.
Comment by jasonjmcghee 2 days ago
Idk if I'm just holding it wrong, but calling Gemini 3 "the best model in the world" doesn't line up with my experience at all.
It seems to just be worse at actually doing what you ask.
Unless you overfit to benchmark style scenarios and are worse for real-world use.
Not really, it's like asking which C compiler was best back in the 90s.
You had Watcom, Intel, GCC, Borland, Microsoft, etc.
They all had different optimizations and different target markets.
Best to make your tooling model agnostic. I understand that tuned prompts are model _version_ specific, so you will need this anyways.
It's a good model. Zvi also thought it was the best model until Opus 4.5 was announced a few hours after he wrote his post
https://thezvi.substack.com/p/gemini-3-pro-is-a-vast-intelli...
It's like saying "Star Wars is the best movie in the world" - to some people it is. To others it's terrible.
I feel like it would be advantageous to move away from a "one model fits all" mindset, and move towards a world where we have different genres of models that we use for different things.
The benchmark scores are turning into being just as useful as tomatometer movie scores. Something can score high, but if that's not the genre you like, the high score doesn't guarantee you'll like it.