Edit: As mentioned by @tedsanders below, the post was edited to include clarifying language such as: “Both models make clear mistakes, but GPT‑5.2 shows better comprehension of the image.”
Edit: As mentioned by @tedsanders below, the post was edited to include clarifying language such as: “Both models make clear mistakes, but GPT‑5.2 shows better comprehension of the image.”
0: https://images.ctfassets.net/kftzwdyauwt9/6lyujQxhZDnOMruN3f...
- there's no good reason it's called "chat" instead of "Instant"
- gpt-5.1 and gpt-5.1-chat are different models, even though they both reason now. gpt-5.1 is more factual and can think for much longer. most people want gpt-5.1, unless the use case is ChatGPT-like or they prefer its personality.
> I hope AGI can be used to automate work
You people need a PR guy, I'm serious. OpenAI is the first company I've ever seen that comes across as actively trying to be misanthropic in its messaging. I'm probably too old-fashioned, but this honestly sounds like Marlboro launching the slogan "lung cancer for the weak of mind".
I've temporarily switched back to o3, thankfully that model is still in the switcher.
edit: s/month/week
You can find it right next to the image you are talking about.
LLMs have always been very subhuman at vision, and GPT-5.2 continues in this tradition, but it's still a big step up over GPT-5.1.
One way to get a sense of how bad LLMs are at vision is to watch them play Pokemon. E.g.,: https://www.lesswrong.com/posts/u6Lacc7wx4yYkBQ3r/insights-i...
They still very much struggle with basic vision tasks that adults, kids, and even animals can ace with little trouble.