Readit News logoReadit News
evidenciary commented on GPT-5: It just does stuff   oneusefulthing.org/p/gpt-... · Posted by u/paulpauper
DamnInteresting · 4 months ago
OpenAI has built the world's most impressive and sophisticated generator of blog posts about ChatGPT.
evidenciary · 4 months ago
I realized the other day during a conversation that this hype cycle is built on top a product that hypes itself.

In any scifi story this would be considered bad writing, yet here we are. Late stage capitalism has created a product that actively nurtures emotional dependence and hypes itself.

evidenciary commented on GPT-5: It just does stuff   oneusefulthing.org/p/gpt-... · Posted by u/paulpauper
r0fl · 4 months ago
For non programming tasks ChatGPT has not improved that much since 3.5

Today I used GPT 5 to help plan a trip. It told me a market is open on Saturdays and then when it built an itinerary it schedule me to go there on Sunday

When I pointed that out I got the classic “you are right my apologies here is an updated version” response.

It’s ridiculous that it makes simple yet huge mistakes like that!

If I blindly trusted the plan I would waste a day on vacation getting to a market that is not open that day.

It does not “just do stuff”

evidenciary · 4 months ago
It is incredible to me that there is this buggy software out there and people keep insisting that it's "hallucinations". Like, this is ""AI"" and it ""hallucinates"".

No. It's buggy and has been buggy for years and everyone keeps making excuses because they just want so hard to believe.

evidenciary commented on I tried coding with AI, I became lazy and stupid   thomasorus.com/i-tried-co... · Posted by u/mikae1
jdiff · 4 months ago
It's reasonable for tools to produce reasonable, predictable output to enable them to be used well. A tool can have awful, dangerous failure modes as long as they're able to be anticipated and worked around. This is the critical issue with AI, it's not deterministic.

And because it always comes up, no, not even if temperature is set to 0. It still hinges on insignificant phrasing quirks, and the tiniest change can produce drastically different output. Temperature 0 gives you reproducibility but not the necessary predictability for a good tool.

evidenciary · 4 months ago
I don't think the "non-deterministic" accusation is a good one. Same as "hallucination", it's a bit of misdirection.

These LLMs are buggy. They have bugs. They don't do what they promise. They do it sometimes, other times they give garbled output.

This is buggy software. And after years and billions of dollars, the bug persists.

u/evidenciary

KarmaCake day3August 11, 2025View Original