This is surely the greatest weakness of current LLMs for any task needing a spark of creativity.
Ordered approximately by recency:
Banking? Clocks? Roman aqueducts? Mayan calendars? The sun rising every day? Predictable rainy and dry season?
How is software the outlier here?
People on HN regularly claim that LLMs are useless if they aren’t 100% accurate all the time. I don’t think this is true. We work around that kind of thing every day.
With your examples:
- Before computers, fraud and human error was common in the banking system. We designed a system that was resilient against this and mostly worked, most of the time, well enough for most purposes even though it was built on an imperfect foundation.
- Highly precise clocks are a recent invention. For regular people 200 years ago, one person’s clock would often be 5-10 minutes off from someone else’s. People managed to get things done anyway.
I’ll grant you that Roman aqueducts, seasons and the sun are much more reliable than computers (as are all the laws of nature).
The rigidness and near-perfect reliability of computer software is the unusual thing in human history, an outlier we’ve gotten used to.
At some point we should probably take a step back and ask “Why do we want to solve this problem?” Is a world where AI systems are highly intelligent tools, but humans are needed to manage the high level complexity of the real world… supposed to be a disappointing outcome?
Sure, manually selecting model may not have been ideal. But manually prompting to get your model feels like an absurd hack
Is this true? I couldn’t find another source discussing it. That would be insane behavior for a package manager.