Readit News logoReadit News

Deleted Comment

jjoonathan commented on Claude says “You're absolutely right!” about everything   github.com/anthropics/cla... · Posted by u/pr337h4m
klik99 · 20 days ago
Is that GPT5? Reddit users are freaking out about losing 4o and AFAICT it's because 5 doesn't stroke their ego as hard as 4o. I feel there are roughly two classes of heavy LLM users - one who use it like a tool, and the other like a therapist. The latter may be a bigger money maker for many LLM companies so I worry GPT5 will be seen as a mistake to them, despite being better for research/agent work.
jjoonathan · 20 days ago
No, that was 4o. Agreed about factual prompts showing less sycophancy in general. Less-factual prompts give it much more of an opening to produce flattery, of course, and since these models tend to deliver bad news in the time-honored "shit sandwich" I can't help but wonder if some people also get in the habit of consuming only the "slice of bread" to amplify the effect even further. Scary stuff!
jjoonathan commented on Claude says “You're absolutely right!” about everything   github.com/anthropics/cla... · Posted by u/pr337h4m
aatd86 · 20 days ago
LLMs definitely have personalities. And changing ones at that. gemini free tier was great for a few days but lately it keeps gaslighting me even when it is wrong (which has become quite often on the more complex tasks). To the point I am considering going back to claude. I am cheating on my llms. :D

edit: I realize now and find important to note that I haven't even considered upping the gemini tier. I probably should/could try. LLM hopping.

jjoonathan · 20 days ago
Yeah, the heavily distilled models are very bad with hallucinations. I think they use them to cover for decreased capacity. A 1B model will happily attempt the same complex coding tasks as a 1T model but the hard parts will be pushed into an API call that doesn't exist, lol.
jjoonathan commented on Claude says “You're absolutely right!” about everything   github.com/anthropics/cla... · Posted by u/pr337h4m
NohatCoder · 20 days ago
This is such a useful feature.

I'm fairly well versed in cryptography. A lot of other people aren't, but they wish they were, so they ask their LLM to make some form of contribution. The result is high level gibberish. When I prod them about the mess, they have to turn to their LLM to deliver a plausibly sounding answer, and that always begins with "You are absolutely right that [thing I mentioned]". So then I don't have to spend any more time wondering if it could be just me who is too obtuse to understand what is going on.

jjoonathan · 20 days ago
ChatGPT opened with a "Nope" the other day. I'm so proud of it.

https://chatgpt.com/share/6896258f-2cac-800c-b235-c433648bf4...

Deleted Comment

jjoonathan commented on Europe's new heavy-lift rocket, Ariane 6, made its inaugural flight   esa.int/Enabling_Support/... · Posted by u/gostsamo
m4rtink · a year ago
Well, RD-180 is not really a suitable engine for modern ICBMs due to the need for a cryogenic oxidizer, resulting in the ICBM not being a very responsive design. But you are certainly correct about the engineers.
jjoonathan · a year ago
Good point. Still, I have to imagine that the engines themselves are dual use in some regard. GNSS or spy satellites maybe? These days it seems like everyone and their dog has a GNSS constellation, but it wasn't always that way.
jjoonathan commented on Europe's new heavy-lift rocket, Ariane 6, made its inaugural flight   esa.int/Enabling_Support/... · Posted by u/gostsamo
sidewndr46 · a year ago
Given that ULA historically purchased engines from Russia, are you suggesting that it was a direct mechanism for transferring tax revenue to Russia companies?
jjoonathan · a year ago
Yes. The idea was to keep the engines (and engineers) out of the hands of the other likely buyers. You've seen how soviet military surplus gets around: the same channels work for rocket engines, and those engines work in ICBMs just as well as they work in orbital launch platforms.

I don't know how effective this was. Did it backfire by promoting economies of scale in a program that went on to sell to adversaries anyway? Did it murder the domestic engine programs and did that have knock-on consequences? I don't know if the policy was effective, but I do know that stopping "engine proliferation" was a widely given and accepted reason for those programs.

jjoonathan commented on SF's AI boom can't stop real estate slide, as office vacancies reach new record   cnbc.com/2024/07/08/san-f... · Posted by u/koolba
jjoonathan · a year ago
Offices revalued due to increase in WfH? Those dastardly Democrats must be at it again!
jjoonathan commented on Anxious Generation – How Safetyism and Social Media Are Damaging the Kids   matija.eu/posts/anxious-g... · Posted by u/munyak
superkuh · a year ago
If it were 1900 they'd blame newspapers for "damaging" kids. If it were 1930 they'd blame radio. If it were 1950 it'd be television. If it were 1980 it'd be video games. What all of these have in common is being completely bogus.
jjoonathan · a year ago
If it were 400BCE they'd blame writing

    this discovery of yours will create forgetfulness
    in the learners' souls, because they will not use
    their memories; they will trust to the external written
    characters and not remember of themselves
- Plato, quoting Socrates

Deleted Comment

u/jjoonathan

KarmaCake day24628July 17, 2012
About
If you want to send me email, my username is "jon" and my personal domain is "jdewerd.com".
View Original