Readit News logoReadit News
bt1a commented on AI documentation you can talk to, for every repo   deepwiki.com/... · Posted by u/jicea
vissi · a month ago
> LLMs are trained on wrong autogenerated documentation: a downward spiral for hallucinations! (Maybe this one could then force users go look for the official docs? But not sure at this point…)

On this, I think, we should have some kind of AI-generated meta-tag, like this: https://github.com/whatwg/html/issues/9479

bt1a · a month ago
I wonder what incentives for adherence to the use of this meta-tag might exist? For example, imagine I send you my digital resume and it has an AI-generated footer tag on display? Maybe a bad example- I like the idea of this in general, but my mind wanders to the fact that large entities completely ignored the wishes of robots.txt when collecting the internet's text for their training corpuses
bt1a commented on Anthropic irks White House with limits on models’ use   semafor.com/article/09/17... · Posted by u/mindingnever
chatmasta · 3 months ago
Ah my mistake. I thought they were French. I got them confused with Mistral.

The concern remains even if it’s a US corporation though (not government owned servers).

bt1a · 3 months ago
Everyone spies and abuses individuals' privacy. What difference does it make? (Granted I would agree with you if Anthropic were indeed a foreign based entity, so am I contradicting myself wonderfully?)
bt1a commented on Anthropic irks White House with limits on models’ use   semafor.com/article/09/17... · Posted by u/mindingnever
matula · 3 months ago
There are (or at least WERE) entire divisions dedicated to reading every letter of the contract and terms of service, and usually creating 20 page documents seeking clarification for a specific phrase. They absolutely know what they're getting into.
bt1a · 3 months ago
Perhaps it's the finetune of Opus/Sonnet/whatever that is being served to the feds that is the source of the refusal :)
bt1a commented on Geedge and MESA leak: Analyzing the great firewall’s largest document leak   gfw.report/blog/geedge_an... · Posted by u/yourapostasy
wartywhoa23 · 3 months ago
This whole discussion is full of devil's advocates. The society is fucked
bt1a · 3 months ago
Lolz
bt1a commented on I'm absolutely right   absolutelyright.lol/... · Posted by u/yoavfr
ZaoLahma · 4 months ago
I find the GPT-5 model having turned the friendliness way, way down. Topics that previously would have rendered long and (usefully) engaging conversations are now met with an "ok cool" kind of response.

I get it - we don't want LLMs to be reinforces of bad ideas, but sometimes you need a little positivity to get past a mental barrier and do something that you want to do, even if what you want to do logically doesn't make much sense.

An "ok cool" answer is PERFECT for me to decide not to code something stupid (and learn something useful), and instead go and play video games (and learn nothing).

bt1a · 4 months ago
I have been using gpt-5 through the API a bit recently, and I somewhat felt this response behavior, but it's definitely confirming to hear this from another. It's much more willing (vs gpt-4*) to tell me im a stupid piece of shxt and to not do what im asking of the initial prompt
bt1a commented on I'm absolutely right   absolutelyright.lol/... · Posted by u/yoavfr
teucris · 4 months ago
I think there’s a bit of parroting going around but LLMs are predictive and there’s a lot you can inuit a lot about how they behave just on that fact alone. Sure, calling it “token” prediction is oversimplifying things, but stating that, by their nature, LLMs are guessing at the next most likely thing in the scenario (next data structure needing to be coded up, next step in a process, next concept to cover in a paragraph, etc.) is a very useful mental model.
bt1a · 4 months ago
I would challenge the utility of this mental model as again they're not simply tracing a "most likely" path unless your sampling methods are trivially greedy. I don't know of a better way to model it, and I promise I'm not trying to be anal here
bt1a commented on Warp Code: the fastest way from prompt to production   warp.dev/blog/introducing... · Posted by u/brainless
bt1a · 4 months ago
optimizing for the shortest path from idea to prod sounds a tad warped, if i may
bt1a commented on WinBoat: Run Windows apps on Linux with seamless integration   github.com/TibixDev/winbo... · Posted by u/nogajun
russnes · 4 months ago
I've found that running non steam apps on steam with the proton experimental compatibility usually just works, it has become my go to solution
bt1a · 4 months ago
Do you use an application launcher / configuration manager like Lutris to do this? Or do you mean directly through steam? There's a steam game that I play often that tends to work the most frequently with proton hotfix for reasons unknown to me.
bt1a commented on Model intelligence is no longer the constraint for automation   latentintent.substack.com... · Posted by u/drivian
Kuinox · 4 months ago
It's specific model that run for maths. GPT-5 and Gemini 2.5 still cannot compute an arbitrary length sum of whole number without a calculator. I have a proceduraly generated benchmark of basic operations, LLMs gets better at it with time, but they cant still solve basic maths or logic problems.

BTW I'm open to selling it, my email is on my hn profile.

bt1a · 4 months ago
i'd wager your benchmark problems require cumbersome arithmetic or are poorly worded / inadequately described. or, you're mislabeling them as basic math and logic (a domain within which LLMs have proven their strengths!)

i only call this out because you're selling it and don't hypothesize* on why they fail your simple problems. i suppose an easily aced bench wouldn't be very marketable

bt1a commented on Death by AI   davebarry.substack.com/p/... · Posted by u/ano-ther
bt1a · 5 months ago
giggled like a child through this one

u/bt1a

KarmaCake day311March 23, 2020
About
no text
View Original