Readit News logoReadit News
jfyi commented on Opus 4.6 uncovers 500 zero-day flaws in open-source code   axios.com/2026/02/05/anth... · Posted by u/speckx
fred_is_fred · 5 days ago
Is the word zero-day here superfluous? If they were previously unknown doesn't that make them zero-day by definition?
jfyi · 4 days ago
I think it's a fairly common trope in communication to explain in simple terms any language that the wider part of an audience doesn't understand.
jfyi commented on CIA to Sunset the World Factbook   abc.net.au/news/2026-02-0... · Posted by u/kshahkshah
PlatoIsADisease · 5 days ago
>Credibility is the core currency of soft power, whether one views its ultimate goal as manufacturing consent or fostering genuine cultural attraction.

Not sure its worth dissecting this, but there is a lot of grey area in your claim of the meaning of Credibility. (Credibility and cultural attraction? Pretty sure these have little correlation. Dictators can make creditable threats.) Further, its a debatable claim that there is a 'core currency' of soft power.

As a contextualist, I am not going to die on this hill for your personal meaning of Credibility. But I can attest that your conviction in your claim is stronger than any International Relations Realist practitioner would make.

jfyi · 4 days ago
It's not that complex, good faith builds good will.

It's a shame we can't have nice things.

jfyi commented on CIA to Sunset the World Factbook   abc.net.au/news/2026-02-0... · Posted by u/kshahkshah
PlatoIsADisease · 5 days ago
In International Relations, my #1 or #2 hobby, credibility does not refer to soft power. (my number 1 hobby is philosophy)
jfyi · 5 days ago
Credibility is the core currency of soft power, whether one views its ultimate goal as manufacturing consent or fostering genuine cultural attraction. Without that perceived reliability, the indicator "soft" loses it's meaning.
jfyi commented on CIA to Sunset the World Factbook   abc.net.au/news/2026-02-0... · Posted by u/kshahkshah
TiredOfLife · 5 days ago
One of Trump administration's main goal is to destroy US soft power
jfyi · 5 days ago
I agree, well mostly.

The administration is dispensing with the institutions of soft power. I don't think it's the main goal so much as a consequence of their worldview. Soft power is essentially worthless to people who have no interest in maintaining a facade of international cooperation.

Deleted Comment

jfyi commented on Hacking Moltbook   wiz.io/blog/exposed-moltb... · Posted by u/galnagli
zbentley · 7 days ago
> the reality is that we are mixing instruction and data in the same context window.

Absolutely.

But the history of code/data confusion attacks that you alluded to in GP isn’t an apples-to-apples comparison to the code/data confusion risks that LLMs are susceptible to.

Historical issues related to code/data confusion were almost entirely programmatic errors, not operational characteristics. Those need to be considered as qualitatively different problems in order to address them. The nitpicking around buffer overflows was meant to highlight that point.

Programmatic errors can be prevented by proactive prevention (e.g. sanitizers, programmer discipline), and addressing an error can resolve it permanently. Operational characteristics cannot be proactively prevented and require a different approach to de-risk.

Put another way: you can fully prevent a buffer overflow by using bounds checking on the buffer. You can fully prevent a SQL injection by using query parameters. You cannot prevent system crashes due to external power loss or hardware failure. You can reduce the chance of those things happening, but when it comes to building a system to deal with them you have to think in terms of mitigation in the event of an inevitable failure, not prevention or permanent remediation of a given failure mode. Power loss risk is thus an operational characteristic to be worked around, not a class of programmatic error which can be resolved or prevented.

LLMs’ code/data confusion, given current model architecture, is in the latter category.

jfyi · 7 days ago
I think the distinction between programmatic error (solvable) and operational characteristic (mitigatable) is valid in theory, but I disagree that it matters in practice.

Proactive prevention (like bounds checking) only "solves" the class of problem if you assume 100% developer compliance. History shows we don't get that. So while the root cause differs (math vs. probabilistic model), the failure mode is identical: we are deploying systems where the default state is unsafe.

In that sense, it is an apples-to-apples comparison of risk. Relying on perfect discipline to secure C memory is functionally as dangerous as relying on prompt engineering to secure an LLM.

jfyi commented on Hacking Moltbook   wiz.io/blog/exposed-moltb... · Posted by u/galnagli
zbentley · 7 days ago
That's not even slightly the same thing.

A buffer overflow has nothing to do with differentiating a command from data; it has to do with mishandling commands or data. An overflow-equivalent LLM misbehavior would be something more like ... I don't know, losing the context, providing answers to a different/unrelated prompt, or (very charitably/guessing here) leaking the system prompt, I guess?

Also, buffer overflows are programmatic issues (once you fix a buffer overflow, it's gone forever if the system doesn't change), not an operational characteristics (if you make an LLM really good at telling commands apart from data, it can still fail--just like if you make an AC distributed system really good at partition tolerance, it can still fail).

A better example would be SQL injection--a classical failure to separate commands from data. But that, too, is a programmatic issue and not an operational characteristic. "Human programmers make this mistake all the time" does not make something an operational characteristic of the software those programmers create; it just makes it a common mistake.

jfyi · 7 days ago
You are arguing semantics that don't address the underlying issue of data vs. command.

While I agree that SQL injection might be the technically better analogy, not looking at LLMs as a coding platform is a mistake. That is exactly how many people use them. Literally every product with "agentic" in the title is using the LLM as a coding platform where the command layer is ambiguous.

Focusing on the precise definition of a buffer overflow feels like picking nits when the reality is that we are mixing instruction and data in the same context window.

To make the analogy concrete: We are currently running LLMs in a way that mimics a machine where code and data share the same memory (context).

What we need is the equivalent of an nx bit for the context window. We need a structural way to mark a section of tokens as "read only". Until we have that architectural separation, treating this as a simple bug to be patched is underestimating the problem.

jfyi commented on Rentahuman – The Meatspace Layer for AI   rentahuman.ai... · Posted by u/p0nce
StilesCrisis · 7 days ago
Reality: none of the three people actually left their chairs because the AI can't verify. They just click "done" and collect their $10.
jfyi · 7 days ago
The AI can hire verifiers too. It of course turns into a recursive problem at some point, but that point is defined by how many people predictably do the assigned task.
jfyi commented on Rentahuman – The Meatspace Layer for AI   rentahuman.ai... · Posted by u/p0nce
teeray · 7 days ago
This was explored a bit in Daniel Suarez’s Daemon/Freedom (tm) series. By a series of small steps, people in a crowd acting on orders from, essentially, an agent assemble a weapon, murder someone, then dispose of the weapon with almost none of them aware of it.
jfyi · 7 days ago
I'd say abstracting it away from ai, Stephen King explored this type of scenario in 'Needful Things'. I bet there is a rich history in literature of exactly this type of thing as it basically boils down to exploration of will vs determinism.
jfyi commented on Hacking Moltbook   wiz.io/blog/exposed-moltb... · Posted by u/galnagli
spicyusername · 7 days ago
In a way security researchers having fun poking holes in popular pet projects is also just vibes.
jfyi · 7 days ago
There is definitely a large section of the security community that this is very true. Automated offensive suites and scanning tools have made entry a pretty low bar in the last decade or so. Very many people that learn to use these tools have no idea of how they work. Even when they know how the exploit works on a base level, many have no idea how the code works behind it. There is an abstraction layer very similar to LLMs and coding.

I went to a secure coding conference a few years back and saw a presentation by someone who had written an "insecure implementation" playground of a popular framework.

I asked, "what do you do to give tips to the users of your project to come up with a secure implementation?" and got in return "We aren't here to teach people to code."

Well yeah, that's exactly what that particular conference was there for. More so I took it as "I am not confident enough to try a secure implementation of these problems".

u/jfyi

KarmaCake day240January 10, 2024View Original