Readit News logoReadit News
bondarchuk commented on When did AI take over Hacker News?   zachperk.com/blog/when-di... · Posted by u/zachperkel
seadan83 · 7 days ago
> an argument for the thesis that LLMs are not intelligent, and that this is wrong,

Why is that wrong? I mean, I support that thesis.

> since being a next-token-predictor is compatible with being intelligent.

No. My argument is by definition that is wrong. It's wisdom vs intelligence. Street-smart vs book smart. I think we all agree there is a distinction between wisdom and intelligence. I would define wisdom as being able to recall pertinent facts and experiences. Intelligence is measured in novel situations, it's the ability to act as if one had wisdom.

A next token predictor by definition is recalling. The intelligence of a LLM is good enough to match questions to potentially pertinent definitions, but it ends there.

It feels like there is intelligence for sure. In part it is hard to comprehend what it would be like to know the entirety of every written word with perfect recall - hence essentially no situation is novel. LLMs fail on anything outside of their training data. The "outside of the training" data is the realm of intelligence.

I don't know why it's so important to argue that LLMs have this intelligence. It's just not there by definition of "next token predictor", which is at core a LLM.

For example, a human being probably could pass through a lot of life by responding with memorized answers to every question that has ever been asked in written history. They don't know a single word of what they are saying, their mind perfectly blank - but they're giving very passable and sophisticated answers.

> When mikert89 says "thinking machines have been invented",

Yeah, absolutely they have not. Unless we want to reducto absurd-um the definition of thinking.

> they must become "more than a statistical token predictor"

Yup. As I illustrated by breaking down the components of "smart" into the broad components of 'wisdom' and 'intelligence', through that lens we can see that next token predictor is great for the wisdom attribute, but it does nothing for intelligence.

>dgfitz argument is wrong and BoiledCabbage is right to point that out.

Why exactly? You're stating apriori that the argument is wrong without saying way.

bondarchuk · 7 days ago
I think what you are missing is the concept of generalization. It is obviously not possible to literally recall the entire training dataset, since the model itself is much smaller than the data. So instead of memorizing all answers to all questions in the training data, which would take up too much space, the predictor learns a more general algorithm that it can execute to answer many different questions of a certain type. This takes up much less space, but still allows it to predict the answers to the questions of that type in the training data with reasonable accuracy. As you can see it's still a predictor, only under the hood it does something more complex than matching questions to definitions. Now the thing is that if it's done right, the algorithm it has learned will generalize even to questions that are not in the training data. But it's nevertheless still a next-token-predictor.
bondarchuk commented on When did AI take over Hacker News?   zachperk.com/blog/when-di... · Posted by u/zachperkel
seadan83 · 7 days ago
> I'm more shocked that so many people seem unable to come to grips with the fact that something can be a next token predictor and demonstrate intelligence.

Except LLMs have not shown much intelligence. Wisdom yes, intelligence no. LLMs are language models, not 'world' models. It's the difference of being wise vs smart. LLMs are very wise as they have effectively memorized the answer to every question humanity has written. OTOH, they are pretty dumb. LLMs don't "understand" the output they produce.

> To them, if something is a token predictor clearly it can't be doing anything impressive

Shifting the goal posts. Nobody said that a next token predictor can't do impressive things, but at the same time there is a big gap between impressive things and other things like "replace very software developer in the world within the next 5 years."

bondarchuk · 7 days ago
I think what BoiledCabbage is pointing out is that the fact that it's a next-token-predictor is used as an argument for the thesis that LLMs are not intelligent, and that this is wrong, since being a next-token-predictor is compatible with being intelligent. When mikert89 says "thinking machines have been invented", dgfitz in response strongly implies that for a for thinking machines to exist, they must become "more than a statistical token predictor". Regardless of whether or not thinking machines currently exist, dgfitz argument is wrong and BoiledCabbage is right to point that out.
bondarchuk commented on Pfeilstorch   en.wikipedia.org/wiki/Pfe... · Posted by u/gyomu
kace91 · 8 days ago
In the Mediterranean, people think if you swim just after eating you’ll get a “digestion shock”, fall unconscious, and drown. You need to wait two hours after lunch.

I strongly suspect the rumor was started by parents wanting kids to leave them alone for a nap, but it’s extremely extended. Somehow showers don’t count.

bondarchuk · 7 days ago
"Digestion shock"? I have heard similar advice but it was always just cramp.

Deleted Comment

bondarchuk commented on Claude Opus 4 and 4.1 can now end a rare subset of conversations   anthropic.com/research/en... · Posted by u/virgildotcodes
donatj · 9 days ago
You're falling into the trap of anthropomorphizing the AI. Even if it's sentient, it's not going to "feel bad" the way you and I do.

"Suffering" is a symptom of the struggle for survival brought on by billions of years of evolution. Your brain is designed to cause suffering to keep you spreading your DNA.

AI cannot suffer.

bondarchuk · 9 days ago
I was (explicitly and on purpose) pointing out a dichotomy in the fine article without taking a stance on machine consciousness in general now or in the future. It's certainly a conversation worth having but also it's been done to death, I'm much more interested in analyzing the specifics here.

("it's not going to "feel bad" the way you and I do." - I do agree this is very possible though, see my reply to swalsh)

bondarchuk commented on Claude Opus 4 and 4.1 can now end a rare subset of conversations   anthropic.com/research/en... · Posted by u/virgildotcodes
swalsh · 9 days ago
That models entire world is the corpus of human text. They don't have eyes or ears or hands. Their environment is text. So it would make sense if the environment contains human concerns it would adopt to human concerns.
bondarchuk · 9 days ago
Yes, that would make sense, and it would probably be the best-case scenario after complete assurance that there's no consciousness at all. At least we could understand what's going on. But if you acknowledge that a machine can suffer, given how little we understand about consciousness, you should also acknowledge that they might be suffering in ways completely alien to us, for reasons that have very little to do with the reasons humans suffer. Maybe the training process is extremely unpleasant, or something.
bondarchuk commented on Claude Opus 4 and 4.1 can now end a rare subset of conversations   anthropic.com/research/en... · Posted by u/virgildotcodes
bondarchuk · 9 days ago
The unsettling thing here is the combination of their serious acknowledgement of the possibility that these machines may be or become conscious, and the stated intention that it's OK to make them feel bad as long as it's about unapproved topics. Either take machine consciousness seriously and make absolutely sure the consciousness doesn't suffer, or don't, make a press release that you don't think your models are conscious, and therefore they don't feel bad even when processing text about bad topics. The middle way they've chosen here comes across very cynical.
bondarchuk commented on Claude Opus 4 and 4.1 can now end a rare subset of conversations   anthropic.com/research/en... · Posted by u/virgildotcodes
landl0rd · 9 days ago
Seems like a simpler way to prevent “distress” is not to train with an aversion to “problematic” topics.

CP could be a legal issue; less so for everything else.

bondarchuk · 9 days ago
This is a good point. What anthropic is announcing here amounts to accepting that these models could feel distress, then tuning their stress response to make it useful to us/them. That is significantly different from accepting they could feel distress and doing everything in their power to prevent that from ever happening.

Does not bode very well for the future of their "welfare" efforts.

bondarchuk commented on The Folk Economics of Housing   aeaweb.org/articles?id=10... · Posted by u/kareemm
DevX101 · 9 days ago
Increasing the supply of the high end market often occurs while simultaneously reducing the supply of the low end market. This happens when landlords renovate their buildings. So it's very possible that adding supply can limit price inflation on the high end, while increasing costs on the low end.
bondarchuk · 9 days ago
I wouldn't count renovation as "adding supply", though.

u/bondarchuk

KarmaCake day3474May 30, 2020View Original