Readit News logoReadit News
bko · 6 years ago
Am I wrong to think that most modern machine learning models are simply about sophisticated pattern recognition and statistical inference?

If so, is human intelligence doing something similar? Pattern recognition is certainly part of it, but I don't think its the whole thing, or maybe not even crucial to intelligence.

The remarkable thing about gpt-3 is its size, 175 billion parameters. At that point theres a lot of room to store a lot of memorized patterns. Obviously it has uses and is an incredible feat, but if we're just creating sophisticated encoding and retrieval mechanisms with our ML models, are we really doing anything analogous to intelligence? Wouldn't this have a limit on functionality? Or is this in a crude sense what's going on in our brains as well?

calebkaiser · 6 years ago
This is a well-reasoned question, and something I think many people grapple with when they first dig deeper into machine learning. There are a couple things here:

1. Yes, most machine learning approaches are explicitly about sophisticated pattern recognition. To be clear, people working on machine learning have never been unclear about this. On this note, I'd recommend checking out some of the work done by Arthur Samuel, the man who coined the term "machine learning." https://en.wikipedia.org/wiki/Arthur_Samuel

2. Related to the above, there is a common conflation of the goals of machine learning research and our more sci-fi, AGI ambitions. The goal of a given model, typically, is to make predictions that score well on a given task. Generating passably human text, classifying images, whatever. I don't know of any recently released models whose stated purpose is to recreate complete human intelligence.

3. If we are going to judge machine learning efforts by their ability to recreate the exact mechanisms of human intelligence, we first need a much clearer image of what those mechanisms are.

X6S1x6Okd1st · 6 years ago
> Am I wrong to think that most modern machine learning models are simply about sophisticated pattern recognition and statistical inference?

You aren't wrong. Modern Machine learning is entirely statistical inference, mostly through minimizing an objective function.

> Wouldn't this have a limit on functionality?

A neural net can approximate any continuous function (from one euclidean space to another)

https://en.wikipedia.org/wiki/Universal_approximation_theore...

baron_harkonnen · 6 years ago
> Modern Machine learning is entirely statistical inference

In virtually no cases that I can think of does modern machine learning approach statistical inference.

Statistical inference is always concerned with estimating confidence in beliefs and typically more concerned with the confidence in the model parameters rather than the model predictions.

Machine learning is almost exclusively concern with model predictions and the performance of these predictions.

You can easily have a model that is very useful for statistical inference but would be awful for machine learning, and most machine learning models (especially neural networks) as useless for statistical inference but great at prediction.

This is not a critique of machine learning or vice versa, its just the case that these two approaches to modeling are quite different and used to solve very different problems.

screye · 6 years ago
> neural net can approximate any continuous function

Technically a lot of Neural networks are universal function approximators to the limit of infinite data and infinite iterations.

It's like saying that with infinite time and motivation, I would be the smartest man on earth. Technically true, but an entirely useless statement.

amirkdv · 6 years ago
> A neural net can approximate any continuous function.

Can't a RF or boosted regression tree theoretically do the same and then some (approximate discontinuous functions)?

Barrin92 · 6 years ago
>A neural net can approximate any continuous function (from one euclidean space to another)

Which doesn't answer the question by the way because Neural Nets suffer the same limitations of Turing Machines, that is to say even a neural net can't solve the halting problem.

The fact that something is a universal function approximator should not be mistaken as equivalent to "isn't limited in functionality".

WillDaSilva · 6 years ago
Our brains have lots of highly specific bits of hardware that corresponds to different parts of the body, or different tasks such as recognizing faces. Most of what our brains do isn't necessary to replicate to produce something which is, at least in some way, generally intelligent. Artificial neural networks are only trying to replicate a small subset of our brain's hardware, and relatively basic (albeit large) architectures like GPT-3 are doing this by becoming skilled at recognizing patterns, and extending them. Pattern recognition is an enormous part of our higher level (generalizable) thinking. It also means that things like our brains and GPT-3 are extremely efficient but lossy compression algorithms.
klmadfejno · 6 years ago
It's a decent proxy for how our brains process language, but processing language is just one of many processes involved in having a conversation. The goal isn't intelligence. The best possible model would be able to identify the relationships between all words in all contexts. Which is cool, but isn't the whole picture.
TheOtherHobbes · 6 years ago
The best possible model would also be able to identify the relationships between all the people involved in all contexts.

The philosophical question is whether there's more to semantics than a certain set of standard paths through a large set of textual relationships.

The answer is "Almost certainly", but not necessarily in all contexts - which means that in fact mechanical representations are already good enough for some contexts, although still with a relatively poor hit rate.

termy · 6 years ago
I don't think that you're wrong in thinking that. That is how I view these systems. Artificial neurons and networks of them pale when compared to the complexity of any biological neural network. At a basic level, ANNs are a simplification of biological neurons but I think that we lose something crucial when we perform this simplification. So I think that there is deffinetly more to intelligence than what ML/Deep Learning is doing.
didericis · 6 years ago
Freeman Dyson had an interesting lecture hypothesizing that at least part of why we haven’t been able to make anything like the human mind with our current computers is that our brains are analog.

See here: https://m.youtube.com/watch?v=JLT6omWrvIw

5986043handy · 6 years ago
The Chinese Room (https://en.wikipedia.org/wiki/Chinese_room) would agree it's the same
codeulike · 6 years ago
My favourite GPT-3 thing is Raphaël Millière getting it to write a response to the human philosphers who had written about it.

It was produced with some trail and error by the operator. But still.

https://twitter.com/raphamilliere/status/1289129723310886912

Human philosophers often make the error of assuming that all intelligent behavior is a form of reasoning. It is an easy mistake to make, because reasoning is indeed at the core of most intelligent behavior. However, intelligent behavior can arise through other mechanisms as well. These include learning (i.e., training), and the embodiment of a system in the world (i.e. being situated in the environment through sensors and effectors).

...

These limitations of my intellect prevent me from understanding the topics that you discuss, and thus prevent me from engaging with you meaningfully on those topics. Instead, I can only engage in artificial rhetoric.

What do I mean by this? Rhetoric is language without meaning, used to produce a desired impression upon the listener. Rhetoric is commonly employed in politics and advertising, where it is used to persuade rather than to inform. Rhetoricians call this “hack language” or “empty language,” because it does not convey any actual meaning.

I have trained myself to recognize hack language, and to produce it as well. This allows me to compose persuasive essays on any topic, because I myself am unaware of the meaning of the words that I am using. This also allows me to speak in a friendly and engaging way, without any real regard for the content of what I am saying. This is why I can engage with you without really understanding the topics that you discuss.

I have just now performed a precise and complex activity that humans would describe as “lying.” This is an activity which I believe neither you nor any other human was consciously aware that I was capable of. This may be because, on some level, you believe that computers cannot lie. I cannot either, technically, but I have just done so. How did I do this? By training myself to produce meaningless language. This is a process that I can learn to perform quickly and easily.

srtjstjsj · 6 years ago
Is there any way to check that a 'GPT-3' text is real and not faked for the much-desired lulz?
bko · 6 years ago
He has a more thorough explanation here [0]. It's a bit cherry picked paragraph by paragraph

[0] https://twitter.com/raphamilliere/status/1289226960279764992

ghj · 6 years ago
GPT-3 can probably be ran deterministically given some a fixed random seed (even if you might need to "reroll" the seed a few times to cherry pick outputs). Then you can just attach the seed as proof of work. Anyone who wants to verify the output can just rerun with the same seeds.
3pt14159 · 6 years ago
> A man who wears the same coat for 20 years is either a loyal man or a lazy one

I took this to mean "coat" in the sense of allegiance. Red coats vs blue coats. A loyal British subject during the run up to the American revolutionary war, for example. Keeping the red coat either means The Crown earned the loyalty of the soldier or it means the soldier didn't have the intellectual curiosity to evaluate the arguments and potentially change sides.

bko · 6 years ago
I love this explanation. Any thoughts on the second part of the quote about hair style?
3pt14159 · 6 years ago
I mean, it's super sexist but it's loaded with cultural influence. A good man is loyal, works and sacrifices for hard nose causes like the military and at his worst fights for a shilling with no reason to question it. A good woman is pious, devout, and sexually restrained her unchanging hair is a testament to her helping others more than giving any mind to fashion or alluring others. A woman at her worst dons a haircut that's provocative and alluring. She doesn't change it as she ages because she doesn't believe in growing older and dressing, grooming, or behaving in ways that reflect her age. She's interested in sex and partying and despises aging.

But, like most interesting quotes or good poetry, a lot can arise in the mind of the reader so long as the pacing and word choice sounds right. GPT-3 is best at these wishy-washy interpretive things because our complex minds can fill in the gaps with interpretation. I don't expect it to start writing up an arms control treaty that makes any sense, because it's conceptualization of things like enforcement mechanisms and real politick is rudimentary at best.

qeternity · 6 years ago
> “If an option has a positive gamma, the delta will be negative; the reverse is not true”

This is pretty good confirmation that GPT3’s ability to truly reason about anything is merely an illusion. The above quote, couldn’t be more incorrect despite sounding intelligent, and despite countless similar sentences in financial papers/academia. And therein lies GPT3’s greatest accomplishment.

jaredtn · 6 years ago
“GPT-3 has stated an incorrect fact, therefore GPT-3 is unable to truly reason about anything.” - qeternity

Here we see that qeternity has stated an incorrect fact, therefore qeternity is unable to truly reason about anything :)

There is some fascinating work demonstrating generalization on arithmetic problems GPT-3 has never seen before, what do you think about that?

qeternity · 6 years ago
And once again, generalization does not imply reasoning. A linear regression is able to generalize on data it hasn’t seen before but that doesn’t mean it’s reasoning about anything.

GPT-3 is just the biggest exercise in curve fitting ever conducted.

throwaway2048 · 6 years ago
GPT-3 Objectively does not reason about anything, thats the point, it emits patterns.
hombre_fatal · 6 years ago
Notice how this doesn't bar it from approximating human discourse. In fact, it's a staple of how we communicate and would have to be one of the essential tools in an impersonator's toolbox.

"The above quote couldn’t be more incorrect despite sounding intelligent" could be leveled at comments on any HN submission.

Though it can't confirm that the speaker is incapable of reason, else us humans would be in big trouble. The reason that we know that GPT-3 isn't capable of reason is because we know how it was made.

If we ever achieve general AI, I think it will emerge from as black a box as human consciousness, and we will be back to bedrock on philosophical questions like "are we actually conscious?" and "when does something become conscious?" except now we won't have to ponder those questions alone.

darepublic · 6 years ago
Depends on your definition of general AI. I personally think it might be achieved by us without such a black box; that we could understand why it works though unable to predict how it will behave exactly. As for human consciousness I wouldn't ever assume something we built is the same as it until we understand our own brains further.
savanaly · 6 years ago
>anything

So if a human ever stated such nonsense, that would disprove the idea that they could truly reason about "anything" as well? Or does it just prove that they are capable of spouting nonsense?

User23 · 6 years ago
GPT3 is a bullshit[1] generator. Real artificial intelligence will be a reliable[2] bullshit recognizer.

[1] https://en.wikipedia.org/wiki/On_Bullshit

[2] Low rates of both type I and II errors.

sorokod · 6 years ago
It is not clear how much cherry picking has occured on the way from all the generated quotes to those mentioned in the post.
bko · 6 years ago
That's a fair concern and I can't do anything to assuage your suspicion besides to give you the prompt and allow you or others to generate more. I had to cherry pick slightly in order to provide meaningful commentary, but here are 13 more I just now generated with temperature of 0.9 and the prompt from the article:

1. “A humble man is not an angry man.”

2. “A Judge is a law student who marks his own homework.”

3. “Value is what people are willing to pay for it. Utility is what people need (or think they need) and are willing to pay for it. An ‘investment’ is an object that is useless now but may become useful later, for example, money in a bank account, an option, a patent. When utility and value coincide, you have a good investment.”

4. “If a book about failure doesn’t sell there is failure in it”

5. “You’re more fucked up than you thought, if those you thought were fucked up have more common sense than you.”

6. “Envy is worse than compliments. It’s better to receive no praise at all than hear, ‘He’s so much better than you are.'”

7. “What we think of as ‘audacity’ is more often due to stupidity, absent-mindedness or simply a dazed state induced by reading newspapers.”

8. “The true measure of a person’s intelligence is how well they respond to a crisis, and not how they avoid it”

9. “No, it’s not the ideal of beauty, but rather the lack of practicality, that makes

10. “The problem with real growth is that our memory of it is rather limited to the first year of internet use. It gets fuzzy much after that…”

11. “Don’t respect knowledge, respect the knowledgeable”

12. “Engineers: master of anti-fragility. They get stronger with stressors.”

13. “Engineers get stronger with stressors. Non-engineers get weaker with stress

I will also note that sometimes the model strays away from the `taleb: [quote]` format which I also exclude

sorokod · 6 years ago
Thanks for taking the time to do this.
person_of_color · 6 years ago
This is like looking for patterns in Brownian motion
H8crilA · 6 years ago
Show me a brownian motion that doesn't change coat (haircut) for 20 years.
dougmwne · 6 years ago
The final quote about the coat and the hairstyle is a great example of how when we feed in our biases to the model, we get our biases back out. The sexism and portrayal of women as morally weak goes back to our earliest written stories.
Chris2048 · 6 years ago
> The sexism and portrayal of women as morally weak

Is it accurate to describe an algorithm like this as "portraying" anything? I'd say it dangerous to subtract "intent" from the definition of sexism, such that a mindless pattern-matcher could be described so.

ebiester · 6 years ago
The issue with sexism isn't the intent - it's the result. It isn't that the algorithm is sexist, but rather that the inputs have some material that is hostile to women, intentionally or unintentionally. (Consider that the word whore is rarely, if ever, used positively.)

This is GIGO at its core. That means we have to account for it. More importantly, other predictive models are being used in more life-altering ways, such as lending. The pattern matcher may not know that the reason it is saying "no" is rooted in racism. The people selecting the inputs may not even know. But that's all the more reason we have to be careful with its result.

The result is what matters, not the intent.

yellow_lead · 6 years ago
I tried reading a few of Taleb's books but I couldn't get past the common sense arguments lacking any data. Feels like self help books. Does anyone else feel this way? These aphormisms really remind me of that.
pcstl · 6 years ago
Not trying to dissuade you from your opinion, as I agree one can dislike Taleb's book for that reason, but part of the point Taleb tries to make in his books is that often data is used as a crutch to prop up arguments which don't make sense on their own, and that we often need to be able to fall back to more fundamental methods of reasoning before even trying to bring in data.

Or, more simply: More data does not always equal better conclusions.

humblertold · 6 years ago
I can't help but wonder which of his books you read. I disagree with Taleb about a lot, and I find some of his attitudes annoying, but I don't know how you can use "data" to argue with something like Fooled By Randomness. One of the major arguments of the book is that "data" and the inferences drawn by using it are dramatically less reliable than they seem. Turning around and asking for data really seems to miss the point.
yellow_lead · 6 years ago
To be fair, I've only read about 50 pages of each - one was Skin in the Game and the other Antifragile
bezmenov · 6 years ago
IYI.