Readit News logoReadit News
wilhelm____ commented on ChatGPT Can Determine Your Age Indirectly   twitter.com/will_thompson... · Posted by u/wilhelm____
wilhelm____ · a year ago
An example of the current state of LLM reasoning.
wilhelm____ commented on What we know about LLMs   willthompson.name/what-we... · Posted by u/wilhelm____
Lich · 2 years ago
> "ha ha AI like crypto amirite?"

I don’t think that was the meaning at all. I think the image was supposed to convey how the crypto grifters and con artists were veering into AI to run scams under the guise of AI.

wilhelm____ · 2 years ago
^^ yes
wilhelm____ commented on What we know about LLMs   willthompson.name/what-we... · Posted by u/wilhelm____
llm_nerd · 2 years ago
"Crypto VCs & ”builders” making a hard left into AI"

This is a humorous intro graphic caption, but this sentiment appears on here constantly and it's self-destructive. This response might seem a bit over the top to a funny graphic, but I am replying to the general "ha ha AI like crypto amirite?" sentiment that is incredibly boring and worn out.

When confronted with challenging new technology that we don't understand, some knee-jerk to acting dismissive. As if that has any hope at all of changing outcomes.

It's especially weird when people who are clearly on the "I must desperately learn this as quickly as I can and try to present myself as some sort of expert" still incant the rhetoric -- "joking on the square" as it were -- as if they need to defend their prior dismissals. Constantly on here there is yet another trivial "intro to tokenization" blog entry that brays some tired crypto comparison.

Stop it.

The Venn diagram of people at the forefront of ML/LLM, and its advocates, is almost entirely separate from the web/crypto sphere. There is astonishingly little overlap. Crypto was hyped because some people truly saw a purpose, coupled with masses of scammers and getrichquick sorts. AI/LLM/ML is hyped because it is revolutionary and has already yielded infinitely more practical impact than crypto ever did.

wilhelm____ · 2 years ago
On the contrary, I agree - while there is certainly hype being generated around AI, particularly generated by the "VC hype cycle", the fundamental advancements we've made with LLMs are quite real.

Part of the reason I wrote this is to separate the signal from the noise and why one should be {cautiously, more tempered} optimistic in the medium term.

wilhelm____ commented on What we know about LLMs   willthompson.name/what-we... · Posted by u/wilhelm____
yding · 2 years ago
Good job Will!
wilhelm____ · 2 years ago
thank you, Yi!
wilhelm____ commented on What we know about LLMs   willthompson.name/what-we... · Posted by u/wilhelm____
jawerty · 2 years ago
I see, that’s my misunderstanding I was grouping all training as pretraining
wilhelm____ · 2 years ago
pre-training is developing the language model's base understanding of conditional word probabilities.

SFT and RLHF is attempting to further guide the model in terms of steerability + alignment of output.

In fact, the InstructGPT authors were worried about losing the pre-trained model's underlying probability distribution, so they try a version where it penalizes the model deviating too significantly from the original distribution (using KL). I don't remember them seeing a significant difference in performance.

u/wilhelm____

KarmaCake day138May 19, 2020
About
https://willthompson.name
View Original